var/home/core/zuul-output/0000755000175000017500000000000015145277765014550 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015145307357015503 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000313114315145307212020254 0ustar corecoreikubelet.log_o[;r)Br'o-n(!9t%Cs7}g/غIs,r.k9GfB >KYEڤ펯_ˎ6Ϸ7+%f?ᕷox[o8W5X??xI[mEy},fۮWe~7Nû/wb~1;ZxsY~ݳ( 2[$7۫j{Zw鶾z?&~|XLXlN_/:oXx$%X"LADA@@tkޕf{5Wbx=@^J})K3x~JkwI|YowS˷j̶֛]/8 N Rm(of`\r\L>{Jm 0{vR̍>dQQ.aLk~g\UlxDJfw6xi1U2 c#FD?2SgafO3|,ejoLR3[ D HJP1Ub2i]$HU^L_cZ_:F9TJJ{,mvgL;: ԓ$a;ɾ7lַ;̵3](uX|&kΆ2fb4NvS)f$UX dcю)""û5h< #чOɁ^˺b}0w8_jiB8.^s?Hs,&,#zd4XBu!.F"`a"BD) ᧁQZ-D\h]Q!]Z8HGU=y&|'oZƧe7ΣԟRxxXԨkJ[8 ";ЗH F=y܇sθm@%*'9qvD]9X&;cɻs0I٘]_fy tt('/V/TB/ap+V9g%$P[4D2L'1bЛ]\s΍ic-ܕ4+ޥ^.w[A9/vb֜}>| TXNrdTs>RDPhإek-*듌D[5l2_nH[׫yTNʹ<ws~^B.Ǔg'AS'E`hmsJU # DuT%ZPt_WďPv`9 C|mRj)CMitmu׀svRڡc0SAA\c}or|MKrO] g"tta[I!;c%6$V<[+*J:AI \:-rR b B"~?4 W4B3lLRD|@Kfځ9g ? j럚Sř>]uw`C}-{C):fUr6v`mSΟ1c/n߭!'Y|7#RI)X)yCBoX^P\Ja 79clw/H tBFKskޒ1,%$BվCh,xɦS7PKi0>,A==lM9Ɍm4ެ˧jOC d-saܺCY "D^&M){ߘ>:i V4nQi1h$Zb)ŠȃAݢCj|<~cQ7Q!q/pCTSqQyN,QEFKBmw&X(q8e&щu##Ct9Btka7v Ө⸇N~AE6xd~?D ^`wC4na~Uc)(l fJw>]cNdusmUSTYh>Eeք DKiPo`3 aezH5^n(}+~hX(d#iI@YUXPKL:3LVY~,nbW;W8QufiŒSq3<uqMQhiae̱F+,~Mn3 09WAu@>4Cr+N\9fǶy{0$Swwu,4iL%8nFВFL2#h5+C:D6A@5D!p=T,ښVcX㯡`2\fIԖ{[R:+I:6&&{Ldrǒ*!;[tʡP=_RFZx[|mi ǿ/&GioWiO[BdG.*)Ym<`-RAJLڈ}D1ykd7"/6sF%%´ƭ*( :xB_2YKoSrm_7dPΣ|ͣn/𚃚p9w#z A7yTJ$KOL-aP+;;%+_6'Sr|@2nQ{aK|bjܒ^o(מO80$QxBcXE ء\G=~j{Mܚ: hLT!uP_T{G7C]Ch',ެJG~Jc{xt zܳ'鮱iX%x/QOݸ}S^vv^2M!.xR0I(P 'fΑQ)ۢWP Pe>F=>l |fͨ3|'_iMcĚIdo阊;md^6%rd9#_v2:Y`&US tDkQ;>" ء:9_))wF|;~(XA PLjy*#etĨB$"xㄡʪMc~)j 1駭~բ>XiN .U轋RQ'Vt3,F3,#Y3,kJ3,LhVnKauomˠ_>2h-/ ђ(9Uq EmFjq1jX]DןR24d 3[n )ܗKj/jUSsȕD $([LH%%b~\b$BrW XWz<%fpG"m%6PGEH^*JL֗J)oEv[Ң߃x[䚒}0BOnYr猸p$nu̿ܣRF]NHw2kp}lrCy u)xF$Z83Ec罋}[εUX%}< ݻln"sv&{b%^AAoۺ(I#hKD:Bߩ#蘈f=9oN*.Ѓ M#JC1?tean`3-SHq$2[ĜSjXRx?}-m6Mw'yR3q㕐)HW'X1BEb $xd(21i)//_і/Cޮm0VKz>I; >d[5Z=4>5!!T@[4 1.x XF`,?Hh]b-#3J( &uz u8.00-(9ŽZcX Jٯ^蒋*k.\Mz_s߭AF'is^_&uUm$[[5HI4QCZ5!N&D[uiXk&2Bg&Ս7_/6v_cd쿽d@eU XyX_2z>g8:.⺻h()&nO5YE\1t7aSyFxPV19 ĕi%K"IcB j>Pm[E[^oHmmU̸nG Ym݃|M$ 6.x5 TMXbXj-P\jА޴y$j`ROA"Yg"ź)\3mf|ܔMUiU|Ym! #'ukMmQ9Blm]TO1ba.XW x6ܠ9[v35H;-]Um4mMrW-k#~fؤϋu_j*^Wj^qM `-P?k.@%=X#|ۡb1lKcj$bKv[~"N jS4HOkeF3LPyi︅iWk! cAnxu6<7cp?WN $?X3l(?  'Z! ,Z.maO_Bk/m~ޖ(<qRfR"Au\PmLZ"twpuJ` mvf+T!6Ѓjw1ncuwo':o gSPC=]U҅yY9 &K<-na'Xk,P4+`Þ/lX/bjoFO.= w ?>ȑ3n߿z,t s5Z/ Clo-` z?a~b mzkC zFȏ>1k*Dls6vP9hS  ehC.3 @6ijvUuBY hBnb[ Fr#D7ćlA!:X lYE>#0JvʈɌ|\u,'Y˲.,;oOwoj-25Hݻ7 li0bSlbw=IsxhRbd+I]Y]JP}@.供SЃ??w w@KvKts[TSa /ZaDžPAEư07>~w3n:U/.P珀Yaٳ5Ʈ]խ4 ~fh.8C>n@T%W?%TbzK-6cb:XeGhl&0Ɠbb__2++oI~!&-[TWvxZ>4(sgz1v&YN2姟d4"?oWNW݃yh~%DTt^W7q.@ L⃳662G,:* $: e~7[/P%F onKȌsK+D"̽E/"Icƀsu0,gy(&TI{ U܋N5 l͖h"褁lm *#n/Q!m b0X3i)\IN˭% Y&cKoG w 9pM^WϋQf7s#bd+SDL ,FZ<1Kx&C!{P|Ռr,+ ] O;*X]Eg,5,uZm8pnglVj!p2֬uT[Qy>" L+Y*Ha)j~pu7ި!:E#s:ic.XC^wT/]n2'>^&pnapckL>2QQWoݻ<̍8)r`F!Woc0Xq0 R' eQ&Aѣzvw=e&".awfShWjÅD0JkBh]s9Ąmʍ@1Ssc;l?ߨG~oB(ъ{zZJ }z&OF wkߓG9!1u8^drKkJBxF&+62,b.-Z*qqdX>$'dW<qIE2Ľ)5kJҼMЌ DR3csf6rRSr[I߽ogCc;S5ׂdKZ=M3դ#F;SYƘK`K<<ƛ G׌MU.APf\M*t*vw]xo{:l[n=`smFQµtxx7/W%g!&^=SzDNew(æ*m3D Bo.hI"!A6:uQզ}@j=Mo<}nYUw1Xw:]e/sm lˣaVۤkĨdԖ)RtS2 "E I"{;ōCb{yex&Td >@).p$`XKxnX~E膂Og<'J[FJ,4N:=6. +;$v6"I7%#CLTLyi{+ɠ^^fRa6ܮIN ޖ:DMz'rx#~w7U6=S0+ň+[Miw(W6 ]6ȧyԋ4ԙ./_A9B_-Z\PM `iĸ&^Ut (6{\٢K 5X\*wٗYS%g,0\ Rk k8P>x7v21՚H :[Γd!E'a4n?k[A׈(sob 41Y9(^SE@7`KIK`kx& V`X0,%pe_ן >hd xе"Q4SUwy x<'o_~#6$g!D$c=5ۄX[ു RzG:_[ӏ[3frl ô ހ^2TӘUAT!94[[m۾\T)W> lv+ H\FpG)ۏjk_c51̃^cn ba-X/#=Im41NLu\9ETp^poAOO&Ats]iSCQ&s~In/SZ % 'I Ƿ$M6rN+LxE>^DݮEڬTk1+trǴ5RHİ{qJ\}X` >+%ni3+(0m8HЭ*zAep!*)jxG:Up~gfu#x~ .2ןGRLIۘT==!TlN3ӆv%#oV}N~ˊc,_,=COU C],Ϣa!L}sy}u\0U'&2ihbvz=.ӟk ez\ƚO; -%M>AzzGvݑT58ry\wW|~3Ԟ_f&OC"msht: rF<SYi&It1!ʐDN q$0Y&Hv]9Zq=N1/u&%].]y#z18m@n1YHR=53hHT( Q(e@-#!'^AK$wTg1!H$|HBTf̋ Y@Mwq[Fī h[W,Ê=j8&d ԋU.I{7O=%iG|xqBչ̋@1+^.r%V12, _&/j"2@+ wm 4\xNtˆ;1ditQyc,m+-!sF*3b\N7dYܞLcn3rnNd8"is"1- ޑܧd[]~:'#;N(NknfV('I rcj2J1G<5 Nj̒Qh]ꍾZBn&Un' CyUM0nCj.&Oڣg\q0^Ϻ%4i" ZZG>Xr'XKc$2iσֹH<6N8HSg>uMik{Fm(W F@@{W+ߑ?X2hS4-=^YgpUHެbZ!y!ul@ڼ6ٳ*BDRw;uYGҦJk#yRʊHބDnC{Q)J. hW UT oVh*c6q?Q-Ns%טCE?Ge먠MD"+3@'V]PXu/:*̀1җZ,{Oǔ6Jy%١oBbFM=$OQYꐙ^=Zza5a%פG,ϒPV3^KPbGVO'daOU%tt!ƖRG9lhfd#]y=DFT8F}$RD<8 ].v\-v:8F+Mt|ga.!! р#ݴtӫߴ]vWͽ2]Q6Û͘`_}KnK"]p<)Xg '鸽= &Xu=y`g[#ɯO"?5Vg3gR(Җ}f`ӀSqUق0D L?U7_nMBLϸY&0Ro6Qžl+nݷ" 㬙g|ӱFB@qNx^eCSW3\ZSA !c/!b"'9k I S2=bgj쯏W?=`}H0--VV#YmKW^[?R$+ +cU )?wW@!j-gw2ŝl1!iaI%~`{Tռl>~,?5D K\gd(ZH8@x~5w.4\h(`dc)}1Kqi4~'p!;_V>&M!s}FDͳ֧0O*Vr/tdQu!4YhdqT nXeb|Ivż7>! &ĊL:}3*8&6f5 %>~R݄}WgѨ@OĹCtWai4AY!XH _pw騋[b[%/d>. !Df~;)(Oy )r#.<]]i-*ػ-f24qlT1  jL>1qY|\䛧\|r>Ch}Ϊ=jnk?p ^C8"M#Eޑ-5@f,|Ά(Շ*(XCK*"pXR[كrq IH!6=Ocnи%G"|ڔ^kПy׏<:n:!d#[7>^.hd/}ӾP'k2MؤYy/{!ca /^wT j˚ب|MLE7Ee/I lu//j8MoGqdDt^_Y\-8!ד|$@D.ݮl`p48io^.š{_f>O)J=iwwӑ؇n-i3,1׿5'odۆ3(h>1UW蚍R$W>u t_uM Wi·yT"^'~i6֬:v~m!׭dֲcUh=Ɩ9b&2} -/f;M.~dhÓ5¨LIa6PnzɗBQiG'CXt!*<0U-(qc;}*CiKe@p&Em&x!i6ٱ˭K& FCfJ9%ٕQ·BD-]R1#]TROr}S [;Zcq6xMY 6seAU9c>Xf~TTX)QӅtӚe~=WtX-sJb?U'3X7J4l+Cj%LPFxŰAVG Y%.9Vnd8? ǫjU3k%E)OD:"Ϳ%E)=}l/'O"Q_4ILAٍKK7'lWQVm0c:%UEhZ].1lcazn2ͦ_DQP/2 re%_bR~r9_7*vrv |S.Z!rV%¢EN$i^B^rX؆ z1ǡXtiK`uk&LO./!Z&p:ˏ!_B{{s1>"=b'K=}|+: :8au"N@#=Ugzy]sTv||Aec Xi.gL'—Ʃb4AUqػ< &}BIrwZ\"t%>6ES5oaPqobb,v 2w s1,jX4W->L!NUy*Gݓ KmmlTbc[O`uxOp  |T!|ik3cL_ AvG i\fs$<;uI\XAV{ˍlJsŅjЙNhwfG8>Vڇg18 O3E*dt:|X`Z)|z&V*"9U_R=Wd<)tc(߯)Y]g5>.1C( .K3g&_P9&`|8|Ldl?6o AMҪ1EzyNAtRuxyn\]q_ߍ&zk.)Eu{_rjuWݚ;*6mMq!R{QWR=oVbmyanUn.Uqsy.?W8 r[zW*8nؿ[;vmcoW]"U;gm>?Z֒Z6`!2XY]-Zcp˿˘ɲ}MV<в~!?YXV+lx)RRfb-I7p)3XɯEr^,bfbKJ'@hX><[@ ,&,]$*բk-Yv5 '1T9!(*t 0'b@񲱥-kc6VnR0h& 0Z|ђ8 CGV[4xIIWN?Yt>lf@ Vi`D~ڇŁQLLkY <ZPKoma_u` !>Z;3F\dEB n+0Z ?&s{ 6(E|<ޭLk1Yn(F!%sx]>CTl9"و5 |ݹր|/#.w0ޒx"khD?O`-9C| &8֨O8VH5uH)28 Ǿ-R9~ +#e;U6]aD6Xzqd5y n';)VKL]O@b OIAG Lmc 2;\d˽$Mu>WmCEQuabAJ;`uy-u.M>9VsWٔo RS`S#m8k;(WAXq 8@+S@+' 8U˜z+ZU;=eTtX->9U-q .AV/|\ǔ%&$]1YINJ2]:a0OWvI.O6xMY0/M$ *s5x{gsəL3{$)ՆbG(}1wt!wVf;I&Xi43غgR 6 ݩJ$)}Ta@ nS*X#r#v6*;WJ-_@q.+?DK១btMp1 1Gȩ f,M`,Lr6E} m"8_SK$_#O;V 7=xLOu-ȹ2NKLjp*: 'SasyrFrcC0 ѱ LKV:U} -:U8t[=EAV$=i[mhm"roe5jqf$i>;V0eOޞ4ccc2J1TN.7q;"sդSP) 0v3-)-ٕAg"pZ: "ka+n!e߮lɹL V3Os\ဝ+A= 2䣔AzG\ ` \vc"Kj61O Px"3Pc /' PW*3GX liWv-6W&)cX |]O;C%8@*Z1%8Gk@5^NtY"Fbi8D'+_1&1 7U^k6v읨gQ`LRx+I&s5Www` q:cdʰ H`X;"}B=-/M~C>''1R[sdJm RD3Q{)bJatdq>*Ct/GǍ-`2:u)"\**dPdvc& HwMlF@a5`+F>ΰ-q>0*s%Q)L>$ćYV\dsEGز/:ٕycZtO 2ze31cDB/eWy!A/V4cbpWaPBIpqS<(lȣ'3K?e Z?ڠ8VSZM}pnqL f2D?mzq*a[~;DY〩b𻾋-]f8dBմVs6傊zF"daeY(R+q%sor|.v\sfa:TX%;3Xl= \k>kqBbB;t@/Cԍ)Ga[ r=nl-w/38ѮI*/=2!j\FW+[3=`BZWX Zd>t*Uǖ\*Fu6Y3[yBPj|LcwaIuR;uݷ㺾|47ߍeys=.EinE% 1zY\+͕߬VͭW_겼cazyU1wOw)Ǽn@6 |lk'Z|VZpsqL5 څB}>u)^v~,󿴝} 3+m𢛲Pz_Sp2auQAP*tLnIXA6L7 8UgKdT)*7>p{Pgi-b)>U6IXabPde Ӽ8Ģ8GɄnb'G ֤Mcv4?>HC78NE@UMc8>`TvZ:}O KT1wmm#K~ٗPֱ3Ď7: 05oEgV,;vZVqJ$%WuJ8w=P"0lF})JY쫒tWW5(x8. ]] >mCstivr]8Nʄ#ak2CL8kAfuɋGT#-ǁ|^>ja#\@[g#M[q*5"A @l$dLZWjT𒧠X娚M#"z.G}r5 Q_:x4r4z=y"IͫjҎ%oIU.{#(Q c13jM-gKiM&N4є=.?* ߄y3>G_"}S\pMfz#Gݑ,m2\8AFwdЈR=g]$9j Fy,peezbx=:!9u^^]`u Na:OXΛgY @+5}[ ŰrUl^W;}۝#Ki<,K]l'o# Súl kvs{ Cx6LFىLd@#IUk{-;7ohܭ15/崐R@[ċ}X4dۛУ%A"K Wi`'Ϫ<qʰ)a ¹bă%3`I=SԷwLA4-ɒm` )M2@љ< a9*9y h`H,:J4~}jʼ)/y^Ӡ'Ou^EBjSG'q?E.>*:rҏ' OZ7 MB3gYz-z^{blvIckPoo@pzrL`cayHh ٮN\ X %^a{g$ b{pK9QMnw@8 j^K5Y#I:}] 6͛Jã)z Ev~-Ӹ><0t=)Rzunv-y_D<"hmx`zECj$y;Z (@͓\{!~3((=&{"Qh-^ÃIQ 7u'g"O'X05=I&3GUV% ULp%WMA2ic-aMHU%XC\a);MQ3M5Ĵ i*t=zqjs58!{ǑYS7,L1>!MtVHSRϾtXt"q{ʕ16]R~Sϥ[ȻHx{x޳@lHꍵT2Bx*0"W>0#RՔcϬh+-q6UwK.-* f9X}q!sD)745esjy"ZZW7 "W[fX_j^u &$5!f6 9\Vw6"[e@U=|`2,6d lLbwuДײ6@Jۄak>{$dě>$ҷ^ @XL{:G4¾¨EI4P.0m7@.FZT9K W \K*lI"uւ8w5<*A$6#lLۧ4-bz0װC #5SYK^K1{`Ҫ&-\^QH֌}UۗbJ/t1}M|[J6_;6 wW/^y"Zczc)c >ɫX0'ʐmN՞iMF;Pvm kZR[L5S b\JsL}Ã/ mukhZbǺq[\w5㾶,fvirEܳH0;0,ۉL#w |; ,+<.e_W&( 16 ,/pԖYI*ռS'*%$\Q"`*S3@Ls~ɫML_iyܮ"-aT;bW]EQl؊`Υhnt 鮵xzIDnU.FO]RDb2Z4L" u=rmY@t3/DU>9VgҒD ۵?DPeWZK_r !Ec"8f=Ja'ϚkW&[fh|6;u𫫘°_3J1p5Ui?M2f hQ ;r *3 &|7,`qvcfwC,4CpwcF0`7Fvidc"s6f&-t+} ۑgW3J! 0Jnh$,s{v=WE6a.!)C'p$IMBy+Bf|zz)/zFKa\Ky.5n{nϥx]Nt_>7?{6Bj(Q,i%E KrmIvҼb7CJ28Y;tjaoqM"gpi|*p%[6ZAvEm<3Lznϔ hL'0:Nwi C]=C%mRl_M:X~Mʰi)wZ5[u<NqU7ϧOpVGQmR;:?rnXc}V`_`v~*z]9䧷?9ߝ=??>y;9voqW3@ :Pn8-t䫽%uY\K>)Aׂ@vjJlr? 1n9=$ wNp@!wݻnP6<52;Xƒ+ ۿgA@{#^-ᾬK; b [`,|tÎ 4d[Px|Gxh%C#"iB͒;\e]D ~7|[J;X7JWFK؉H%@b'76@A]`1C{zF K!(g: 81Lnx!k)ϺᴂPaD6A ?`C#~xW!bp,l]wD#:k!w!$ ndF(u*p d@0yDZ/q38ާ@ȿ\r*{@EZF؛JZ)A~MX'QZY{'#Eb)x<= s']iq٧EWZZ}a؅[z);bmw%AJѭsv/h97HAK=}A-Lz8d#ZXV@u"o;^Gf"sR{¥ǘ(";c"3*:/XN Y y¯|?RPI;*]9J`v=΄t7-H@@&.B,`fpO9CpO@!;"H0~!3.MCI-ՙ曄ZFSޮH~Q{Y|-n/1+ <`,Lx'CHgkƹ:0C sj6D^JC|~1Iɫ,lVGN6T{GL կZV8:F$Pʬ`suˠR՜|tY Øwsupj`NPsV櫏NDRP<<da .B*A! {f|U#1`+o 3HH ,A.n*0Rj,E,FFG;QΥTnRmpP@eFRubmDI/څj.8u%}=rjMR.7+Fz0yKPQKC*4{=noAOq sa_&yͰGeRUIQ_-븰P.Ő!B+/R`&ŽOVR5|\5$:A~V`P⏢(SШ:'QEpQ*c݌uC^Ne77W5^+i=,0VÃubc@r{#Z\>(<[m 5h]u1(1e5@U3z+!t0af|K';L&,Ȉi"Ga` ]WQ $??d6 n߬5?65y]y2|(|7 +sHif MD,=/HX*i1 84}7dU˩g5Ȉ|}0lŀV&Ѿ3zw~ͦiQEސ$olw}W3 u N3s˹A=&EyI>.$OͯgQ}cT8^氻ן\+nў^ߋއEşjh}(;(|xX9Dk #CI? /E%Bz": 9>t0ʆflKبbʠ^'RaLZ%H o+4zM5¡XקSt8qU?;"uڔx3mg Oaٹf}' r؝]n=T Hvp'=܉Gi mjyc[~dNH* h( T $GIifO*]8r恠?NAH7< %|OgG'|ϔu&wI(qC/"i1q%jnX/(9-(Yy1(f;"|1n¶nM|?r Z/fәWm{Ce@e_F,Ze*yԬa+"D!_̏F n+QH5[.hFXqfVlG~!PQ+n^}JZeHFnE/f`ޝ.mEЊO2S PoTm%>&G :X=ʅTwrҺ >LT K7UPQA-3B^1hO[K([%mA(۞P4Bَm]R͘=qTٮʶjdwڧ΃身Ժ[nO4R)u [%ۂPo{BH|P|{B;ʷ _%߂P{BH*<`GB--*FؑP*>pGB--*FܑP%]9Lj,+j|vJT[}mӣ A!|8M]) &n(`^6_9I 4B*4#(gfJ픃F֥o=ׅ]ѫ3K~>qW}8/appLVMss y_:ʼ }р> {iPN'''6#9]u.|i } Xo9d0+dy[R4BpO\C]ؠ  ASح5@Sm;F͐23RͲD--xXoUc=GGXz[aq>hK^I1#KdCaa xCev:;**Y861s~pz5W:xH4Le.KOU}o=*  ik7psP4fJ$4Q:^ph ੓x+l |X2b}D>B^;@_o8:st9&F|ɔ[ECuj*s5`ғ@MaÄhXʆq#f]1XUSPK/y;y0d{x ~8aA~],Ɂ|@W`DhgH^j_ Tª;M#2n2 `ݰ~~9A}]sE!ѷdx/Q#u^j;oT#wLڎU;'ϡV/BP|uXv+d~p ~2X iWS!B6QHGo>:r3o6W{ t96j~:P<&Nu|[3w@Hb(miq;CY-QC5JƢSLp,Фvk{A|qY.5xn iV h쯜`?7>6/5IJ1:U1=7M:R88U6b)ApTkrJWP,pq*R)|M3|Cԫ=4UxDGh]lȪ?eVP73D3:֞G` \ G`$e]pCS&gTx˾#g1ͿD"U ҿ3@,ULj+Z/<>1FsPkR B3|u6($y}AڄOu! $G_W =?CWVnSCd{X 2K/Y]‘i-hGn AHs Tx/fz/[z5K]63IHlTϸKܐ| rU~QNbՍz}>:mncR{׮س]y7t{׭ahVWyg]]$̢ej[KBkP%} vp G]ES;t.&\=@Un TlzSW. n6/ץ_|B!P'2jeRiv\3MS!bƯJ388%rkKe Z-&a6؆ZV*Ӹ("6=Qs\yp!"nd+6]Zl"zjn.}Ʃb} C3_35۽Po@~PLZW刐ݒTdt> gU.YvabcCTOt@3 *Ȼ޶W!Hu , k1=@?%Z8$q)^\0lI3.{>Dǻy4wۮwk}w&{?Lt)ǿy?kѩok~8|x1`ӯz5jχ 5ߏ~S_\qOUsW򱧏^s?N_~g|1٠O}xx?<줰u?OR>NwnU6=o1==ub!8; qB,dA??雏j!/p;jz.WŮ E}bW2hgB.~=_xp^xoq=2g;1uӻzx7!:AOľ`SzjZ/dUÞ*qcU.uc!zϴ憥#KVMuJ]l~ɿih"~<:;^tQJL+'ŲT&ZYW{-#|HbGSiigYCVZSo9%-gkFՎMګJsg>yU%Y1mf Ldxp!Hbeqd9 ܢ|SeUGPEBd71CנL-'0X\툴 =;)4f+Lh%1n+c$5apn?O0&R.zg[eR$tp}} !ko#8mBC4kY eӑ)xe et1칦[FwX St-q@Mn=saD<R}XW*Q᩺;g 6qt4'ь"GX.0-Ka{/E95h#B$ʳ7˼SbMV 1/,8o轄!H H:qey^FřZ%˚EM{E(  F;TI޴ bUQd K_` C$V7O珷Bs >?C2Exdi8$kk[]Tc>%Dt)ihLdbdLb$Ak4MdktiTT ˝Ag0o&&Ќ$L b0DwE@$XO/ƕuD(2FLz ޷љ3֬l 3p4De!9ee2IUNsIvpʿ g@ ,XH})c&/R^pZ ߕ ^~;Gm-] 6hʡ|O:`J' 7}ݤw%c |Hz# +NI^H<^R'MhP{3653:xhq1bE..=+{4}BM5cƀ9H0FTpJ4Kp;'^o̴@G9.yk1yl:uucivgeK,Yx˘K#:؍5]ĢM蘗0ǭQs¹I't?$8z{- A/ʁyÜD" 1|nю`3ˤG@y3pAf)ô;pDZ>trsK1дj\hUG[ YEk *k1dIB!XWjwt(HMrΆ"94Ȍ8ɍ\OBU+BH   ;J19z3éh=e[ Ծ>v0,ctXq~ɔ> B;þ!O>P5 hUAѳfȫE )Z5@K0߈0O7MEk1>yajPx%V [>xaǭ [wWsnQjHWt088-8c퇚0PdLfy(/<5 7l̆%1rC E_jpi- V(i Yd|ݬ#ћQyD ()M g!"4T+d9 !6H雧/ejM;xvuC5MޮGTd5v8? *f+fX mL۪-dLb$~]TӨ5ughn 6PukdZ1}\`<1bfts؍+1rIITDBy},H2X1%"Oaj1־.Ф2#syTE]LI%$_=1х&wSຜEqd%I.r67GSJOK -㑚6qJpxDhV _6i\C/8.B, @T ٳ^?>(74xE|R֘ӥF_T;?ƻu}$=Y}m'䌾xHuZ곎tfu\2x|ÎWe 6cgivp-3lṆ(opXEsLQBD :eE"$਋D6MԚ)?|֢Luˉe%il*ICkC|Ps=K78r;Ϻ1V‡,Js-ض^ॾ'Z:܉Mn@ nWρ[9OS~~!q+s^&ͬ ޾d1UK vW+B s$8j(D^`T1̪Qޢrn3G<N৛t+ipt78k[y$*&ߗD\ <32#v?%!e8xkx2=" OjHzvGt#M{XG _üC˱iV8ʳk= ́V(T4`Eih5}[WjQoG{z[?nEP$\{AvY?5â_9D K svHo1G٭TPѹ3GgC̔ N Tu,=ݢfP@т ;J>)Go\8F+%Sh ¹(&;N$8R|+fHJ.3{+*{γK"0Nva[Ip2V/\2+}7}spNl←yŔ5̓THnCcDtolKz^ #EI{ ܜ\nX-S_,ta v`vXMg.=U(;SK&/))Ţhg]eXŌ4Ð#\< ^r -@BvYJBjGvZuxay/~Ď!d IyHp C!5&pf1ߟŗrkHpTfQ s[I0])E {;zL?m/גĢ6=;0k@?[P,iRr=k%A+-iK0>5M-oEjV& ":ƒ e*Js_p/Pgn nQ3Vom'ehPZ\c_=owTeUUT OR7$!rSЋwq.ԑM2mn |%O*YB ԍe!$!V;dȇtzKXݮZBbBZg 6 GN}c$ǑAg.yIm4V/ ݀q/8ٙ_k$ L^@n77$8&e9 GÀ bh3»i$hSAnH:9B`sko ќQӪ& VfkȔL";1z<-$yt_n2?W$ܸt# ࡀsٳ>|{MhcsK40ٔ 率sU'm" vQC{33wn?88p 3vnp.C^KM95nTJJM]$n9{6~o${Foo]f;2}9cHl;;BfYk3ZUvP-2+tU wcd7TB@˓1&0,VabGy[ݤؤ$t[?]uԭs]NKғvS$S 2ܑN0\{bEDy봴CT{6ɹ546G 6X(%;KQ٪49\')؋"%uk(Ze޵}fOa-5*"rt|^rW0LG4q3|y,|Ӓ/%"1 7H)fIct 4xp1^0ܷa" 5qqa>=,O/L.U}0u գtIB* p{L{ Uuhҕו;(PVOҥ_h-~;ը{r"' AcTvc4}1:Iw<#>R;a},ǧ @)t krj|xK`ؓ&-! ߃" tI9p8NdpOf.@w-w9ЁtW Wpin'Uyi9?_tUr{m-%jrW}-:Ӭ =v7mIo߰QBQ]f @Ysg!Eś=p ~13B54 eMW2=Q^$fyj?QvL[6卡xmF+Q75rW5rӝD&R Z0/̀Ii,%C8YkRVOuf,7(VI7ԾgZbN)Wf|q9h'ҞLnct>*6K8AٲpMqry[ryMxp<դ*th\.TSoMıd)EY]*N[ hryi4:RU"n߽l:A2RBW VdY 7W\_tsQqUfsR@]Y&?z[ xrw?= O}]NbZ3sʯn#x%&U-V gv2.{#*ϽA`Vpܙ~+&=.^x4|ZOY-&)yF=#Jժ J8i2&gW{9s 制8ޫcZ8fE 0o64 oAa;k8FDڐmHV@rR=[{Bhڥ vZWJE (LFacvHm\g )Dl{ɹ,ƏriT-KIK-T{֠]ڠ(-EEk ÏYMi$hU(-chqSaV^]r~q"yֺ*Ȳu}o^7Ѻ74ʬ39T%_/lBkE/Q(Ÿ;@Re\ &5nPifEKZ2.j'0 7z0b dU\N; ˌ*:&\z ./aU%XBrA7I꿅a_?}MaҪNMy/=^7e^r4ytBji`˜VJyV>ά*Yj}U+]x a< (0"rbr'>pA!Yhvg W5~'{͐IljU p b"Hޫ tp+,cu:Ʌɹ# m̪ViZߴ/.fTXT`#IMm.LBCshŲ2  ^e7yK/xxƅzZrGӎѳh%gi94;Y^^:^i0 zgzGovmvo~ }",&.z>.v|V=Of1*h^o18cR{!g -rÇk U\j V}8@?d2mZVq!Ye׿oJ_߳nю~θ 8O7fs.mө(ڟ8Kj#@v %Gn ڧKptP,_ v+-~N|]-WI yP,8( X9߇~T}@ _=kj [>O;H= <ѤEl?gaz0SumN7m<dS$CaO$d|W9F(?}4 CryO՗⍽GO+T,<o;|N~Jv vߕ|~p9ێx}Lpm^z^kc϶ va ޯwDb٦M@`۳^f$-cWxwYilw'%[wZDѨջ/+%4H Q*_b|VU34CJ|8*N Yk(@#:'wQKJq}CƁ__32;ӌ34C+9M@c@8J"F m G^Ghq6JkkFnGYGNKNu{7^^jq8KvUGshFrʂa( r0m(0~k/܎ln!93*hfc,M;z3*,#Xg]>X򲚯O/޾JS\MlF]$^Gෂj6$x0( :P)aq8o=n-AM(W!b~+X~(V "9"@i_//+3*=# ̻v)9MC-Ҹh>@{h 08YsDG'et&quxU&qқ f}oC-%XTuLW W^SA/o#^DIU=&*or1<_=Gп'v:Ǫ׵+JhtZɎXP|槗?~m8w% xuvpg~s ÂOVkJ¢Hi| `mC4\ LbAs+_$sr/nd4ݖ&w~ n݉ibÞc|I \bYg^q6b$o={08P3Z}+0rEU$ V%XIqTXZ2F!RqƬ18l41 pe--m^d٠П$3"qEdL4ǘƘhL؊1Yk9'a R[Qyꕐc?rvpG.gDKQeBrQW !HVn !=O@Zu>soXJav`K1&V_H,W\xm)1Ih syt <ZY2VUKN"I0J R- LP0g ֋ )i(Q&aKNML}llkblݵ-vwyyaւfR+&f 94LĒ~VaѪ nSK3Ί~Sዞ6In v\j'?@Q>,vC?,7Z 3~ٰ,Wn3^3ڻn7]J4%^͓=bD"ަHi ਾkmtܱMR7_={=臃+P`k܀j"=LXaп8EQ"H$0' X{bdEI5c炯93z̚pv+I[} s~o9# $Lt)>EVƽo9l0np㍘=VKF Y¢BTl#]0(p̨W>A3N=\ BSi!LT,J͐jOUJm2 1Qc3M8(,'Rd5>,ɳ&m@bp`F#Ie*=t5m$B?&PC I l|4m]d#JvAV7)%4%Qql>Uտꪮ;4<UNE9$FpĢP& (;WBA?Dٓfa\ٴ`@>+YҜ a`ۛ9=Is>LL/cY~qQpahaG9 hPo '@%G`)",#M8(` PHbIpl"Hf]J+N%nn}O87g'+3d`)&uCo2-\A`r2T# D#j,vUT#W5 *Q8/i쬤A4YIMrF ;+i̤As [P* &'oL %Zr8#EZdIRd*̟SÁw9Z)KL=|Z$DHNaSp F *,il R112 $ yoZW y>]#%CxKF&c#"e EJTZ#CsP""0`@ *yFR)97:FHbb EBLLiBb42JRdUJDC$"cBdȩ1 c2H%x0 ƊК]uvU$*AHDI(Fb-]oa :Lm- a92&aiaϵRW nW2Avj7~PM)OYc^z.M0^V4Pj5 B@JKF XrlV Ԫ.ᨘ`ôO@,$M O#Xv2X71~: @j ZJڸڭD"MGW#齸 : jv Jv09MB X@l FbINx}-b[OzξB?KϮòEZ55D0 ׊H.i68%Šo|θ`8MNMyBwoVk ڢ!~%2qK[m(&=܊ ܬF!%%Ql޽},I4AxF)ϷR|:.F5T{j9>BFqVEYå+uvS9VpLڑf[%2oCYӌ.L TmPSqڱ+r6jǒɪ \ D%ʖy;b(fꥳ`rTPWX}ɺ-~hhYqnQE u" H%Zn(\%2ͫj]ɕ&B qH<+sOx' Vx \ R5:yC)ޱ ~6zǒ{>z (G{ ;$ܲՌ8V;# KU)rp^?9Q>#]L% mRƛ:RLdj7S+ g׶1ʷk-e&ޢJoo{=_-$ z\7+f=A=['S8{W`ww|7ׁpÕ+O;AKF{ M986P+ZzNU?GwW,Eel@o^C|؁`Lݧ7 ӌK^kL `b9I7^a=nQ=(b3RT^hdeXN[ΆlȎAU)sZ`ClgC>6GiAtWx~ fɩZ~{^Dr6vY3~|.qCP:^Ȍ`塱ԓWk ߪ%)źDo]ظ T/XXGS{@D{ЏH31޾X,&$\)r3e#zQyEUa*QQSgUXƭJҴWt roAlVr%c{ \=9E `G-1Bb7m<d@5yQmxc)KlsHENV< H̱ޚ=ߜ+[ hяx*K*'&h:h`>U`RP׳!|)v7[wmF\$bܛôwkmߞ R\+7˜Xe_~q{+{=鴗;7`4Vw8O 6C({Q~>pվw~NMO3 6S_L~R3~/ A@MG ZfRwЊt$XX;h[TF_4ݲlZj%ݫuHTM E`7ɐx.r}.0cB3!LOj|f]hRs*4u`z&C}dc B`UNb,/lܺyL/So3C2fױeSFw({VW` 8QZ>N)W"ГZ`O4k,տk 9 WwuGu0]A[kvHֳgneY]tC7Mi.,9$~٘'0lw3(GoVD%2T#oFIKyg8 Zqb Io]{k2J%gLvyC^Ȋciu/|s~z<ƨ&Տ?U]wz޽yǮp2KtoFB|;#18kQ2 *.8ua3 q_| XztVj䅙_cM`zxտ|36\0BbfbSp/~gUlYͼZSc<4A저xX1|3źų`p{i0yYdOv&ju #sO?=^jT|Hgyf^gl[6sc)C S/2=#"v XzmFy .%VajO~~獓{dv2;(?徇FWdtoEnzRDfA*B70'^{کcq~B E%g~Tz\8AyeB.FI R,Ai$qDt HrWIfTXL0:6C[E-. f$`qВ<1Lg`8!5,#4@G\\g5tcWx"Ͷf(Fi=& ]6#l6 @\`MZ-i1L>Giz'ؓ׌n R֣\Bw_?֣Pu\5y3r/WKQRHI(RA{WoI"JeuGkkb>~8j+A?OϬeKˡ9$ǣx`%{{jzwm|3>|?vgR+_x5un{5 >t+d80H'(uwE3s\lewы{)؋w(>|ֳC8k~OOg8~Cf"#b,dcQ3E\9sYbѵúȷ5e@U 0VipZ%bC,cCP ~1^,Yܰ#f7) f f@)L{7)Lh6nyƳaJȭn.Mdm[0bJ0ɽշ9xldw}66Ʒ[@LYRWLίL ݢ=y~vya3mɜ,7٢] 6vkH&ե9A9s?O()o+hut2K4xx1HՓk|Nz*PRZLJQcRj]s o/Fx&xFĸc^Ѷy{a"}-AAKʣ +\v:Yݣ sln0ǀ9iՂA]0Qp893LT\l[ )V(@B-jP Z 4gek1Qk}`-i}`"J ^xVtL+$L@ Hg>CA'6F|>89%KIvS)1X(.D bLQ iNM3j G1e%\[q.ys}ep7/8?א37_ywyRyxb_Nlѯe~km14@˃b%7bia%3 n [59`(ՁQyh+aD0#"b'RHDEivpm+;2~]B9{OA.m&ѷ$<"/m^[a8?<۝f2ׁR>cQ ƙFRLZHcEDc p[HV:qXVdwЖ@GQyf} T۫!| Pm3T#DAR $ak! ct .0I(K! 4641 $%4VL#,8ZI-/q?@8v}MZy0|kk[ȴFw[X[$1& G㋬E_ , 7]܎EG2Itk|UdGg0͕^2}a`"Z.x?ߟSղ˶Ǒ$F nDQ(:2 "0'f("IbxcTaEǛu:@Sł".Ȋvo%d%z[Z\r~^烙ؤ!po1c VjH8'P5*`%XZW}h9{s[&? "- &^h/.6BxMh>r k/M&^hQOxM&^H6Bt m6BxM&^hvxMЪ61M1Imk/ m6Bx]Ko#9+>mns`. vdUe$U7RtR"fS7^,$9x~^9x*M5yV |?վ[ɁG Fvpw?}6˧jS?:mOu,h16g A$jT850vo3O@Z{VHUN7UV9?nXEk(I^H"`H&Hu$CBc@ǑO]bկl3WƃHPDoI/&[qã9|*F|gmCKs(N4$ȤHJ;ɜV<##E=PqSAuF&P̦T' Vd"7rZ eQ1w;X( VNU],{sG*U *8DZÉe)D` hĺFOB IJDs"3M\6]gBsG1{@ȧ]fclK{B;C36 wc;,`Rʼn[K&zi ư衍#+Xl_g'prQKFF{)@,(ˊ|nl}e$H-;(*ɀjC[m( 3ݦگb?<.1NU_kGȬٖAhڂI^o&ԳnC~nw-gWpzE$1`*H#R"h˜F:N,#*,LIsHm?"U8/}"Y&E'[C>T劁#BAmH4ypD#>8>Dx!ŬA# ~pnrqʞ? $ lO,Ti{t9p@RۧVo?톦\@?}s"/r/.Ege)\m!)_181N b]'gsF8D~ƽfmo7?mo"%q 5 s+0Jz\bőOD0]i``Vmn*gr+p¥JYi8rޔPI*H\MTI0yV隸E>jUq}T*{¶|jxZqXb&ژHd5|m7 c 2cx ǑOn]}{:2 s<9|*Δ!LJ ٤ Q7o4,$BXҎq?ŒY݃G>f[x[?.6)>Us 9vQ~kcDj%dVB<wQS!iīԥ(#H ,3.G mܒ3ōXݷQSaDX 0TԆ3f3+fI^C>Ԃ!cG4(Ӿ B(emS %4$ _G/0NC T\(f8!ЊZQCց^#։2hr.'BMQs2I^ZKM,\Og̈́i&f:INg/xTYgtv?L3uty DWyȧ^-KS3dȜR w^PCT9,CoOӆD])t,:Gg|GeD9pxpSfgN(N!8j0 GHlzB4"QGRZ2NӾF _|a= Ղf6oԜfbYT^dHB"QK҂W4)a+qM[chLԔYׇE>ڂYtEoZȌ@ YT+)qZJfPj{>(;]S K+꒓\`45J$tC1,~+|"Z|6bkaoqp3w47" aCn2!7!74I%N׼0/@%u}Fש(1VE?QQ,T|1vn5?%> A \Jb*fiJGOZ,a䵱m.H& 9 䒀.`ȧmMU v0 Zǝ6z m|6 -r}3hShAc!FsC2IF̈ $)6& נ,v/¡/5#_̷ojЩSm͘˙DR)9F0_"9c\nGe4|u*I``e8\fd@F xTȹ&Wm*ILk X~&Λ󑷘E}PQSA =?Ǩ يd"ʗCUn肁!O2*Ϸᎏp#ޫ$%8l9%HhLq ڀ%lɀ]G(sYK60wb2TW11ULXd@FQˢ/ Ȩ`WnfDe6*= p*98]Uqc*X8 ZSTJC|s׾&~Z+ݨ(qԌM01dT< K^xx+^邕chQLT]kkKDc)mLZzB,W=ȧYpfqiG?@U4y㢳F\|XW7h:sG<'e\k}mqWRk̈́YG;(򩐛*hpK)yUsvvT#,5{`)kqMOcwP.x(235]L赣 fokL43.C6 qã9|*F2Q]F(cmKߤUEGĚ!U%7Y{U)6 0 `΁/m"}faY#=qS!&•7!VϤ֋8f*"LEPO/JS1S"SȧB\#8l҈AE1o< ^y~BAy2j)+y|X.݁-AvsߥXqL|{Ƕж]߲c3GtNȨs Sﳠ[ū/~|G5/d貫qP԰Rs:ΐQ;2Yѽz ⽆YW( :cgZ?qR+" ˠ" BD0 >C>rYn;oc=Pg_yVd+K*.=%UTl_X7Qe=jhDs,j7=oVj^0*e5$F7`]?NRU'e=U2<^D5mk(Q\`d}e EBTP V*v'p ۚMd^pDZ'._.^}MPqȔ,qpp4o<3rG. Xȧ>GP1sdTrcoc--yag@FhUpծ%<l0Ë]yÎ] lhwVzE0=eAMKA\yTj~91g kw+GƟv pJ\o~5+?~ݤGXU6R|7x Y5>SyTxOֱ"C̤mLv]jmFs^R"UAnITgW@PPGaz6#L_bš |j~YȜuFe x|HՉܮoDФ-/Cu(9E sYx 쾲ܯ U~|!nzAFFYS /^Ai |×:u-sA N,*FI*Tx| cBм ;ζT&؅@{*|WA+f˗fT C?aLzp\諿g뇰x|~X4K!wva}VQ +e2{w <0`~l]?m?y?1~Cی܍/k_i®U|[, ko~t?x>_7hzYx ]|ɻ!z/;Gɂ\?Fas`TWvy?놡J)"?o<{޼%ʞfMX!6û/ϭrq`I^:A$8<ީ}i_:1nF G?]_aЏR!Y/FY zcskgq˒^2D fL8KZoiiE}y7[تYg,"Lϖl|~7>kg*aŲr͗Ed=EVxl-Ge'oͮwsm^we|6u+gC~>O6W5 [ʍ_7_V7 omؽ_=wy:f'tc? Ma/<,fwYDy;_uU *意C{9A Y.- ʻw/|6FKVvunj ;OS(!e/UtU۷+O0UaS_>T_pDgUq='}"?y2hbadٳXޮ <)({b;ĵ$(ǎ-%g,'UW0=?Q҉m dܭf #Ux_w|C֌37;@pɥjB!ql- qV97Vg@F;43V,`S1^qb=a: X[@ă`uCE{5.js$a;W2%"A>ؿ|6[}7 Fz<>jAh]GT?gm?{zjwmqW1:&NGq}~?Ba]$+:[yWX5Z< *^DklA,U>~W1";ywì(ʹ~:D_c_C3\3:`nhlT+:;Mo.8T³wٞlU;Sz׾ʾ۬i.s٣_e1^:;`2Y?Z,V# ~Mh(܋?D6/ cBG &aؙV*ms8ӱGycgwlm&g dH*>SC; (@J*tQbu?J)pϩ+ُLՙ;ʬQ_E2%a^7 M~KzC5:CB 1+qPguw_/q 1Rza)A3w>bB>(8(cET>0Z(T`qb)1 n,Dv8tB|j+Lw: >2n@H fE<HdES4 &kZ{&wLGb MuFu_|(6/1e¢IuwrO6wbb'.8ϑ "S 2s6f8V1mCBqR&?Նۢ'SN^#1 ex<]QTS>R'/qXQvZBS@ ~IWsLIx)T=)e=-^Xu/2b粁O$h@0DC}ڍ΢<ZLBu1"f%-,Q2K%R B Yj1Y0EƖ)]ڸ$kϑ1ͯwQ oc J(IDq% #E8BZj_-9!ăQtQLb1j/-8pYlxR:gX)uXrztogůН]Zph {o^~APo_ F15T12 '@s5;~:S ƴ=cA ЅUQbY7Īqk .3W}l~=-ܳcR$\ɵ.Ɛ`1)).Zm~],h#8*>N絔L|JWmGʰ.Bib6˻"BGMM3ss~ژm8=ٯ^lK}'[n$GWz7֘}LQQ"% E C5A: O< zc4#mf 0஦1U79Q&bU)Ds/=A zQ-G1X1 GYB G'mY3A)YUHC!W4*~uljY772Z h-~ ǷtGcjChAG[d~o 0O0 ͏ rCYGfWv۱beůjv.Cw78l<&\Z'&v_.~{ œ[{Og!`O͢`kcTA 5|qth;&%V-UT%81 o*e̞O܇ۜI<{i۟Vr| ,4*嘄*.8\笖7֜o_ea8Z<VPh5^pon.uamqǹ.48<-̯a/ǹK|Ybi'/(7T"{c :?&͊J^~l8C`hƼkF1]<H[p(oiOo}g<::N2]IC2#bKsFDĨ`h\jSkkX0ЄrKhE6#m*?4zHF>֝ :R>-熧?=ՄS;Vwo-9&YdQ9rn˳`qL)]2qtd+sǘ<u8mh|[ nEESǁ!w,UfN 6TiK0ϕ\X҃S$r)D <{m_&+跛vn 40{?t Ikz8i%JJ尣ymTT]J߆i+Kc2{شMȤ% IW)]eXGdd+Mymy 0y9!dRŲEd 0k^#/-?0^e6kAhF0N zP6h@F}],$b#uυ;( ω4%T(x]=h=1`o.$=5`fGbܲ\6VUfrö/nu$!nIq<de'3<ʳSRQR$/M%uWD?glLZ9?,yF)X(R֜Ti]EIUUÄJTd_9s!CPJ tԼnj:/=I9wlx挽G->ŷ*x%*u$OA*bW*) tV__a[|ŁydcMmg4 pW 2GϝOgzWd񻟝>].,t'qnnFt_8UF>?2~y:@7 d1r|$ɗxq]펿^EwǻO^y!Dh:# US|"Ґ<ۤ-b$ma0m['ţK8[ )od Nw}:SC_N-E,xz\ssX*LH&gJzXOul"Z6b̄EJNf &8Υ,IbB?WX.9ÉG ]۴c9 6`$%ã!*"yj6DД1QqY?X;I Ӕ,+8blPFDҠ(*.ldZ#zSPo[P.4 :KmԆL'RρCLx̅zBvӃ}{ӃƼ@]V&i+0`oK"i+4`Ϋn imF-1t`WtT%4*F|:I$t&! M2Lj^)2qtJ|2p0RF`k4ACrnbɌ1IQ|ak0|B%u;PR:FtM?3mhvظ-FF"H )VmQĭw-Sg<夣10Q{E',ɄbJ 7VOS)n ɭ-jθo=ݫom7|M 0]Mc[T2nn0NƋð Y &aWrjK6z^ed}M~6? |{BqNɌ)[ |)X .uU Ѥk-e9Wzi~r͠nhSj6XJ Z~"pl'gLI=://M,<-mCyF*ʭbgڡ3벪̊?s8NMY3XbR Mo\Aʧ*'rJ*7w b;I@ݭ9r>+Ƨߢͦ_00cn^ :rɱO_^T. #ZT5*V1@N8\^M9(&8 |SiU_M~]<'lt g,g$ ˲Ɋ1^*y7lT3VY?14(#M󨈲iJf/[OGıW}E/ij6KS={ll2eVdt0zdQX|jx2p-8$"p2RS>AɜL͠=/ \MNh2h8{˥ɕΕos̕\.һ $S$r)D\#}8z&Nc)r Fo(| ˒*)=َ9|s,Tq@U*=gڥ-['uZ=DM T =G.a5)V\_0J䵣Gv`1T 7*Π}Szd9z|OWn2t7]C{&N֥x_Rw E3+ۢd>XUAnUsˏcya-qq\asklYE"CqȐV:Q8B 0M<\nm{p)xgVv|2+\ʁ D1m# \$7pߔp-C՞BS X*TQ` u'nM0z[,8]cYZdpFSsaCr,A@-g1[8厵{&._ 1/$${BXѦ"yMPx(U8&Π粺8z&N7oRmכK0Ջ",m.73]F.j;3L *&$7m8gjup;zAez>K}\ԥ }te귱۵X  ;~D*r֚*4O3(_IK\IrG%5Z'zV"rq:xv"R]W3w cd-:k\ܵ!,=z݇B-l+Nѽ^h/҆E &O3`ibؚf]*C lʋ`Qr/|2B/Y0yKTjE]vRC#RF%dʕY*B tGk#^נp'pL[*)K|0A.?ApAX& xUEl _wW$_Op|rt_ .uRpI `Z~t}<ߏUe僊CM W>,M5ZC=:g(Dž,>0X#+13:x춺:jQDJLiL@24r7;Lz%5e<<1vhTҒ VV ~zZ.vl׶vX+gmX7w9Iz-I]ldI)N~gHmI3G܁ b7d%8"&((MN3˔ߑ4ƝU)5A+4%#GdT;`'cQ7q~9-;gԇ齘Zuf ol%66j+e Ohl a*N}JeS-%27zskk cK9|֠cECQRAӔ/SvY ;7*Uۦ%=k/r:!}(hPNlEI ~az8\g-=#[0dv^ॴ$gtP ,u n9䅸( T_~jz\K"H@Z#) Y HȜؚ9_9HkqC`wr9],l}#( A%6m7YV٭[=~x顁edЩ 0zFNջ\%UĄuUʌ${9K$g>}.[>td`A; AM` 7&zejhC`u;8 wA̮-k $oTH3w_=w|ԢKASyC -e1Q3ļ2IaM'g$B0S 8؋ J;Ø,`:|-\a,\Bt5oAAA=#by P priL|TwgCr]H 57pb Sdqsj2.f$Gi3X~Qnz۴$9|㳠X@rUl%EOWvx0zO[bǗB_xx;ܠ~5! FUnJ|-j>stCp{#>R"i[b49pb|=#~*EvJ8>>6v !YDj$ФolIq`}E]gPGA^}m8E̖PKxg)U{_Lvءu4._F:6,sm>v,mriarTX$Y}{H0wg /h!C`J:ca\!-vQ,Y3 K؀}7$B  bVegtP7Փ-DiH{C2y >sB!W3o %@u%Q K 0E`?vSԺlҪЁ;ɟ,I r(}C+ڬ 77~3!.Ås 0zFNljq6Y&@g1NE_^s,rۦO"<~UD&SKuL5M9%~}"C Y$)8 =Z?i:6iTe 3u%İi7W/G$2h=ݝcU(,v:Po5k5U9^vDui."wXx1*zyⴘl?o(7ވ8g yƜg)%L`D)nE> N6фg X3 '6Q h`D(DRWn\XŸb/,˫c#e4(Et(q岈@ (6r4 o fT^9p]n?'Xղb.Sf,Q& 'c|3?̟o}V?ߺNd 1E X 2EI੢Jh`r?A~(CbB_:hc,;SZ0d솳T);ԩԦ6u g(CJ`Ĥ2 | `P@Qǡ %Yƒ`osO⼆h(aRUZl' }/ͽJY!rb?w^|=V,y''h\2Zt gY6EpJ"׏[-*{s"ؓ?V$[a_D艣2VWOyeqf̫EID1"%_Wbu`-]/xp?7. qTcweLϟ|ŷ7&_x px5絽S#eK!FGG4!L1I&[&{gQ~(NQao a&xcL,7+WF8 |) "zjT1(VN&e5_ ZL6,c7}\AQ .凨[ZЏK!CB`%JAn wyْ@"Hܫw#Mc;GbFUYl$;*}c ㎃ g>EaL$!$&!0Mlϧ {sS7ޝ;_sNIW]f~<Ծz齅JXk %[p.~kX|z#jb~Ft<< |<@@JʇɤE'M ŔR03Emά!%q$LX(T TCuz~{BHpNT=S%POZo[VFÙpEjʵөj&z' ?M\)Gv_<_O^l)1cT۝y-fQ봝S0c:@H|nMRT<<'>] |IyrkЃp'ebjp"C%q0,Kei&3Ju`8L'z@}[g'Q(VS',ŋ=yie} ۄ$}&oH3BAPd\rFMFšST+dyr1b)MeA@IlVˢqڐ~$'> 6Iq= 6H ^p4اdhz}[!RY8+N(i$©&?G4׊P*WRk EƂ?N,#)IL%6QBifL) ڵj B1?PrABZqJMn{el-y`ons Ճzŋ]ļJD_W:WW>c>GX,qF 2ܟȻ <ʅ8E痮>]N&7/+"?]m@;?`ݼUbF֓=̈ft f/Pm}J_Nu|p>nvꛐ X~o|d̴w,u9Y0puQM˔R$Dqfa'S[푧¿~XIqm:%RiYAf7g]Xe|s5Ud ~ JeqF4]lg_H_mUWcV VzzMQ/An1 YR %&JK&2 )2ǔ&8K,A 'xL^u?icqǚzW$‘$43 =\ofT%0S IKL-?V{띑MϋzYCȟ G(DSgYK_!N9F\oo? MCF:t;oFꖑ*[@Oޟ|(!N YsdXGR4UgqC)J<05XK%vJls< Ἲc#3jG5<smܳqoWFVN.r^J&3'c,#@ j,\m%ޏl#࢑3.S oJpfKz0S ɪ Q'v#Z NGđSDN<p(A^6@N3̨kK5Xs"AI.mkoo9[o0U|ήyx1nHD^l>{8ǣr>_}Pz):%Jmp/AO\2T7f*|nitz<X~^rY^/fLXcG:fb[vl + -r'n!)e FF0ubyLktVɧ*YktrAt_cS*AC, jZ.B-Y.܆G<`la1/6FG~ՠ% ꛍ,nG&_9~߽ͫ?߽<^yw<{Wg{ Ll4NχM@0{UW4,*+UuGTp+g67[[1/u^ߌr>qAL֬d{Wux@b40 Wvۯ@?nHpDR#r(JqQ;s{~ 8 t<*L_QKI닽NhPdGx~lU޲fF"X#̼ZI{U•$E%M St&RR$'K*3MAAr"+^xTn0:.ZNy| F4Ϡ~FXt0KU<~q$%Lv0jZ=A!RY3amK7Lpm!I*`[I:%ihetӍÍl i0TcuVӸ)a8KYq|ʍ469׆$m3Kw`\]TK;Ǎ4ᣱ&B-O[1 B JO5%_ n11(׫P<⾺P6\>]Ke$BLo$W |_ڼ ? UvfvK~t]h؎ٓ[!/кGVI~.+ E,O>H,z焈x|-zրBXW :207u}ͮIB.o }Q\J8.pR$>.RPUR !#zj1dk?J^2Q;¨88w}06u6ڻv7 Gҷ {3q_"ށ0&;#{Xf~]F3|Mgsi|#'[Qr@ޝla\Q6Okb`AAյU_wJ#$|w/8Ok<]ˎ *={T7e+jR ÀR+(+_,HZ\N%4XRaiΤ3k0R8+.GuW05`7CQYV#7\50 \D6?e[x<hOe=׏|:&}P,?rMv*^e?/h}+>zNEyJ_&\TTHՍuc+}P]RBKkR2D'RȣĽ"&8e (-TZy+dqu=H2wxmE^qM7wΪ?el@#;cVDHDHDHDX4i n$@" $@" $@" $@" $@" $@" $@" $@" $@" $@" $@" _:@$t$@" ,DHD@k!šzXSaÚzXS!.J*m,C(3_wܦ V;u#|lD%y` p|\ӄ`![ D-ljtV}|{u\PaP^C>tTU6%4%, -|iuVYdrBO)UrJsS]_Cޝ]#uP٣~#ޑ{^Fe^Fe^Fe^~X@Osʑ{1Zo˒͍.Q:=>Yx P%Qe 9M^ZRز1lowzcX&hC6bd#Ft:2Pt NNNNB=͈etWʴI<ggc@u'>?{4QOfg|5¢\g#"[a<\xe<^=|dxy~4'ޱo:'kd@d@d@drԁ:p#S2u S2u S2u S2u S2u S2u S2u S2u S2u S2u S2u`hBrFE|@dԁLԁHkHK(m]Ceװ]òkXv ˮaٵ\b5|Xv ˮa5,eװ):ˏ kѻHXvT3 kބ&$EZ$EZ$EZ$EZonM1p;[L S2u S2u S2u Sǃz)ܐcv2e.ZaRw# e.c,= ,f>T_b=t*&;k)Ee:Œ>hF\te@J$uΐ.*- KCh 4@=1BIA/Lіm{uzެ>c;Yb`sJ1ڗS%HmڐebC;# Aܝ 9`YFxi _h$"%ʝJ#)\oB\9HB"Njhd Dg0]8_"1ҹҲ])#[BY i#VhKGD%ou I:KQ1{= (%yJ}Em͋69@@,_Ld!Ld!Ld!T@*d!R,B,B,B,B,BAL,Z Y2Xfˬcu,e16ˬ%-Xfˬcu,e:͏5s25sf9X3k`͜CB=$OP3J2+3 9'BNac^˨)T!G 9UYrcltywf-ùWET Yc& 7.9IAåJ*#4uOb0 8Oګ~2$~;I'ߏ,ŀs7@E G8l%u!hʅa\"6PIvh:2"ZP/M%JLkAו%PDi\K!IqOP:wXK .v62yЖpBp2aIǍJ6YUP2K)e Š?ZAwV,RV:0KI2ԑҗdDYW%8k4`a$v:zO%N@6$Ƭ4.2xo)-9<Z1rCU\2)WvWz, U# qgl>}Go֚T\WUf 8EG^;7S1g *SP?X[8oKd}AK (0#=#ߵZ׍ԟ|'pvӉL?kIN*Bw攗:]ɯSUNoJI>|b`>VVsqKzZ'Kh,Gq74z *C 'ށH pMqaLӧlb+S/'U<=x Ŷ8uzn]풑 ?O*C1l䞤AȺ4_ے[:] [ߌ%kY,/4V`,Ňϳ/'gVlZ];Vh[g@ܰpǩE6ha_W_**E蟽\Go?yӷz껷O|Ky{/{ +I  ܅_vڡW7hڶ47kѠiEt›+v|%ۭ~qA| gً[y~i!5A#gA,+m8槕=aW^{4VWao#f y>DKJ% &:)f*,Cvׯ?ӏ{EG"!fCw\i{<ݱTx:>ZvcM됃mxkbl^XJ/n"IegųI|s n1 uRj_k9cJŻϯ,df)B":_q8!`YZG׮Z=^^y6\FE= ndO,I|'Kg6+9L)^:.NѮT(É)T#\>l[T^qATt_%7zp&Gv~cqS^L@=W:>4T|Q>`^GFUxKqݠ¢_> a/_:0I0T7RτsUWj,g6)?)Lǣk.smlsAZK#?m3_ ]mݙ*WݙRiwFLgK ݐ71'^ݓw+f݇ .rHm 3mm ݝ߻l[Pa]B%,ΞռT)LSEi5@i㒆r\9iĻI99YNk'-]mLlvM\5m^5rm`WUZ6̧Z+w[i5up Dp,}*Y1:Ń.>oovwIX[lbw%Mhb]™AeKkx76f&[1`M0؃=u8R/SHK"p*Yo!SKlg{g.{%٣8FGZΌޓTl?e:n&%4cKAzݑ:Iӎ ~KQU븯4Pb gCvY^Ε ~n<߹qϼ3gO^9$fj.+K qekel ]1"oC>0[F^l%= nPZ;ђ书~!Iygej-jRCט(?-a`h`'>JuQYںXجnns$ pCn T;'"! 3$/ _=@3VFP_h݄ϑ0|8YW\2W[(*9tm L ~F]VʊRAY LP&ϐ0|4I1bNJBM^YYtU{41ϑ0ϑ0|0GB(4Z_x9(̔@. 3$o1vl BPaNU-ƜrA :ɸYc!j&Fs$FǍʔڪ2uc2[/B/Α0v Q%D IiƺM 6?GΌ&jٓB킊CԒXH[Peee*}ܒQz{WZ)>O0|GSPm2.:U Z[)F/KT [֊6ؓvUS6?CFW5 /FyPAv@z"\:Gv4q({RBQXcR2-sUԭLs$alB gf{M,K⒐?CގNLBʘz:[(#wlĄ7GNYq[_U< ;͡eȊtmCv}}ې/~;[WjW_Sк*2G ӈ ,.Ec{V t7"` X^*=CYH2'<[} /G 1TKO~9^7z?ѼVF͍ OR3 |d-b!HYcFs`N;b vAHLW e&Sw7X-v{z 箱 bJ}>Qة=h :#ΝcajBUhY{3 #CpCy^VE Q^Rr[?C@㳊2n{U#НVy"9G߭lQ;E;;989J}%`Y} NBoYO<.DzjqyGu._KR|JR8z *Q檈Ns$ qPde%u7^EFAx}OK4|Κk 21e7%A 3$߯ Bij^)(2A:҉Ds'<_LF)Z%bS [b4)3$ OZM-U䊇( .a^"yLIgS6J7}.}|Dn$Bu.Wݜˡ ha] C$LA mfW*O/ NjY䵻wR`#@o-?^r_E?o_wϧ |?נר{RN\J❈_rHrɻ}ͤ^1Osʪ1ԋSpH>:o9VfJPEBS{DVBp CG;zc`ɸeB\yl,RP`BH|Fh,tͩHDi^m,x(F9~t(/8ʶf*^_j.a"?GF཈͹*, uBnBm9͑0G/U&V9+J-q>4MȖ0uO4iV,U&T˹PȂej C/ؔ\Q=KiCl`uq'#axz42V:3*j[i"|?G h-(mбkD\lC)_x'fbǮXI*GC\ny4e>jO\wَ'v AM kQ{Z Q0o!yyܦ9#a54>O [^WULM[T)ŝD( MNa;7ʜ$Zl2_B=יKFޣA=/iSv^<7,݊޸pڔjjXJȐ9+FRs;ǚt/e!fP*Ve {H#a/!sug7uzrxP>^`8AZeg\nao+Y' 69"M(䖩g]kf.ӏBkiX۲O,+v)0V:2.H`Fh3I?eLu,q1| ZʰL c))j,oX-R@L]98YF&VrIz="wXMi r P/ݶK%}a\FeK p6L%j2ZJ=[SsBMj_udc::mIGs2ͦ蠅MnFCI۞Ǟ?33.|{ [[{P=zq{쁧=){pE֟Q9qz\>-9\$GΖjlg9[CV B(j);>z>8*8&]۪XnPHB`bEb]B')+Vn 6[ ll:nո^ikҍ:YC_ii[W7.]<+=@N7 ͯVhPU5*V׳_̍>ծǼ.]介M1b U(X2%N2dw0بg*#c:.7.]qBwn<Wɟyq: EppXQ GǾ/pHy6{z]֊咏ǽLG{|{?kN(ZJ̐֜9](:֯[MwGynEۮA||*fz*E'Ung8MIdհL/^(oRzjt˽2SϠI_v {Y>؄Yt=(Bu[ꍗ_:13bWjq~=8nޯWZhq|rӇUQm-%tt6s]Y^V>>YOSj.pjݭ^VҮٴްY?,yq}؇=X6_Yɫvf#oݱ<⃲ܜk&W\Z0ie Sզ {WǍO@&K2.`wr˂/EG,i%gqHdOkdeN#HL<$H_tW_Kب`}ՖꟴQ}쾂v?888yrƯ$k}>N z<>:dţ3O__xo?o|wz^g<_}5 ƕp'hۅta#`f5']a+n?~| #j={y·@ps3{s^>$i$ӳY:݂/KRk,KRkl\j ϥ`5X5Xj Z`5Xj Z`5Xj Z`5Xj Z`5Xj Z`5Xj p.,KRk,KRk AZɈ}u>5R J>ѹӱܴDUMeVqy]_a@Y{U Xu:u_H z$H%>A?[OWoO~GJig٫dBZf*UEmRRo˃n]`u/44fw7: +hQ[h9j"-^ n7ŹYlCmvlsXGb¨mX&mջf^Mk|'GOd[>n xP{pj`kqY7SD@s!ݛRZcgm3qt-JX3Q*ԃ}`yg۬wO=nWoٱ9k 3ѡ<tۅ<!C]jjnfa5a5sYl}fF:9W/Y%(ɆN=qub%T+%_C4 fg700 =%&__2w`w׍+S>}~tR󁠋?ۓ} X{x+Ne]o uler3+Q_INiEdxU4U Ĕź-s֚<=`׏,nQn:c@ n\W |frvO IC 丠GBG'vΝa[S=VVd_edkI#'އ?>ǒw5$8.BMj*jq)R#m3gz2YSMq Kn!GQ ͚$a/I%ڜ&nwQ|t_-d+ȷ[cS_c+Bw X ;q=ѱ8dEƮ.e*,t<*xAW S7$i/Zp`\T52҈9!_RRצ'[ FPdx(r]Hq%86|LfKiB6";M==|3/euٻele:Qqᚸ`Obt=Uhu|*q!XV/ ܕ@J- YuǮ$eja}\y/8kȒz *dB(U_QStH ER-q⾂tof!yC<%bkןJڮ."/Pq^J䣃V>Mxٺ P[x{4ry>bn3:|҆uch<[g~'ʧ?}|W}}t^_=nKD7'z?qFINi3}}6[E伇(d(4'vȊswjfӛ~U9U峉|ʗ0a :tAYga bĹkMؐgߜ.|&\`0?)q׶%[{7QZ'Wq޿׋75n|+/huquo7goNO7_?)}9rQ'c(=vָɥm8S_=c8F$_ñD{#fjMܻ(S>WyAdyeyx|:=;:ʕcF ӕyj+8ҟO|(:tw''7_)S3f\vwdpjׄ2pmf/)^Զ%ah,b K)oV}L4Y͌ۧo?3S{!Wc8}~fGwڐ|l+6MQ[|5p)PmTjyo+I^<ϭi7LݷJ?1| Mrp=j^"hdPf\cY!X_9ODq%IiR-cS857q8Rͻqcr_N)۲xIfViI5.QSڨU`RIVTFɗV8:Hk4l%%¤P-xTCG_9Z&lwڊ!v6֚HuVl )Hj)1]> 0KΌa k\@7T{+@tdR5ג 0>;,˜I}7&^ٶR14aWN'aM:Ζ1T0W!B )hxzOW1S x`IڇydF`=Dءu0b__ W&ŔuuI^{ 6"XKDI)C*9ΗCHbqsR;ht 嚵Gf8gf0ڦ{I ŖM!.YIs*;%X 6THHI?C}ƀjj. );ַ YbDŜlənl&xnȄ&5.Pv =)K+#o "u&G˂@4Z61!D9X[ aEY@<1`-ilN#Ď8$ QT1PlC mK2x?%[i÷jFZ7(ƺErAqk zձkRXIi^ Ȇm45c|E l9ݲ6 (ʣѭE܂{vN EPCO[5qܼŮZBʖ UHFN" 8T&z)  k}L~^DU`xPl#f@57𮤵4&c2M!lB S9rm>8fu0 c(?%Q a*)Ze$80)SbK@42*AΛ:2n p h PpojBDC#)'t0 "ͪY2[`Hl Ύߚ"@NFyͮBA*Ѥê޳8S[-q~B/`ޏ2ON BEju"2 ʻ9̀Ȳ DG4%9>TVczR)@<3Ȳ .[ ݹZE3f Ƙ*FS;Y " ?{`lv=ֹ[E7XMPd9yjjRc(O Hv9!N f.B m%J# I UxC#Ut2-,EѼq`(Kʈ%Y'[_x+ nGf6QM‚Y,#;59*0[lxVNO x~2Enx1+ 7D2J@RRPFpaL)ڽAXsv,J٘y nz6R[ڝhK³bFH<Xt-w0 `1)J1㴰63kDŽ`. R*l@ 2 D2Z,UN 07k 鬚:u`l"X#ygjvA(֨[Ko(^B]$lƋ  Y3ca6Y3L 1>[<_MϏ t=_鯛2ѡ^0h<":Gqgs=0z$l F4໳f1xkV5W7ZG[-k9)m` 8h |Ya4p Ȁ*rxnYJC6tCMv r5+}EAtp l )aU)ҌOlrAX7 2#ʈ fCq Hf-'0I)I+aX 0p\ sUp G 2"z p z g9(D ( q hc0b4#61x_Cfz@ִ^kNǚFLJU0ohC*!?!\ly~9[wIܘ|57nD{xhL!PDZ3%ae Ԃ3 _ XX2LtX$Kም{:h7PO,6*nIr`~x{ "i6rJ޹q$ z] #">r6N5 "LHq^%RHYe=ꮯZV۬}Ϣ.5 m`tDC7KrcC#Їp=(@@[@$ѡ@xeuC ND4*7(j`0= *yx 5%NUVég_$7}ߖ[ ë{T-qV之6T ScK›{f(?*cXC0p d ِ@E\ TezI"%HDܑBI $BI $BI $BI $BI $BI $BI $BI $BI $BI $BI $B@sH  6iq~6$H+)@%@pbBI $BI $BI $BI $BI $BI $BI $BI $BI $BI $BI $B@ H VB Ę$χi@C" I ΄H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! rI .,#ωy>$+ճ!@Z%)5C%@3ƐBI $BI $BI $BI $BI $BI $BI $BI $BI $BI $BI $B@Ct~Pux}Z_?R&>-@`%Q%VF .}3h(޻OϤ(]ꇯpx+:Wuᓋi/:="\06Z,ˬ$GYᛏF Y-$Lܵ F%} 6WK4z2LNQW[Y 4`<:D,j0c!OP3wѭP qҟ6ޓVjCh~nS jVe00f" "x㒃HdY+Lʆ1m(s!ki݅ /YzSE.S _hͺixGchCO>Cf&׎eַN07ڿ#ar~*e/7r͟s+BC+|coRU].ܝ(nǗS?>#OI 5fQbRsoݬFlTwSըW.\H;=;imJp4>RB.T3u7N5*ʴ`s37%?tY\5X wb4C;~fqҹH%@)<'8e}RN9XAh}h ajzn(@H.t@@t=~W(IƯPopZ*iiAiPOQNu;jKv.\݉WԟսQw|~]>T1BðUP^Cevj g0|ӅL;M4:o=UKsm}d|6=plTc,A>5un&:ɸfu o^4x%TKr}I益adc1ˍ(mLa?A Z,p:Ѣyo|YTFn.uJP.:Jarf:,t~^(uS.=Cyͬ*h5 ӝл8N?~?RfN0j` D"0VߏCѶY܈-VͷV66_yCԧtIrkgCn@t~zFa/i@Q7fJD%UXs j~,F0}D-W`-^8ubIR4G6f#<4'\'%\oqjM5N}T'$Y-.ib4f|HI`%%MAr"1^xtzaU3E{Hi!U2uH\e}]p,-ܠX8㢎Ҋ&J FµMzR,eP1*.d2pQEQO.|a 4B<hBK~9;w EwJjEy t1jc]u@`(Q{,|%p%i=~iQb}@¨;;Ю>qG`L*ι.檤tl !`{$&ha 2Tc}x9$ݧwt{OZi֚jw]1Y1(묌yՏY,pҏuSڗ42@~!9tVJ{A@7`ESd_7psۼޖz_hӻÉٕwwƣ%߿z+$ 6;\ fwƆ}~Y>=4_XjͷyZ2֚Z3-]YZ LLw`B`rGݤTj&qI tOof~zQЍJ0jD?#$γdU ^X."qSƒMov?/RWd J%Q4*ep|LΊ͓Sy3GAVkM$BVg;yk>k~ڣ7n|j-b|"jY\ZrA}!}t~e3[X{kAvIk){|J_uPIqۘ%ȕuTBYd\˦eK ƈV} TV{j+=Y)8 ВZx_;K9,%] $\VУMHXۭkcYA`J-1륃F On/yG&ѭ#ZG;T8O;nJNL.\IŤWΊRg@wG9P@5iWZ{*&~1`x! \sEM!T%8km#%cuvUS<-zs7 Ⱦ{+fKj>1 l}ϙ`Y+rK$eR^&jhJH%0 |$"% #;˾!e#KAnh7O) 3짎+At2Z5-t.̭4kX9U} *JR PӱPh@{kk Mj.ʲlqXTˋ~k6jp2K%12RTdL8kKBrviA =( F,~?3ꨱl}Z:F($IJ1q[ luVYdrB.&prUީNߦo˗2eSX+7AGh hhO3ϦnU ]AH[}0*Q+!l_E+2J2&h >DƢT9aUMiGE?tg*s 34Xw e#X/Lڂ [Iud>+DS5S+'􄹝z+"y#QdHD8B̃: 'J1O$bJ1Jq̟Mud_ >7H`Mģ Mo=t,eaRݛIJ;K0)\La1E-"$+t| ЅzV+g~__=[vp$|7XuF8P E\7¹u:tu2oߦ,![d$a⽎:?7_ˍG3R2* kM>ZSaTXk*5֚6+@%MƕjR5֚ kMZSaTXk*5֞F &LSaY4֚ kMZSaTX{\ \̯4&#'̧8O08:i)F&M=}poV,nt:X9LFpI|H$h.M2Bsy! c0Ib[jVys >WL坤)ߍ`o+8KMӔ\Ki^2I38 zrpL+ϴZt*+e)y'+p@PlWpܦX\jϾ͛ºVnk8 7KеL*1i`/S唵LTJ9caFot۝TrGҎٌ1/:qKQ6e+<Ț72 0 ]ͼQr"#\RU/@^_t$T񦐴 /ų~5DeĄ`(v{YDDW"BS֐>k9Z+yڻh'[gٳ">U)M!>MSԎ`JnvujA?Ar$ϢFk0 J$juN=9^*3=[nk]c ֖]?`'|¾;p]xCgRި74'>uug߁::}s}|xoORfN?:=~{;w d[Ce;j Hx~vWޜ47b Ҋ׳_eWн&>Ja1u I~$a7VV]tP](`F͚~ʂ =wb nČK=?P98YhUL ={m)y"jN&)NL^gSꙒ PHECG:oK)Ҋzr*$ks/$Ә *f ǂy2MlUmmri2+e98,S>Bv~,a&06'1PO!n?VϨ_M“W]ʅ?G7Ym<$ܪڤD%%b^"J%0%x9|5lSg:QE=\n' `ّCxO?G/`Rh=ИjZ2%cHM#`juQӈjPr5 մT+nF]FBBEcmU5Z#\ꢮm0+DYWҨ磮ԚɻBZF]!\SR5}@iTWZb]`Tm. ъWWRF]=CueNj.B|M!種,R! ~B&-0Aq@t2>٤ `GWNXx+ ϰ!Ņ?_d8uniMa29Xs?92R&M|(Km-c+,-^s0[Jug_DL/,MF%?m<ɐW&Y>^+7o5WR\ݜr+9yPzڄV]b;0Gx$'g] ݻI[x`fv%C{BW pUl{::8x&f-[ѿhG]?xPz2N@rbr>*`NeNtLPSΣd3;@hEo B>zg~"k+fr&kYį]^#u8E;:ny'FݯN/n{ 6r/o%LF^vKIZ]I2I%b0[~qLnS7x.7T2I:ԃlYpZ~p'û-.[~ncȴր.i-;鲓ދE1CeBX •h$ͱkNlNXUw`miI]XfdRHoHdƕmX:4o[6s'C]qyu['B|7m.%X\%Ir\yW %! O{]-fqQGOgW7vqwZ00}d$ŝs[*2o'mWr~ո[.XX\6 ^ ZK{q#ze 4lɸd|b\ byꁗ]>po)}SڞuؑQR2el=C1ԯ'׃jLy"͙RN(baT'h;_s^eq#Wζ^̣Kn0/lñ;he:,F/ݷ^<ٵ﬌zmc' 89TY<TRZ#u%BCﻺB?Gu)jBE]ZIؾ+DIIRZU'u))Kkh9wu(m3TWZkxٍGE]!?(l3TWF[KI~ʤRWBcCQWP]Y%wM}F]Z^]!J,RkNj#|fcd.'j -}Ն(~+RS975RWXڨ+w1Uz)*eaa; |+%W^ۤߺzٍC?a4Nc(#ԏ'^lá@>3&<w- д]&ж`PFF Kkc.u1b}7bx5F1E[SК ㋾i0F_D5Wɰnc_V=aTE6G쒘'`ʲs9F[@JBfE % sz ^2O_iL,RJ7% օ>l?Bkո h/W6Z+T6ZP7|M6}ϸxnQ4iLټO} $ɴ?'̧828:i)Fr|hyRi~ y4H=y뽜?lJ _>|v| ><:w}h9u{@fƹ|џY4?NU }?lp v#< V JJf<4#l0dތyO^^Аٻ׋uAv<6s:R4 I>\TIe67i4i')A^64hBK> 㯢EbP&UEX2kxO#NX5r.]o\x3qx0pL+ϴ?gp = \:%ovj%+! . t0F2OY`&c)S EauR\n/q]YHFXCE ,p#g8Qk%?O{yl<{6Q'*w6#ħ xs*Be,ƭNgA"㔪r$fe/"&A7´5k/hN=9Ō*PIoʜk!Fkm302܍*$ԧ&iz5!TKmxp8>8:fKh,ESxM Sf9R"$2wƔVPdJDX.Mw%m$ ^>6#cѧĘ"e%Y)R!{XI ghD|;ƈx%y;oMǚ`NpJS{wmqM!E*JGJ"(0!xicVW&$:lat}Us GQ/ؚX4 vk+x෷Λ:[;d^5W MO%S6Qw@E+sYKqU,{Ӈ7*08P&Me{1|!L'U]F ٳD#βw'Y< y?|& џA^14 sϲ~Pl_+_߁z.8iTee yTH%@3D(6PS\Ofڠ>d?| 00<]t M03dTZ捌1Y6ғs.鷶Y `\'4P0vY4a#A@0Y@S4XP2VٕP;& S eLSO/aKMKoC(REg#/[>$ LA Csp_h /Xr| j9Z1parV ӢL.Nϥƍsv'͘ݗ^o+  zѽ ~NUMt}́jV%DXm(q)J( jW!s- 4ac2-M<,Axo*% @jkXΰ@p15 RP))SRѥ#JI!mJ "- Yc,cSIV9c$wrB4QP0T#}Ar0#QΠ.zH Ruƥ&XEd."|lH%1Q d,ռƏrkڍw}1dK%ieV5NGQJAR3AE@ QkbX`0TM[wYCIgޭr ښ\ jÁ jjL_ѧOU!%ae4&;.\.NE[^t'1G]=9owXVi=r9#߽yUZwKz=*larQR&[2RJ@4esVE;<|_EL2 ߺ[}oZ1yʖ?MʦõYs0ånkٸβG_ۍYwnDMfWez/*{/^}Uuj/{Q'$՘݁5^fc5^Y!m1IyQ^q)W2:?y<y,!<Ncfo |Ww.gu!=^&&VQhYIe-eJRQ:,ux*5[x:e2xGLn@$iyW&b'UV\؃o*lk0[15'99S`-3M!\9%;iA^ BSӱHB1g.`/N``k%qHgNy_fϲ.NK2[u_tUۚCl՞F;-K]Jgd`ɳÉ\NRzv8QPk@;,_} s6\e?zɪ/:.ֶxsȦ0;ytΩ LXEv y%S/{ ?կK\|P~a2P|= T߲i{f >y_6pp l 䋳9,λay.d9F[7W=8b GGLrg&H]o+_0,>pV?R\܉j02'`yqz kOmI3"ENVz5O-^5+@4k^OyC46'_u(H.&ڴR9QzJ[z z?*m}(R/ K(Yo[auHϲR=nخ}ݓub.95ONes;#ݛ|8:S%&W&ݽ~d3 с[*]JitB2SqoЊyп%Qyt}\?S,W3IbV?M{qnĄ經>j&j6gNVJrP26C>r_ }\ik=nuر d;A|ݨ^op$ =3~V)&v '67`+Yԉ$S{UHzaꬰ^_4ppKn"'Jќdsk,QV8dn98KWtժY{d+}Hq,@^m*FAARjqO<ZJ8֌FRRn&x9 Ȣ`;X+\$JH!*}-tDݮ\!=:𼀕y5zg6qeetyZֲ8]̘dVϚKo6v*uqllϖlTΝ3AΕe%v9 X̵Q*g er+gyEAuP "FAdo(HP%#WQ0aéE֦-48F#c1fJCQ`EA[chBb+t.crVk/qμ*0 j(m:"(h=2++{p!Dkc#8:*N4 P9cs&(sNbt/țkŽ k,qTEA4$I/LyW3n̴hw<B[pJEs$,Ж1–g~FH8x! $p")` EA ibeыVp4ճ.wJZ'$0pЅSA(c"rJ0Qi0*YkP"40VLZ"` qԈ H!,J(oXk4Wߎ@ػOg&DBǨxXp Fa$sZ'E 472+C54]$ ;M$.Z~ q"0Ts,=i27ܤIEް79{atԁ8VBh1#C14XZF]ebLbp4b80^l&_+sb+b x2 ຠ(3"~LKF Y¢B4n6Ÿi"hpޘQ{/^+ͼ\qPmEDmq_7:kl`k =a ,*$B(BL}H, XbT9+.0v45C^actՙ~.,d+Y'Ҕ@"BHh谉H7jmGRYF8cÒfA X=?BH ٣#)=RBfiAxdTRUdE h: ]zAƱMl:7 #$ 6R[c= 8vnxM7Sk8yr=\J V}:v`K1JIpeWzĂ* &QJ1YNXG"\JlB"x$HQXe)A@GgMtM5i] 쏦 5 |!U6p>}ArXF05 -.DeAP@_ښE[ jÁ; .f{2`F 5C!#HbR+%kC2bdcD|\ aE|K]wU"e1DzD>%,@2Fϴ,sJÃ&6kڰy/ߋ4'_*YBEpr. P4 $#Y]td}Q,;^'.s,T:1S/ߖhaC'3H2\2o_P8t \9}Nl~^ G_5'DRjcyqϖ>ߗHv5lnz;za<üoŷ&8JLL&tkdʀK$L- M"66]]}Eْmmf h"|0σzhߠ^S4#+ym;6`fF<I[b͇p;rAn4 WpeyOtO,~6n2I$340 qh['ٵI=_{hiE@n DXte6ergR\Ǭ?Wzsʯ!պN驐0,X"R.J?%iT*TݞT|@Y~"0EbRΪ ă*FXnO)q\Geu7NF"3O=A*Ӂ QIJ`E ͫ8aI# fz!ձϵέlVJ h TwUB[RtrYy%k)MMǪCJaUG 0;*]q=GkaGP"3lTKsNnܫP %XtMfFn9!NKH0A)*|F/;d{i"co ޟ]L˃Qw(sV`1C $Q%BWHEEF|yKcta?ֈe9gafHPo0}HLM)).Oi$QzXlw>͙;")9Kd 3â'i%Y@냠#/EjU\ˆob{;*Bົ(#X h b;'nYFЉPHj=8|:q G'2V*?uT)_OyA"Tkwelh ݴa..˱nlDhtR97~e$imM)*A]5jY,Ai} #F1av6]AoX4kuھBd^k 9L.=Cg;lt Da]B̉z]_t~xc~?cw F`RlL5W75@ϨZ7U57*F7jש5zO2Iz gKn@~u!3xPyp "C2:Iێ6l^VϷ}8l!e!xnIq[@xNq!b쌗6 lQJxEZ`BBȦN&WN 446' h0F˞ҭ*v14h\j^SM:s㹇yu&WI#eR1EYb30cMpTDEig+j_=`k U cJ#68R$ <As%%cm),vG[lzBpm֮xuqUX oKdؿnW *1$qb&H9\!⪖8*肶!L +7b@sVA-m=3{9A  2\X˽ǏD RJ'mh[ի@r]^|X2mIX PY~v>o-A,"fvt)SlO˴/ǻa/]:ݓh:7z<87ywm#I{ ,~? 'Nv0v70k-K(9_5)ɒ-ZEڒñEY]_U(ц|C7AH3fL2)M_; fo]ṒZP4UQSI  6r"ĂGǑ֘`"ٚJ'0\0PҠmPV# $01Yp4%ũGPM۽M:*bFˀ}PyTVf2GSxzh S9>P2TGE)y΃I\ !OW FtJUwۘfj' S:!d0p*eLDNTH(gsBT@ 5*gEC%D3b;2*Tz[8G"W p+!la9kl95JGFӻ8-BZi :ik."kK~:EcC: F9X8,!rvYŘdZ~ H>,C enrus0-6 p} %+n*H r˄.7r㨔ӱTpPyS))R_p2% ˄ϕ<%\J 9ҥTiaYZ052>0T`s@:6ZN~^ Gq,DũE !Ie*{ C :N[c=( 8vnxM7Sl(s5 M c[VJ+Wm " ^Rb4DA(%L#-a* H')#).H 6nFn!HfS@"BTvwڂP6R &[g\eUDj͔#EDMOUϰ4V:=*-W]Ô19Ȗ J  "5kT3nTX:ŰDi(~vd z=vg[֪`P|p c૮I<\jy9-{UjdkpsPBkpKTf<&HY:~ b'+ni/4W@c3Θ.)΂}:fۧ-W?NjҌuFN܋r ևݏRDvj:WzPbV1ǒ[3VUm{PUhr}X{?}KwxW`1 ȜxXfI}w!7*<8)P3ٯ _=lC-mAAJم tcv݂if ns 4 mAt^Mq4]>Q8<8K0,Ӂ >(*X߸8,0F)󵽨3~Qi+ղH n8XCP얦 GCnޖ%ܼ_z4e ;ѐd"|upGCpM EknH.g9  }URYIfÈfFZiЉ;-*+C||D3*a'YIkCU.vqfӒǻ9 =8PKBvY"=Mf̌F/a>kaOB7J'7$~eh:ZHFkeysxDenn (zF;WY^umЗY}8 ǜL$y20Lf-p9D O8ԖX0~ferP{KK{q>3|kBcͩr*W7Ti4ҋ7V Ǘw)+z)K+P 4_pCz΅;JףRaJw\`D&yLF^X풨vG.J /_MYrيh<, bnD1T)b+t@HmojR5sQ 2hV]ՏoUr[U+y(G(kC/wKI_ø7bE~m rwAWm}7/Ɠ/|i^f϶mkx8Y/.^0pay磚ق[0ங^]0- 9_6[uyֽqC:Ke7!S[?/í5MH$F*L<SDcb)%р-wN-gw7,ЉZ]-9j|Մ Yz)^VKvr=.J9/:/Ҥ|o{{N0y%4@lK-<ƔEL㨢FH*&y' ;sl~ˎt2h(3Fb7d[m{ؘ&TUW9vfM3Rix`,f(pp dV0b98U2rD l8ڠmI5yw 5T:g7Z;{Hw KknE=f@mS`H@iW}:tyh\ƧG+[EƱCGM#Y"J-:ZdZ%bv)A$I@L}ځpQ( }U5;G >1\N7{-bO/~ϧa咧O?|jךLeў\CarQL +Klp 9+rV8`I^i" >H]f)ڙõ*uU[&)"˟?p6gLANzˏo.cz'+?s +;t&蝾W:*;xnLRStY^{gEg6җ[DG/Ǹ> Y1tfY$XLjAuNHdM#m!$*8ʝ7\ w1qV{tHbaJ軄=J+9V{6i;8ޏa~L׳뇒Һ ۺCZwCl3Śh8[FERYG$XyfK)*|F/Ո4ذ+1K#ucϵ=]z7n3=(-U]=t֣ɧAŶlY!A=x~XXsSfSeRH{b CM88JOC9kdn4rA3%)S@ 8xQf܁[VI$bq*C.*fh 놰a,T1I .pе)Q%ZZDVrҰ? 7cqS:d:,tD{BBX"cL/6(ڨ"AB[$0e 5+HUaׇI GQ hcY1รY,KAZ6bm|Tퟥ5"E5Wƅ?fV\xV> \[* N Ms/zfa'/BoJw׃A1x'x3ឝGQ>8)J PMK;3P0%P0 *:mqk ˥C_[jonCwH:KƳ|21ERDwe %RQl_F''FRF@`6]0r{FP_($˰\hyvGm oL~m<^y{ 1ŸEʻ›|:0۲[4_8~:/iK-mׂDd|VOVkPLe31}˪iH4UN,@h}O LTLV|2, 9z8*g%YY5k¼d5kRFԥyMb/ޤ<0o-U *'~-=7:v߽{~~xs~yw3L}x,~8{ ?Maj0rjXn3/k+'>I%-١[B(q7]ztryNqttk5#&+l2hbbUYt{*D &_j|P@E*&6U/)X q$-Y>u!3xPyp "C29IOmX꼬18tI" P}LDպ˻6blf}F +qIEC;ÀXt/0i4AG]mT7$J%Y2JEɈ/2.J FµD=`4}ʦ 3FEÅLVQƠ7tbԙi|$z}ACg4r@>fh꿲V ;ԡڱ轛6kkK-"ሴD(e/rcڤ_\/] e#]Y%Z։3S06s *\BL!_{*!-u>Ϋ|;Bv}"p8DͶGZBZBxoʲA}~p$L i~ro0wn6&9b zAe_!Db:Q"u;L=S 0Nrȵ*b+Y+OKL}+p뿹jjwt=ҽwף7BCח m: 4?̷WU*;:k[<]qc[ ҜOt_sK:+ز9oUE:UoOiF^yY_-kbYy*'k轆4T|1.Th? rZ=&LGpGm`*Qfs3DbKv/5{ĦO>Տz˪{M C,}]Nzωyz~Vă2yU`)U: gf85c2( R (cnг<nwk3m:{"Jb. ^>Ud"w0uя,tp/x5 E(ɦ=¡e:eԊjybck{5^ܥ<0|7+']wwԤN1o&ƃ2&&NJRd=Cu9uR)\B.^sמܵ5{ʁ-ue{L>޼U Ç鬗 '#EF,A DF-cU)@"A.dBˁ~}A.x#@)eŀs'BN1FIFقvt:bycPBPA NmRK|PyIY&4w uc-=s)99LaMyޘ76op)c̩6m?X (pB>cN(ss+PD?<7d뛒sW!#+2ӵȉDoOV{!7=ſ)Uw&¡N, pKN@:%ajF4B$S3/@?l~)?!~3(sncYyVT CZ&g> N]A(y5d>ϐp3$ 9C}3CB̥J\)q-j1R%!P-}(ǘ Ls*<27io2?AG?:Z_ u}o}ϾUMߪ!4DbT6j9Q"YLrJK 9eG>8!xHI:+L)((Q(hRZEI)!o"[⦿<%7xoq4y;jB|?F9۸&pp8ҧsQ6]La`"ܼ#VPZTyWp`|7Muڏ /Lx!=R{`wtՅm&y,{tὥ@ "oK=}Pus 3ðYJb ~鯃*ll,|j8xFq*n+rυpj:2Em䘍dw[g<#jD=2:pBFQN8e%#O]XS%EIiޞЌU6;jz?9C;̮“*vbIjx];|rN &Bj^R[R.g,mpDΈCׄ#::d,%YrPE:=*&;'5WP#8H`ɥN+lt@4̫`]TN[HI($ 4=1BIBdt gx4| {E׊.SJHY%8O^9;Wozc 9z8%sgd@ mrgP id9^GH'5>:r\ϼaZaLKC$KAtIF"|UF1}~ =]+4xrOxb&˅cU@LV 9%Ț=]M de.kg8ٻ֢Dà8Bǁb.PfX|PB db^9ZHΔt%S) Βֱ>B8ɇRWtUC|j# =mJk7W3|㹝8ۘ7F($DIs$yB80*'DXLR! %,gS鶤?m6qd ԡz ]t6^~co^/xt6nY1NΠ#,<Wx%04Wpz 7:IxHL${8mΒ(SA%\ IŎ1Ɏp/<=By KKS`Ҵ{D%fYkf|jh?LG+Uwn=: :ތ'CGMLGCCT+hx?w$@&PJ"mIg`z2,ꭄuNt VvJhi .CtR+eB#QD@NJx4$hboe?Jq/9f$R~Z+_/[PRPOen(&r30[ɀGg̣πfdTF4SNs]t_IЂQWP 6LTNq9/+ټMR̲\(괗p$PeN2O Jw|4wsb BT.{& p [B,c( 8G]lħ: ![ #b({->(M`N*PƹDQB&Lp=1p}䝉tg$5LYB!,$p7%\ѝm 2|p䳒xlV6Bג>RCM'"JD ZR:D`Q%M0ݳ@,-ӗNQ׈2!$`RO߸2[<| V0CΠV {=דΘaÁ+(kc$!SD"8xxV'bq4DpLa4>RlZ7>&V,&'_9 |>-.&|jBxmZ ]3V*u ) gƨLܯQids[w'"8 9L _Z36ۇ=$Z3*P$U4HZX ˶bhW,(CX"W]Qȡ!;=>Mo8tMIQ7Y:nHdﹹCBw!8&%%+4fŗhY(+m"YUXS#],C^T*!-uΫO5ABv}hp8̃ͶGZBZ\R}OCS>L?t˘"$g S(|.F1ҦNU_BURWU>/05P21C7D/(18-(K$F)R!T;e ('eRsLJ|Pʚ'_[Wْbm.uZ|,`wѕ}f Yo;wge'xs~ m:e03}2kO@ l]?.ۻ|tfK,;Ė,jjݿmUƻ;r;=LF-n?^y=ӗtWwt|='a<]zc[9龦tu es&KbXnk~k Visz-ÒVh&E䑢MaA/qÆ" "4T|1._\NGayTJpP $NrT*g8Ps ff;K}=bS_'=rxqY!ij@ļqij-Ke;R#˿ЗIiCڞqe ND*lJ2((RmnzM)QsglrD`p'&6H|9o\6[]$eOu(S]J 67=r3N:e&zXP[WF/g3TRI%Ѝ-ʝ!}q<"&=*'Ĝbr0ʕpV!y +gӺWHI,VrwӴ uQ\F'?B?s=OOZ ;@FP"ԭJe\-kl0cjx:нƈ!$5Qrc^C#ݏugx;\ˇZ뼐a|:#8k3쬙8bXX8B 7ϒI{V*ř9 +&5v=vjt7i5Vf"ei>3[ ;}Ƌ;`2B5&zfڑ3X+Mp8gOj %ؽ ;u7-\.a঺^5Awkb~1-Kc'ė.L^+NgA_RW͟8vBfJMrcF47Ҳ\3-v[.!ވեҮzY>b&e@ʙ/vV޸.҃Tmdϋ-GY|%zz {Fh|O!Qn?Y-9'?rrXAV2!oXx+10BB20Pg($^2#YobA% 90g ԤV:qZ6pi]rZ9h|d/\=+vDaI ^ާ& y ~˽i~?>l:Uz z@pU  H0žVWk>A%Ln:&4` Œ3JXy *]^/۸oN[[.PW4"}iX{d[kB}lQw(sV0F f(8҇re ^"D*(2:p) 0RHl8r(us͕q,슣sQ,L,KA 'h,qW#jqn\0Welek=%>8XR(VU 3ࢃ4a3!x3} اl |~EW} r0s,j(™=/f égͿJ`PUtb քʥC_+G+՞DAa'|&(&->^5.fwmNt&ͺv0/ZHs2ui"d%qq6^=L1X 6W37:?r7^}~~xy_|u^}x4?%p~nbܚ_~zԺafjSW u Memnyͼ'av*d J?e?]|y9GKAVW̸t.ڮ&GJz3| l~6-jUsiULksg!|J1\ʋF/_~n t|q5;|!o% X "bd]9exČuFNPy8@ apHڵyy>?G3O8 GJ"(0!xicVW&$ :\avq%z5h@/oVv~Xcv:5#OU( Zt ZaW q[O% )vY;,rs p1?i!Kn"Asfͭ$G~Zez(Dz"Nkr # zDGzg+Bq"d2)JhT1ғ4HJ-Q:e(q &n8{s6ȹ [Ȱ, F6A[R.k%\"IQk˹8ZB'q-ݹ&w>-+dFϲ,͞Jp0}G/Z#XZvY.CC"6 `8c~$,h?L2F{&'Lٗcq2A(؉EҺ)A`x/m%g&W$LJLס(E^i*uvx _A{du "vNOWU sc8$}IE&܈wvM)w-ٞ)v$\4eiPv %`Ua)|\)%' ˴:<_+ثzW^콇OޚEP ?j3O> ۭVnWDnJyJPW:E Z*3)CyP{U8ul,|PFIo m0$g8~fw+,?4a+ͯǣQt]+d`Qd\vS z~tl2/ JYJIY-{wǣ?/{ ìRJtpIJ~=n\`w\%P_K9l`2HP _0x`?tDKn{_uf7}Fb#<|{ѕ_^+nk$`],ghDDAsCw b#Wcug H'pc{hƆ 7#@e{rE.̡^+ .Ez ҹM|iэqd/ AXt9V3cmV&@8rUkSk" ka(uFS̖@S@"#Eqf%N*% !΀=FqIJ['y<*fӫ[| z a< $- 4uec2@Hk`GT%1zitF#"rAv7v89B1^3gZFB‥N &/;0Q,KINY~դ(ɶ(m^HfMQQG@R!/V5PQ9e-ӛh0=zzVH-+i/J+p4_ͲRe,nBCFM57dY"ji9sE(zH1s,t7HB(/4JM* ڎZϓ^}:Wd‡0P:uMVY"ɠ$&)fb4aDR ECVrB3 hh&(90~ :(BTN*W8xKeL3R)ͩOg{,C5-5YM\I<(A™uJSqi3fZyѼGt5 2Hs{!o>uSҗ;JqQSVf$-'*(%H"`BTAh.yF)QrΥ"s;$ώe0 A8H0Jd Q2r0L4S>s$j;EYB!=S@=n@kK %i1ɠ6Jyc>l:;YW6k8m Ԝg xiBS2?"JD Zc9=7&2*Qhd/z9}`*qȉ,Br$+EZPs+`= jٰz;i/Ip ;FRILp"|xă:mDbG'r}Fvֺ';d~VOjSeqS{Ax5. HV+X4ބ3@cT&kcEtT֝Gv5*X ?0_g$ܘo0j-)U $$Ao*e#1-tBQlce_^?BcNЫiV J86)&dșl:tЂ֓)Q a6I?C[wu$!"TL Q yr*CIH vGR3:#)-(D{ULfLB 5W$BBNaV|ЈuT$0₀R85df^j˵C$EHpޞ$F:vS;vWS Ip"`Ubb`sJ1MYSyZU3gѾ,)95Xt. ZS@D\hj{De\}ԖhI]t]Mg;@S5~|O؈Ц-QXEHG`+&BxlDhBS}' ITAQ׹@ϼaZaLKC$KAtIFr+=Y>*(ףs1b%OXoQ̺`BҤb>1Bb>A%Ț==#)G~~6}A;@@4MVZvf . Lg9ʼ#M.kĀXZ-EaNe>2PxE^BF9n %u4qZ0_FEr B2*ZΧo3 Q<PJǷN 7h+i2,~>x5ӿYP y' i^ZFg\Yή q@ k.i*b}.R\W>駫K1K?xr߄ y #{R%>s kǝm%]93jfmyTQEgg1| l9ԛ΁ ^ ^:(3)pM޶V1xE^,_ G%n^'RJu';?WT-+j2(_yb=a4Q<>*k^N_P.OG9AP,:x?|A!fi2˼V*ӍrV[9+O~d7{(Z8: sy\CTiBb0QWb}209&:(4$id$6j͙Brc.<b,9A]`o w6uCF~~їr$[)_yݞm|(]lʮ h1JFi$!JKo'ꨭNIpBO$,y;a޹xam-&[vzaБf[Yo{ẃ1PxMs8R/r g]lmq$J 8҃qWB #Asz7plp4hNrOϡz6TygI) u.bv<@z%`ice>r,9rkn,*٧G"=8Z*.jU=:::>B26Q_; ㄁gTN|·pQX#G5o5ggnSSyPA;h1)b4o8D@(}8*7%drHW! c03}`Ww] :Ͽ?p-*{Vţ#)1-"[M٥Q@ؘ4Ҥ##@bŽ)dt][CX!5:e- j &w_~&0/VYϬ;%#רq9ge-ʜH79C^e-S)X=0F@2uJ$LbDYW@F,#[FzH*4=纶Kj0fqAཥ4sqx8c䆂US1)wNH}]Tޅȴ:tf?? S+{,O =^<5=nu _LUNYo?N)3PW`!LMA~P{L?Cۮ񚡹; ho=_]]r˸w'Ÿ1; [1?]}{7 u^&#j7Z-A\Y\bGpuT`ZyլvkV!ВEDa獘Bj*!Um>NoD/`{IV ځ&)NcBEq2ќZЂ"$DD0^HzjoC#aE]DiE$񂑄pm!QO@(% . 3FEs9E~SMgU3x%@:2-_z&/v%j6*h"o(gڨQEd<iQ@y?ì-0V6sK*!@.!&B'_Gҥx`L}J)6]1p8yaVm)v iIvi@,Jx .'Li(#2 SW'yRrMHË_K7Dz#i*;ԇSHhF*-myhK9 \Rfc6xCMX |(~or3k@Z%ZJ}:)jaX"ppYkBGY5f%V y8\QmY8j|zEƔƖ!9[ Bs#j 6`qb;z,O۹C+-7d<_swI;BQ2d薎:c57Dz ټ~t-Ym2ocǣ19yn5Y9N&d%mvg gi KZM;MtA'#E ]҂%^KG#z UXk9br](/'ޕ6#"dA006{ 5%-ٮr-o0%lSR·]L*d{0-9r/R(L@QfLY1,yrF$0sjVTeOܔ4t|>RK<&!6lgۙJq 2{Uɖ#$%$+3fgɁ֜ˠ#\%9^,'hg7^s+Fف2ș ^P:( ."36s !e!BXJ$JlIbbAM(96SIEǜYF&`0@,G- MiiGb ;2_2¦,TPb`uX I摄 o9QXٚ,&N&ߏ(j 7mѲ|N>y \$GPވgg 6 <}_ڟ2)K_F-u4cjt~9{߀Q|]^Mo}ÕoyhՎ'wQj`kE;2DN?O~Vg>=ZF}ӂ|QCJ@ E[5ɵ򏮽EDŽ|O wT^+Ṿ~7=0ouqc:ZPgs`68bzQߧ՝028jzy;.߾JV0`s݄DhCaj8яfR:<|q xLZ 9#Ґ-2W yO"β(sQIi  ,X#% cCe6ѐ 4r+xkb !$vwp89{Zcśts>\ڷLGJv|ZiÎ403vDjbv9IKrbc޴(9i$s"2I@T 6R{Ò`WZXoG=.5]R5)Z)Dg! o#D$f ZJ k9}4=_%iҏ9On?=wi{^,I/Ob+}M.gbtMN:!c&}٧.)]~n׿Mźu2)&{Ac󗲱1rytrx^d/ o1m o ?PCQ=#>"$ڃ#U-2\TFe(O藤%ia?8G+ˢ>A1\R3&m(v(aE`(LmMulCxA'g4)R"Kd2j'g)Jw.hzA!TW& QuM}]l h8Beiq}-Ux,e=TF*Ce;K48\5 ~!3"lW.mM7(Xlcrtr1r8RWcbǛƒC5m:Yj'˖F\J @2YuJufG!yhVg dDt!y jmLQ`%: ="VYz{I-䛩w!nٮo|ZGX9r/*U.r2ceUnwȓ3"%QE3o_MX,u'W?r3/wM^ǹClټJ-}fQ׷3[T/ 2mJ@Y%QB 8=mnrr52(90Wc(c|O//s]{8s :ޚgk(`(̢_fBO綀4bHbzDħz9@%G<@?W if :gnU3*A?,&4 6|zoGZ94s6d`SH68NIShުI8;]e:uCF_wNխlنP7Js0w綾f19g5Ǻ oWE_?gBpŷ~¢ `>Vpzwū1GF9FonGmdgŷog/wt2C7?3/hjd7ӗ-+SD%QJE/#>A1nFo7]Idzvh!cǜU&D AP,lxIJp>00ʠNw4 ]:UlQ5ڜZȅHm0t*X$qq16 -TOw?.\=7޽zӫ~zë0QzO? (N& ?^u[CxT5YO|qYS^2=I*i:nvt˭ҷůj\ɛ#JR l Wl~S @M`x|By.bYnp_LۼJhTS'i;|$-Y>u!xPyp `Ee6;I>mht^YϟxpB  ):DlQH;e*=E(%G-0!!_rSͱdֳg-׳oZl"Qw}h5Q(|uFgI+LIJJ-Ѣǒ M TXy:݂++_Wr96X96[ZmZZjMKKPDg=`J3ОF.U@{ۏg=J+M=!M =h%-xaҥFLFWZ|0GRjt&l,:(G{, \c% R`BUh@R8]7]Os_Wc01lJe@I#WLXϳ~9x{FGDĆ@w25WAQ"u' XFu9JI:%x92J#"( -J))$쟷F9ht3Artw S=KG61cbbLn|9𲖺Ҭ~oDZK ))dR)K BrصP66nln%7)J?brp$˳T"4[vI/2~^tI G}ѷUAN:m^4%X`T9F$㑁Dꥦ_C ` Xy$RD[+wNA>ݎ]r6iP+y3jڭ>vI O>6lz*@B_%c9]&Q ,Rg2DOIHLDHY,-,xGI4 GڐgY?U`!uBwVx<`9Ev2I?-OSٔ\\w$3 f g%t\]XPY(VN?cyƴݽvVQOR6;qlimC%4|0iZxt}/]8+O~ʜ-hcW,L\~$LGn9a3mf4I4hR ,Z9kM;6~IHL5fIK˚y"MOvM:цuM5{h\_xf1嫱|;mअg ZthV-oGl:~?8G(rh %9D.קsGsT"=C3%xgM .[ j< KmVKnӢ$JђdKk,)Qp2]AQ&mwO f2)JhT1 $ R#kNPW V1Ni'8GӾ8tFۚ(0u^Û`)S U]VŕG{sۭ!ӃO~57Ez_,LNLpvб9b! @DdD.Fg&j ;v?3QIy3vygW{Ijns7i恁 ~%_o~FVv0/1. -0|]b}.5LR1JNHMjmD.NT23TTJ>gGLUת&qF.65_X 欘;4yqF u^@;!eP4IRE:NVaotL {20e0.X>gZ.[~efdWr+5ѯ̖PȈ58zItZXJRtkgrL"(hcn%TH֌[0hJ #ZiY;- RNP" F aZQEfθEk#H-s)%=.[xduTXCWeu>vZkǭt ^KV 2Xn`,7f /ZHcH*2Xn66f#761J5fڌWq73fƫr& 2f,6fڌWe"fڌWj3^mƫx6fڌWj3^m2^mƫvU4xFLx T°雦xI#\# nj2? gtjh=2\gp3uf י:C1B#Ju 1-){ӂ#+R2$SZa॥[9\!흌f,͂Hk`GCR4`iNzpdh*ō8`1waʽ3ø)aeS5r Į1Yb3 pԻi,ܽxOI/^~YOtNO3hl^XBMk}2ȿ@.AT;o n![~Ѷs[TH6%8}c c c {cH>%u{S n7!Bj T` Mq Ɂ!N oFbWs?Uγ vƽ}'o7sz 9ϱ$qEb^j0,bGE0BR4"gk]0&]B7_V_ޱb0|s.1ֻZ|2$\(Pix`,(pp ʔV0bJFw Y 48F#c1fJCQm1T)b+tp6HF/g)psQ 9X'>cmFj5nC[cn2㒘:9iRp_#xG7l:ZȠLn zA.HG̫$\r-99b,ϰ=][O$9+2g| a}3>lhm!_tmW UL7{.TdQ@-@ΈpD8Ent<3zڋV!,\êP66 4k:Ò/d΁xRMq%/?T vuu=n!J y_NXXVx7oihvjy`ZgZ9^hYˬAAH`|s?ؓS z@f1z^ybu<{]V8uh^Yp#%mRY²^>C&R3=:]Q~q;Ujwu{ڨov$VSZUy=MGsnWOlcejF_5/roxuŻeaziqnkwa=7`~_t6{Wpc}^P72=]FfUnq˶13NF޾L=\974'D l7hPH0>9v*Ǽ!`i{tMz;~/XTVrBzkVD-f\bve#l!P̨toK Ƃ-"i]#^)L%D Lwo:tb>e\j%tSVm7׼8O]! ^\W8y~ysGuS\wKvu-!p)"$e[eRTyC ^[$7R=K g&:@plBm5LS u6JٚVHcMQ*gA^^Hfd.*{wǑB۔T1 ƍF҂g9덜嬕1['na PWBR*րTR!19(5}Ψbci, DORQ-ڱP&TBAPٻO߸Ţ=MQ{#]V1' VoVO?i»߸,@D% E)Ƣm1VEyP('Ow߃L-]iڅ "*(Y¿DlYA@9[5zA4 mEDsj=~ѳ}Xd߄o ENɒg*m.b2mZ&B]Q[^K#$(ѱ+ˑ-*jb[UZ_wd>IT9$Y:"ɛe!I yY"g]7 쏤$rw~pxdH&OOD6"u@VB'*St[(M|CfB,Ǣ6ׅ$"*WщG"RЕWkBV/*-W-c4idldLxò%QXlHhF笊E - :Ƅ ,푌`=}` п yb\R0+z1+ oc8  %ɔwqa{J}a3$hg>}vFH5En3$D y (eK4gc_.Jy (8r^I8dG6/@nLSFIu1O!EkIfaNaZ0v;-Q`tJ gG?}pp'<|,[>QZocH ǵ5+/ TZИ@90h1S(:;A"*$ Ѣ;]0WK}7+TZc7l_@7Ηp7">1py>x%KS[ ޒzd ]/x_"{BA:IVDo1/OOp2Cssu8n擸^sʴ3*rUntQeKc3$GbBɷ6;MX ׿6g^A=_Uխ:nqu7ݪe||茇Vu&RT.V0dȖJdkΧZg }~lYU,txYξwv6w5hkwt /kK9ZcJ%οchk$SƋ ?>\|kB^0G4<%h\Lt^5!܀ .BG ,xpx׿7uy7xt6.5EgmX@mLf:?ls:n,̢<9oN~—SuNg+־?M= Yt%-b.Z* 0qrVZ4 ZAv5Eunفx~nv9KˣѠ x6t@4 BYJuG )]a^/ <%Qa:>a}EVu!U_`7S l'M)H&{- !b.&H.zU@lA^yBXl-}eTAL%+&LYNZ2C:B̃ 6HcJڋGuG5ٜ>*jz{b1x9<Ԭgu% xkOZ(J&m@+hP2Œe(}zz{ zOFiN:xRe1Q$J6֟/XQg3A*"=}(awzrZ1\{'Zۃ:G5γi>DT,%|v+#,.+= l^)됕ّ3$`AwPà=U$KzT `RUPtBSZx7 N)85ȰURL4"^E4&'F 9R'Kg(])AX(QRaʱ-hAW&"sfHH)o0$(ʒOZóAeFIޜئ=Mtg5DCPWgg!F蟋xS[/,>?]jvR.o8Ws0Gv>L34ZƳjf0H 7oŋ$/@Z5GƾfHdG?!PgYYlF`6g&Y>[}G+(sUI>ʐqkwג/y NO+tn= Ot9: i[폫ߖZppFZ% pvjCjjJXnqe9vzJ~~|eFչ9nŜG5ȿ,>b~MbX4cU%mqZoinO٢t*=NW&{jpUu=ͣ&x6Z]k5<@W>VzWbfk=mP-`WLRMCG:IKoIH@NZ>SLDV$ Z5؃" 1V@qp:ڰyM+>LjÞߪJRh_^B#v)%SȮVpR U6SJpˡΌgr7oYx,=9yE&]5/=KӚMgY|uBsda|:6m|ȥQPئ_IiW/nIʘ eLn-c톪&o j3YZ/E'= EɕnSJv7"V?2oC)eDߤ|&zȍѤH/7IcA\އ9%)T`cx;2$"5$R6#wZW>39wBܣ;r޺nr.{]7"r+iӐ,Tm{׵8;`2+} n_[7ۻnt>}5e-W wmy/zhjţVOu\y  \v~Kr>W5;ڬdYs5U]3en4_yJDU>CϑG-XGwǢh=!%KY g\z+(xH/ajjG7|tx@x4_JFxG`9.TKk',}GV1IvL4v<-8*#\-rD,T!g}{a=v=j0YR2d,Ld,8ƵY#sꢢۉA@tmҍ=?S YBN脡Dr6KGkA &:ԊOָc$q,ͭv6{2 Eb V/1,r&3GZ @kxPNq8+5L"7CeLSpZz#x<%mʪvDme%_kK؄xXYQ#wzu&¬4<*,K&f򄅓ʀu(EXōVȾҡZ5n^G|Q/+U3_D G$t& 6q9ӪxBR\Pp#+_(P\[Xwu!)BR.$e 1PHm3t !nfQ0ϭȲK8H#<4N YN$,S9ܛ6bAiq| T[(K }>#EE1NzNj Ṡ>ݶJC Ҥ&X@,R d`ϗtǮ9ަ1๺ Li4V\aвO2 څ:k~/NR_qzcVe'WjmS!;8Ǖ-tjmO` KɪOo&%f<Q<@CR匂*Z+B'yNi:ҞБ@vSpܜW3K(as&B-S:yP,sc|EF(v-2.h;tt0td9]5ذ[s:];2}>wkuᮏ=Zv*W f|+zW0Pb34i)WePJ줔'D\lRaN5,]7ZxYjnUmc,m<^LX&lS\Z{}kՕ-[w.7p8?/',}o%SG%mw ߹ÛJhK7::>K?*GMչ4W:>w[ ZwڏnC-q@lS(Mp3NnU9/V;і]Wd'~!QGaT}Z]FfoXKުq˫/^-5!Ѹ^s~tcI o&5m+Z~=wئX5]lFlƱ,/$ Z? +Ŋ_=џl뜃 y'7պ%qtƹ-?G}5BY)?4߃.̖B5ǽؿ7޿~˿o__osa߿ӫoxIhlSw~_=p꿼zs][US}󪥅UOl mvyCħE3ۏ [ zo>i2y:D_[%f]ɽ4 5bz⬈U^i#t؊w f#5o]=oDp'pO ;Ir0+)"fzy'B+A1ȁe`<@蜤m뼪ߏqE6yM9U0*ZŘ43 Ak-7;u(3JeFiޭ;e1Qf|0_l:ǧկc Bbcb@f˝0u9:BeF쁮Ֆs oT`Q9n3rk BH#2z'm zDb eHї8"\} O^6AZkNCn>&G/d>p;W?BP}eq9Pp< |77(paח72EUt.33n^rRIqHg7rkӐm{ѥԏv|B͠1t3']up; ]O?lͺ@z7jm^U˻ 9+ J^zsz(77=M!ɻ$[8ΥC.lp8&P9ePǏl-6q{zQnӋ@PPX7v_d !cFLds< QywhMCHMN(NH=B6=;P=0tΔQ] Z)묵fqS_+bQ9oA;j>Ǭ m*q#Sl'5\G T[̶&& 15URckцf|E&.1tH«=C~syY<<' i6;<;_=2uk5.glh.PQ6zbcE݊4S-6ĉ@XRR.SqB *+KA,TI6_՝`J'~|9al+Xd0&*b WB:rͅQ^߻(H.:؆ i&$Ҏa/Sb;Ȓaa1;hʝpc)wGk͡;tv*w;QcC~^,B5T_=0'ԪHuX}oGQͧ߬+Ngj~2Og5vcK:ermlr~&Vftl&]OD/V"&⪯D;?S % 7Tә<䔕S5C1հKmJOxT9m;ChZDIAZւerBjı|CQ¨a{T\LRM>SBhb{Df.Lkv䀰цJ}-tHu6ң7>U2/Hš',<ڂ$v>lE,b;#.G>L| oݬwʻǠ:}! TEJԪLTgJ h!\Sk-W% bQS?&}%4W2*s2UCw4k({U|r|>OuC脊uwD1D1=yKG'h$p%N6CQ3Iq\0G|l[} o|c#vpˀf`v0`Jo!\éty>)Tfݻn/\쫃C zrk>_+~qV&yp :5o2cW]Hv 1V1uFߡw- J<񅇴*ʹajE-9u;.c5blJ,}ٲ,;R5c1;srV6n }iUridYW*7- }r71zWK 505yϞ]`BiNjC1l{Wqi@<u FϷ53߉*Ӳ(zǟ?V`f(E_7'k7r0secF2Y#Jfq)$́-HtJQbvp?OUks/g/7p 3J3Eo~nzT>yJwH@8&H_{35j* HCr3kt;ĀSNyAB^N76}ܜ8eeI5Nj !ea1TUnsRc6 SR3I\ !FZrE05L.#J6ZSM6Sτf^W%w1t٦ [Xˇ/ԍͅV!b[X Ċ(eQU 5!E+6;rC>rO8{ S.&*^8o1qIBUBvNĕͨ8K%G WP8 Q4.bg4X 78m8[l_V!8jOXr_hj+!x BA>gv*Y B1!RQo,%Y-_)-5F[#-$ }⁛* d=POTk4Uf4I7=MJf"dMI+>R3Q1!{g8FQ9L<32s}%{vO3m4Td-R[nQ$j \E5G <*fBx+":2T}Ch"rhR@z P @&3"a 5j*3:a7dl(_cIsGFA?λa-iv`aqM~.XS H2 Iެ+^eay zܐfw͉BR Ib ,+;')eD{+4Q*[ɅS X JKmTZL5!bQoCbV*X*VUIPt0o+ԣG橣 iZf#|y$_mXٲ6dB E(q3EU Ȝ'EpUWq.SKJ6$T +Vބ"&y/U[ I-R [RVjj:Q>W _ʄC˜^E\K&r6ArT5Jl.Fln@ bظ #0 cK)S }V_{d1hZr`$뎨pGr7M>گkZ>8[#GʮlH=M" PK0,[ψXp0b%%Y:Sp8fDOإz1 Pon{-KrV/O ?ϙ}1)9fb0WI\E1`Z45 ԃ3o.ػ)0N8z(jfO{Y+W?~d9qE'ɕgU89i䅞wK6=gg6DCw  .Hϳ*Ry#Pk=`F~Vtzg(՝çiM% L}.Tj-.Y+=d;{xl!>]+Ku$5D… O%q@Om9Nn U'^5Z7_Ro_7֍re~V&Uo)ي>GqbsT)wem$-C/L22/\a*qL2X"bT$RTuXYÃaLY4*r^/̛y[ݩϟ ATV9]w- >sִ5Kg~[373/(s.&•=iUD-U]p%*9*w v|,ip> X!k p8'1e6%;QG 1֡&{'eVxH:H)hd* J3,MAyɀpˏ3I0 zP `nCXb d-a +U YK+К(9NiD?/=]豌xLB:=!NeEU$HhKൠ&txan$IPf69D9.2Eg8:Pi^zOZl& MP}\Tָɫj}g\ôٿAN1ڇ_V EůXCJe>_{7fzsV5Zua~:QMួ D0$N\ +0c C"ruXuÇz * u5o]  S85 ˥C,!? ܗyK:Aa4IX|^?3C7^л3iF4_?|OoB !a}K.h] GopO&_ըYxLѽb_Wݿfܼ~z3Ӫb`cqk~ Fq8-Տ{$ƣ_/ $m]+ݶ-CڗѨu+0q#`.[|<>,6~\sp;z8ZW%hl[YSu9'.M}_8+餳ɖIM6 ?8ӧS/ǿ_`|O_~#8"am$pj:+x?uaT8S7uYGnY $\1fny6 D/?>_ 1pѫ  )\I\|/6%HꞛOP 4@gF(9biT%7Jit_@&IZ )"(l&B)#NM\`@A al1NmXȼ^vJ!En):DlQ`B02 J h~.0!uu:ը2ۙx#*_yX]׫Jy.8,7K;N*Dƥ g@үô5ɒ{"QE)3ZTX%,bLj7UM~u%5햺CjZ](Pv߆Ʌa^0}2̼2ry;[;9Q<,[ѷ9={ng:;Xy,ˣxC3eZ`cNEٴ}R:r>R_܊m~&\m`JmbLw=hmaV \6UG YjT8 HB9UaaR"r7UE~ דVr')Y]cvN^XǾlcNF]杰]Ǥ~= F 2F`[iG;FxùJ[G˛+GkQ@&b'8>jK ^Ȍ ^`A ƜB*dF~ɸ[K9,v2,|p-yzIy`nfPy7.9 5Ԧ RO`*Pc\Hf"8hJlhd 3 j-P&ҁ$D{:tBȏBDN;9%qtTjR[]4*97p9f`҆ b Dh32GS*H%|.cyL+\'`(QHLd$&F:ł,DFbRw<)V8p199L"?Vٶ+g$+)Y ܰґ4_8Cؗ0po2d*y sZTD)Z1le% ,%S5x+mm_r^'ˏÑY0`z%BAQ0HJ-Q:%u`+R"5: sL(QmGs+:2,"1HF`rX+?B"Inb.rORKP,enGU '5YG԰p|k/IKhOk}1JqrkDt?QIy(R0 9,&Ϙ*j.*ư{"id%Xf@v:yg q4JDl0z(sZ#-Z"u' XFњh ]I h2"wUrOg83GckJU e~m@hֿeP6sB*XRE+n-T#)ʦ겲Nid<H7$7@j6{b#`%3z۫ӗ>.re v\lK|,г˚˺uuĂBv/Ӭ " $`"{&CD)RRX,-,xGI4 GrM l2MAנdtgkX Fo6NQ0_`C<{(7 _v/p,Y|aYm0ʼn1TxY^ZWk']z֕[m.Nd6K<=t xcH.I xt/ 9xMg5~}WaG@ Fa~}\3ꑟtqvHyez@r2c)+y_t=Á]dS& w8R|^iW-xt<}nٺ2CaR(F0EE3VXR_p2]ꃶ\mAu>/YUگ~LR#Q`"^m*FAR#j(lRJıfT\穸ޔB]*/Zi]n7'h6Xi/bR֔XR]w N-ֶkQ}U(\ukJl%6D(J!* 诔JtQ4E}ӎiGN#'s%%:ޒO%H%a9|2P1 pč&$dI ]O$GWZeou.Gdf$?^{%[ZZ鴣Q>"oYa(hffh+""㕿8:|`E^Z9_vW]^|L}6()HH2UBŻ xYX i ^Ax{ޥ%.;G˧i]:j6/yϯZb.xZW4԰.n؀Ѡ a|rRn')w?bvvx#03,Vye)' .M-%v]0* FCQ^*E,: )e[EҺ&z0z!RNx)JP$zNFiCw("l}^_ƅ:WBǩ5e{kq獯&g+s:!P7߾+vFuQ\hW,0C> $$RD*Iʶ(ZgJ3/s Ar{[z(r l$t1Y'.MHəЂ)>` F)[i[fY%h^Hfd.*kw^ǑB۔T1 F҂g9덜-嬕bQksgmD RBI̺RZRIQԝOsFHd }_ /z?٢[N eoJ%k$[,3/7es`jfj' ~Zxw](h=DaZz4Xt2Z UQFEG/cP{oF@_Q&6˴Bub?s",yv ɜ-f= Lb""L5u7+ٮ_o,k·x`"dIzO*m.b2mZSCQyv>)%@adb@Fe4=۩sgtz0]VRox&:cn@ -ɔd4h#H훐Df-s)$GnnAW|'L%x|T@b~NOד.~ y!ᗺHcK0˭߰9nyB6c~ ߼]b>)}8/ޣ '!s.[\_;/8{4!9$Wk 8' Ң q'~0cOU>wd 6PH8-d 0x!R'-Lΐ @)& 4Z@E S3-Ry( "-7raŐ6[ء<‡i?=$nnw ]#rGk6=gG<}柯U #,YobLF% . 3bdVZcV;kҔ1-楴HHޤD`RX(H2BU{WLo-7Y_MҕR^{U穕=OS[!2AU0<y TĦT~MgmSG],|}~wyo+ԆT~يo 矧i-6$A'QwIm.EP¾|^,10@SEWr.1ړ"梥bC *XJ,fAJ[3FےoR^D.:tᚢl@\lwo<]2t/.hx65 k0Ǡ#/e{vWAѥC( ZTHBTUTKl2YN S'&XDE$=kȹ[c$o8{㎇jmYk㠵 UUC+mML"nQ,t>@"Q}l~!A*c#D:U \aȒNnCmb>d#}3H i=j {XoDq1I ɟݮZj"kVb*Vr=o4\|5+^Pv~;@=> $8u]_! \ EQI"lP2ŒeI}OZȷ o zN8g~UU'I)2A؇T%GXdgΙRhpwy.w+w VBq9Tt>䤆h Dݻ[ Y l/uLX_HYR];PBL*T,I L=Ym8g >Tvڛu :k+N z#úVJS{9Xym+˯ҘEB5(*K.)"vTWQbD͡'m"5$c[ ԯ =)BwB) %:$6Ǔֵ_.1iQ7W)cYŐ>֨FiHя4sj}[ΐ 솸j-گ~(L2ɏ4ZU`4{4ŏq^Dkd j $#P|?!P{bYYlF'bo6gxGǾY[F+(sUI>ʐ0. ׮_7 vג/xCK~DKJ~=?OC< \Q 2~t噖W׌ن8S{MbX4.$1bnۍ h2r>o^2'&߂]#99G:FuE06d':uN/V=s|2p lGUQW5j֖Qg5!J+fLRF2jc_t1-7-n6,NO>~?OzO'9EQ&#>φ/54Z#l1}>r·|q?.2Kn+@ZrwxͲ*n5?~jG<WUד<ʫ0,'x6Zo5<@NrBPgښʑ_mb(S#3/ӳ1/%e@)onT>1uTRJG)Y-L]U2MzTSy1W#)VC3lUThu ,REHEJ AIzik&}8yus5ɒRik%Ґ Y`N/N1|ڗT#S}"=jv;E1rnpKPja5d_K K0S;aBU:uB,CB"S 㯤pكiícfCYeoٽ-_#I)(qQO5\] T:ʛPqjݓ-n그R3؄9m wA&.h:r&q1wVZ4~Aݞ v|">|Gx٬IM~]w_-kI_3]^XȐ6 du7m{YxC6O՜t(-;i9GپuMϷ7w>Kkݢ祖C+~9oBÞ7t< r[CzKx5ӟ7pҦ%!~UۡΖXmw{m!:\Ch>7A'dЉt*'u7z2JrU6ECcGAGM4.D?-8`SC"P-K1V1FͦFs!Z9BٕR~F |+}TsjSbH"6Rbc?٦9msyqEW g\z^.ʹu=UWZufMrEv)~&he*p(Jzg̘E&6s{N떑8xK:p\a=[cU}.R6u0WYT5jmP SKՑ]9c5^.nomb{ْ> (gFU^W j?"g%XCբK9PIXOZ϶j'z\c%N޸U 6' @hQ7Q+ߜzҠ'CWHa x|0EQ7T9 m}$Ed-&HF2{۞x%XEe xkڑvJH58kK[TגYVf }ƛj,"B[C=mm>*9[d`3L`!<`[%D ڂ Rl"{}r;̏h4֋Z"MI0+_-zn0&6iyZ5Y'+>+$EDVCC8ņ(E I1{.$($RKBVhLbLD7 r"CT(JNwWʨb&*fUC*+&CBQ0P!ᕁ,t',D.=˩YNx)~s5$eQ6ɠr%L]}TƳ^R L ҌW\!ZÑ'E[,0L| +bl6Ԋ&<\aГfY{0|;cvc>~38Ag|~rg>\[ED ыCt}%CްӀ3(8;Vpw\*F'RKӜbPZ25&P.Ug!rRfD_@T#G9Hq6rhEte&G*!<>uɊrWI.ʶk %e+arAqRi-BqVP{ TWPQvPE,UfFwmE>QgMG{H{⾗iG#ɮ9$Y)NQ9M%kԂU4޵ģF>}QcT89GWqrLZL^۟-޼9b)2'7U+r7SVTS揟 _DΜf͛GD~1}On>>x,=@_NGpKFO7\{Σ?RFv<8GX5 &,o߄i .'瓬`[v_οz|;ͺZ9*'jԕsm赋Qsi>έOc>G8c8:,=׷[ө=NGGl?qo?w?;@}/?ɍ'Y-s L?F} ~СeCO+5قo3.9~>HZ;QfOʝ0ֿOFftБKdrbM8<;Ur{TTi!5MӦMMc L]u<) guA71Gfk*z哇\N%mBĀU\] Ě(bu mָ<=)=ݴO2:m%B fc.-:\$=/FFwѤ&?ή ]n<]ݖȵ/ŐUl[g~|,d\:]Zwz2ֺ^߽,hE ijN:fun-7=6,uZחk?]'6{rTsfNb < f%ؚϛnz^5m XPϪZf&#TgCԗ5A7ߑn.1hAHo:<ЉF{ЉFeī-_=+dw^Ay&j |&Ug?{05躢 t#P}.ScL/bej {aj`C`\5̨BP!;à2QSuK7ԈIkILT±&kmVE.t" 6y:=(%dE].x:dwW/\èC]_Y ;#gjț֐:Ƥ! $mg"[j Je(3*Km@$D7) NEK$]JQAU܌~ p\eɍ_Ck[ybW m0xa VWNR yЄ|dyOemn٣.Ę;`#vYCJ!weW~duK"Nru 3v_5O܊)ytݯsXΆ4g.͙=ǟww>1VIf`~%˒ԺHQHr1ZPօT v!wV$h[.nEtq_k2 FƍJ+L9Z+\j9!JBqYβ6ϲV-β' ] ߰KϤ`WI;N%%JT\aOcb>/;`r:X# !gKmgӁEdCA(sJPׂQ@c(,ZKf얌ۭf1oC>͞冾L'~1 tKl ]AzR=;TK~ֈhStB(Ikc1'*M%r ?,)bSd2N8_PbQk,[bI^s."qǡR5nڃC}Lبd*&EN` iV[v_ヲF34)Z=d ًN,˚L"IҥZ9jYb.:zc9ܩ_u/*}c8H;?Q/Ƃe" fF !OJ*X$cQ mg6 H;k((?3;:-hH$;Q"z+v뺒gnf \DoEs bC(7$"!옂kHJ;䇎:sڟaC(()wɺRtl0X")8ZmZTJ|,}A0AIֽ#ιTdXlmEnSҾ:,$qJ['M3oOƔ.n_%63v{|HLS> =y1N\LeJub/4$^F*&/(&C1C}1<^G4]tW ltBW6"VƐE6F {ڴ=M+rN:2L=I ߑSJd2c"im㴲H诡B-+Dn8}(=ǟo?l4ٓK ۫T}Mbi9sCW+vxѵzů}WqCrbg5n{ᵛWBnKKx~vŔ_Y& %nbQ&化ȐkSYt҃*JJ$$rRd T?1@;iU8e;,ߣewCGf,d[&5a>tep9d>0OG 81P1(L<凊iK﫽ZJw"l \u/ܟFڶ^zz `Rsr}rB;usq2/v2wwDoPgc}{G/҇}噮`kJr+<:лBx71y?.XR 6Y6^q>٦xuN[SiMYW"ZX'u|r+Z6{8h|UVtqx=6 4~wy%Uʮ$(!'c-9E 6ߤZr\dUN1:PlW<'&gĄT؛M_5sl꽬:mݞ`SOST{y%ب9|^ɵx.+ϙJo4(п&c׏)衢^r7Wؓw\4/fbv`GBP_~zi>wtI3Nw/FR~vj'VZ VWiv=ݔ5Q$AĄf\;f^j^1Q/-KvNGĒP20/O'&HS 9ϓT(h-O 4JB;~w?;4IC%rv94yAŽV|Klf/Q֔Gpeu!-:,&Ứgֵzr"WVE)b·\ %'hFU#W ~Ϩ#Wu\J}wJVHnpGNx:#qU F{6⪒͹J}WJ7oQ\>u%?oF_nɜǔ~ǸzWY^_<\gΟf? WE."h6>2d<%נu6v/A/t#a֊\wƒSu^v!UlΣ7)AFE'l*)1+^F_1Z{U91k:&y=[?6E?=Z#(䋐J6gۦ^N>ニ@38bI Q*Y瘄3u6{7zՌ}rʏfܬR0v,c@鶓oIoMp+~>!O(!*eWK#ISddd [/J_5vJ֧31)Z^5T1,Ji2ڧkEΞMb RDֳŗĆP;Y@">=YjTp,hJT\LAЮP<kd0]q$A[Qu!X2:ikaWΖCl^P<Wsj}c}˜eYJEa(z!<;Z#c!ᄊ199`GT)M2dmߖ`8vpP. ]4i kmFEؼod`1L,9ݞ[^ QX9A M*D2`#{:U l]/5Uj=D ϭ13Ѣ$@ÅvRH s~ 4cJ3_.kcJO; <|P Ƀ3VbN뽢\v[|JLȑ`4Ф#P SJL5 ϔL)GFWESV9#8h(@ /@Cq[Vۭyw3V&8A'@s, 4\bF91L"xm_+^(TI EW)eZ\҉i\Rע)e@He1UobQK ȘzH9Ci{ΉQ!@҂ k-b뺷tgY9j:=>hs =oKX# qշW*;Uۢ<_? N8j~DED(o8dK,8 | ;8#Jrv`xF Ӄ82o-I*yDgQN(OT{~A.HAVK\wP2ʏubC +u*~3 Eؾ_Z7::^0q8MVpeȁDyω<>+5* ՁksZ̛gl1.f'7ޯ+fwq-N8'jm.YW|o8IΑ@=i0{C:Y,iCOZ+W1rӫ1٦ͣ.'5j\sը쾌K >4&i!=ܿnw jC? ~xv/(>_~??_O?ey/?\YqE`O&O#@>[C mJ -ZҒ^,6B)w{}HZC:r[෋o?Ïײ}S-aVAX~F6?Ӂ_tRTT? cvLWr[?h] u'S]=[f}tH2 "U6*4$QH4WQL т"$e!2PI/}۰y=\]M*I#I8 pebD6\ ui-NXRbuq ؞x%oqW^լWp݄[rJAJWq|  f$p]ؐ\&)TͦyU47[{RSp^zeͫ6cb7\?l#MK1nXqMb5%jT2?6O4o\ "¤N=O;S~gw(G(Bg61J0T+`FςT sߞpLnb1J&ijb (wm)K$)PaQ͓1Y69[2es ۨV_?xm svo[wd|j}E䯻^Ng.j☽)}דYg'n5~ޖȍM'sl3]^?}:^w=#l`vu3o{\NWmͫv-l,vnZ[=_5EZ=ܼ' -?'6{~2N۹_r3_`<};Xjk.l빣"U7(lھQ@i/C-j6$=*:+*D&wNdjװFzZt"SixE'$eD/:qj .Bպ}B<jzjtc Kj4V칮M/)rFBFmRIȈ&ꕈI%D:syѫF %p) bTmAL&ʌdXGHODLblyxov٧Fpp44·_Dܞv; U/V۬];!f_d JFis^P 4yd_I1AkcmU, igRi6r:-"uI{ST)Ƅ"S%Ola.F| 86̍g{_kE6ۮa;gRanQ)[fzT%  ksJxcp=uBVƮ;/SHBtD<- ˹]L5gS$n齬k8g y\M?, gF!3~QY~)v>GDDGX)ʕ>E殶X̑xo@PCyDHG4ǿh>CQ 5PDpJK㽕"~-fXrVG欮P%rVW.a/3|A i.I' xIN a60ecV)hR8e(z^ q,tJ`U Jn)` Ғ9KZNdaQeېbg*~a˩_x^yƎFt%N)tpA}0 `Ny#6y!'5TTbs2dFsHA\** *!uDFD҅%v1rn2ke_vX Km]v`7xI@vy-']t':JTӔPh2f(g\I AILYKP@ 5C hE{ʚ,2s B-Lȹ_F}.x> 1 ƾHbJ*7D3[#5L2R a #8|A4Ȉ\Qh4%5 %IEyPR3. [BQRSL %n}; F0l>WcóY޿ F.X JXF6Lkހ%qֱ$'L{J5xG=f'%R=yf3P8MM"-crQ$Ө0(MELLS'QJA Sz^8scc I2& eO5Zks 3%ቕ/EM)ԧT 8CsJ'LqgmQ5fC)Y?Sy-C>Q}QZ=*=*eb_J|djGRZ P |D0GFe eֆZ$#d@m-4 )*2YWbpO+&eVdp|571q(2(^2F+J暄~ uFZ&/~KZj"V(HqseV_\>#? (Ec 7)i洁$JELʦl>8P֏8 5Z6f2gs>; n0? z5~fEEg+Rx ]Gآѳla[^oi;/ U"0kM^ U!2/hf Jq&82Gb% zVo_&wMS2E=̟l39[Z~n2b KwV_w;82fe1 ˺rh_*S0S\] UL[zӫızБZZm \bJnOdӶJL|MNv{$2Rq$q5WH%YիWlG! `ar[q0jF\=ʵ=@\*)#=WHp~/*L-*ޢbQH\!Ww_U&L%*ޢգ_u'b+7i||4xZh +pD ~D  'yӀ9‡h(>d3`H"ڷ=oVKBFLgr)싘 w1\S-~;bJLG\oE3ϛ_G\FHau.4@ orzOk߿l.4*&d,Bc"N # MI{!@>Xn|O-dQkrLV0Ϯksܝ-wlPm߽}_{Y|KĤ1t>`p7G_&wQ+dߏLe=ǴYW UA*W \p>=0*M-4^Qm*ME6զTTbS1TeTTjSQm*MEզT5*ME+MEYTTզTTjS\E6զTTjSQm*Mբ*ME6զTTjSQm*M _K-HPV>NgNoG4k[ExThbk}M9]OsvZ˒Z%}*\R>,SRNG*gXL|yKn)k 9R&'%LUXqn& q@cAln UPrl@HT}ŗ5RkA0qFZcn>]Plk_yAŮCOo]̷ k3W« ճB/֫%+8-.?_L>zO7Wm5P>5+ht?1-A%`GJ?a!O黷tI.iOj77 *RXSl(IV?1\Aa9*w/l ȁK÷Q|Yy]:tɗrqr*hݻywN^Nj/[]{>г[`rnozKL/LS zQ{ʷmL+tT:CWT9 R{)4JI ީϞD7eJuR?KU/]־XUrlA!Cb5.XK222GM"^ڈrJv^>ʘ;&el)1.106%k,@5\$[}l'ZN5K RG2>CV;xo`7|z{7* [6/y _/S}c:ۙ߼ Ƈ6ի) yXYbJ۔ *&L$X&>x ؙ *f"$N8!D*ZecٓK̅Mu0Fz+sf ?Xh˞/|VgwTZI{x/+;9 |4:ibqy.kpաvR@>}@IB%&<5,gsoU>*{S(RhKq6\E rZWp}ۡ9v${w=E~~wXo&;`+A=M7c 鱭Iov֚cݥl3ezlg!To{J[8((|5Mz M>y32AHƄ l'ϹSީgo?z~N=M!V[F+6kzD%Qp"袈2IJS `"Q)B`% \cY"s})MV|P)=^r0L ݇ђ#Nu~9k#d'S֥vS7~9[d}GbwC-@$Oa 15՘ŊXYU$ ʫjcɡoiaNc"I->U]-"\ ƞWaϑ/X bZq:UdlԹyQREfz uIQlg9Ζvv(j/ot HW@kV @NzUfi,ӯK,gp+ 9-y$}˪ Z 3|I3O3/0ƌX(%IOA,NS3Y1Z q qGUWtN#ZkbZ}`3hw`l˫Gc4[G}br~Jg[%:82˥[@gϭܸWʍ`d}O  OI' nRZ))SRѥ#JIsJHцɨ^9K.6C1V qRj4lMa/8M˵x0lK"0e2) UuhM} b(R- ƩُSB$guwCW(l2{ bR*1U%e^-+ͶE":c3XN]@R"JIߚw*Nɹ 9) 2M9(qk`-x,%e9[Qf\˘/@HrzIApҐW_z`Zmu{Ld1%$R $!KhS/Mjbjl*B95vTig;&39m%]+&#(e-5k G1hFMs#k^lF~akt?}mhvp+I8-ާztc}\|ĭɧpd L:cr's7ޑvkl-)9CPBL9IЩf(57dgql$RDbh B*^^6SkN \P%=xpDd8ʕo>RyO크xkvqWR-_}ݍ;^[uU^ږ0v-y_|{Ar^2v3,|ЗW71.QBm.!tU:=t#.ީ[x+$Փ^~z8/{fΧPV[v.?mF޸r>\_^o}v7{vχE;~="./V  JwީחJKVO=}^ޡSM6sxn^[a1c-MSu'?!gݟ登{.ZpML}RKNk2;h)(ygB ؜\[r5FM>4/R\MymV :Q{Aem<@~M?ǧ}zu}ǭ%F=0T>|X]_֛?:z0%@jF&Hj6\Tm]8ChTr s.P=9pQH&R6ptEd8-Qqmݹ7k_3s+*We3"4|z\Ud]L{@(sRS!㦒O٫ fgBk.;{t2@-V 3ؾgzMDmi+[50ji1~:b>Vmcĭ.=+H%,pbԟleKc+}7LQeSbX<$ÞLRJytSQPX~q   l 'e\g¼`#lPRbPiؔdϪ-ZB( 3c1Z#bmad LVz֖"DW/Mh8L6D) i  _x/snɸ:y`&oY r~~lq{M X8Y~T4b+q1yhA[RmߞtbbLdsMM3 *?K3p&~Mu\O8kgEUL1"%*LK!)}6^>bP3c@S^+CVd^Y4;IkGQzPm*DTF/!g?\JwS=bgKUS F软DRn9+S={!310K*K \M޶gz%BrʤOTxta_>jRXo9X0-y_tf_o_'6@A jGa+@ IRT˕$/(x? g&؎]Yqs:+C{تP֥;3VBN S>NbKUu{OˇЮ'K6%*+Y!X=ř}\@A9nrak\ w\x\u;=@C[^˽g|ko RNUR =z ULnۚ\}*$͡>,)M)^ =P1gLV^  [_@U1ky/]Uɘb_l !W]Cq$,A}7Jⰷˉ"ǜmWBeFgk[-pcH1Z 6ZPL͚T.5Ax !5'&+u3o\bmX7qvZxՆpbz Qoܜ]i/޻S}o-NT.U_ۉZiNy\T}6կ AxN0 }>;(3'd}-tZ?gʼ{sjjb!ō:O.ocワ UKw9\W1"M浽Y5B_ΏW? NV4/]#l>yQL^|<^/zzl0[9:[%h}NvĉluNBGR_jd6~R&T*`NV6+78?0G~xo?zÇOOݧ߂ ,8p G]"pd~} kuET-8gM,g;ڽ>$-mtQۑ~0mI'O:p"$W~Y)7wn; Qc:.H7>h"pߵ'C:ۉ{6xܥi4@_$i&HGX3Y@Nq Dp`aXd[f"w6n>8"s#D[ҵl2q J (Q LH(:YuavqޗXlsbXՃZ=h6*K_R$^dC

h` |ÂO8 =1Yw֯iy9,-8<-0v7z]~|Mت'4Wr ,^i؁.&ik|_Cջm_UP Q_LOƗtd=w_;zIg/\csդ Hs +E%Y꿹)B/\%q1y)*IKtUbAsJ s^J2RUV$,1WtWO#~18:zR\=NZc%X0WsW{[e#Q_UͿoXz-Kd0°W ^a+ {a0°WXJ°W ^a0°W.0°W,ba0b], {a0°W7>2o Va04F›ؼd)rY0f,W \a+rY0E*rY0f,W \a+rYn\bP \!+rX}a;D \!+rXbB,W ҢR(|T J.]3]m55֖;:8m&]M9ړİ?A;ɲZD{,JU2pp`J/H: y!Jєim<*kY^Sgs7޾|o\+t>f{>7 :YM@ q1aO2PMS-ztC2vh6BJ.Rl_WM͕ϙy0ԷxrRy*&qpwzHx95_jL}s,& ˼ieDYi0㈐Cĥg%i9;qVRB ga&Vq0s :E)N7NF"kO=ց>r=-'jeFl_[[4Z0 h)# (I0b1G" 46ϺH rH# be}0z)#"b1hpHgj&Ζ[k\ u8> .Bvzo|m.fmm\T?_A#)ZDZa8:P_P`!Rp)Ӗ@jӘG"h@QAsgDR*:;#u6q6k8+nn,X*]6^S;_늲v0o|9Z>?lޑq HpQsBjZ) +h? Aeu:R9dсInrTOp|r ۨٽӗ@&dv꫑_ ie#AE/g8q3SX/?%|6ڝ/%u^ȼ UaVlYћ7,Yydc1E4^JJ8%"e;bi1g;Ja8Co/ rWA;!O-_NRBu֨w-Q 7Sˎ.ܴeE3̢E1wVEC;U♕bmoX'|1uLmW\&Rs,Bf9l\yUbdlK%w?>b eJ_30S{a>V7]:%=t1ֆ%,amusʠNy5i;MWd{-/.x Vu":)kW5gpֲPԴ8Oy3md)ΫHx' )Aczi5ᢣ y©w*썎 7 ρ.0EX)ucPwN6V6^RH1CKB%B-C^2o/oo`JvZ!Kl89-jAƒEmCp+L;EГ.5?go'3IG͂{%BAQ|4G(Rb+R"5:+hcB)7j< ZGEdQ0A[R.k%)$17h&fv Z# tvimkiG_gƌط> 9QDǏ ItpD,/ɟ9GD:j.j+ZK¤J:%vIt{=#XPQ"bu ;Fa!Rg1qKe<(,t}B IgkkuJx W%3$iݒbb*XBYrWI$w+gmjwbְըrÔU4t 5Sbp…=)35n'-Nv& <^!rn^[hmecU 䂐 B!B&R{.Z0c"&1chPg6Fo1.tDDZ.#Ym T(sVAvro0CJf/ȟ-. p|[WszhB'-#úg?R>S|Z@3o/@L45͎{KNx5,gizp|ӫru!%cz.44^x5sբq5Qw>MMN8'E~h;t$j,si՜bBgix`,(U3Pej )$Ťʧ.:]&}LtYP%#WQ@pjA@ k4cLp9P>5=|»o+nbjQ|܀ɚLGGss_5oteK0[#$xэ喩h>Ic0dp0@%N;n2?vH(=:!i%mh>c5k^B&T\(wp)BcF@1gz4‚S?X][ 0]4Xng}ꆔ 7Vކj8N^WWaT͠\W|h Uo_L[߽wh|5ԐdW7R?w@;x0 ]PP'=CɄdG֗Q!fz^zrJbSeě*w G(i `&4}9>EZC[3P`]\c.Ip7z"b>RD;Rd=:Rd3ۜ]UQM D"RFZNCPT IK(=\"I/(9eEpԩ5M=A{$f.H\0CQPh[1hn6]C51wJZ'$&}NJ8%(gsBD`U>0+U JB# 9`Ť%|wdUR#*#+TTต6PʟYϲ5zWf 8k&j T!cT?LTH yH ]H QH )!)(DZK &X` L%XE\ "Vߥd Ĭ8աa9$LY|x(5tH(ı-RLp7޵u#؞0@nqr%;$[N,KhKvXs4oyP6#g3NMZ|rzHW6[! @cWW#I[ddf [/J_-vJ֧1)8EJ;~kZEd:XJa`C^dڧkEv ]RGR9ϖ_$nB>}#>=9fT@VB%*Si+& RCˌ%Ǣ6񗲭 QCG"R(ѕkop!zQ,u;$ްl9BgU cg5/k9+c%BŘPedlJoU>ڴѷmս{6vG=[|ޝu B`M=Z'=\LO[bٱ{P$A+ */,ONؼ< &* `qsݲrO]ou'DYOY݃.,l;?~3z9?/e:e2ɬ;~^9/b~_{%',jDU <曣sk4/bx[;`?|XE/bPJAor*j %'ۍc}"}@5]XVβ²k#ܻMN/ Q-^CZm,63vQ^'{޼7{{޼7{{m=ly{^׮+z^g0vE^vE]kWtY+z튯`;P 4({~kWvE]kWRy{2xR='d ޓ{2xO='F:գ7Ӷן0b}w- .faV׊^tgBW /QQX*JE6䤵5\J!bsch[{&gd K7(a}N=:[#Eݐk%uߧ'i6}jp4Ϳ}(On5 iA<$~gM}GI/D}-c=߳Sz쁬v#%eRY2:x }֙XKʨkT}SC6=+=4UX[Cf>ܺ~(e[Ex[X0?&` ϦIE%siRQŃoRQ7 T(t=SNsʭj:A;-`wk:^.}@5tpbVd ! H% h*KĞ$K2`d1@1Tm]-,i 5DT=R:I0:Ouz+r6/TzHI|Iۗ”9\XI!w=8=\Rb6l ,Hm%3/s +L]\f;,^w9{me.MHC("զ)*)m (Y%('^32je]4q$P6%Y jp`o,gRFȨw1*_>TP o 3JFQ`rԐz9B̒-`~g"BQo]9I㟔J(enXg^)*oesZAtq'=a=o)KPdXT2y,-VFFG1b'nX+6d(|Xд*ތ1N+Q❵[,gN8͒gM$b&u`d]0mSj١_394F<0)Yy` cB2"N{kQJBWuP<ؐ&W#8aK7A/gհp:z!bM >8= Hq{'buo)=At=E>f뻞}CyR= b$5l+~M~==>ʃƶ#=!-4(9pyʗ_ -{?hWk6ѤH/Ný.I2l9e2؈d<$ ȶBYP*$-E5ErԠbN&E]U20ٲ/\7Rrӫ =I~'Wo}qkm2[5n]י:kfanܟYnx o̵|q-6 Ǎr6ڹN׬"X\K}EwIY& є,KRS F$嵵: I0B'I9m(I D 2 ,7N*5hrtN`jo9[dmFCk'[1-r'~]y>7o vJ;$"\}yl10 I6͢+v29EEdCq(YXXZUPXI֌٬G){х8¯C +;k/n 5!݋?0o' D+5XVR{Ue4"(T%c%WVe_R6u!I]a„돒䰱nFf'gyŹ<݌;Ʈ;C%:кd*6E(tHa3W)=he>}P 60T aif/:)rk2Tkc YG!b2{B#flևSl_hfPkD5b׈qGPa 13˨/*C)R*5RjS,m29?͢8dz|(ˉl;~s,w`4h1 *tVZgHJq?RW2^+߅Wx7$\*ɹiY(M9 ݳOF(_]q*XAy ;hyk k؟ƟoG%͎]o;n'9юiP*n dODU$eL"DC|""G+Nܫ7,?!'mͽ?UBR`v'*\X\zf=jR3 ZvIϞ5j!#Aɰ1XdwTJPj%o#+w9D8'+xᴊ蓎Fda En7#gkmzH Wsrᒍj+Sz-gŬ$/G@h J3K%_ i2`HA:0Y l/3󮷢 $% fAwvd@Adt#G_Nt .):]VjI_00p $K?B4L)t{jw.W9k !=Q 2Bߊ^wW<6 E{Bhx<zQaI8e10"Y˥UʟqF*J:zo TifF]W\>,l3g%$  kh#pCjR@Wh5\0E)̱߃B(pzH%N߼8WElTdzO}ҬTOch2]!H??>p4Οh02*m$(`~{|7|]WCx =zS'|qe2w)"i9n̎R6S&4EDa[/&yģbpq=IUZ|&5r٤.z*G&%ţԈ? R;q&#<5rX6_xt՚ ^D0 \ H\5֫ qƵrHc'/bY22 >S/6,D^ݠ[zqC cVܤQ*Ƅq!0,8DLe'dTi=Ntxv~җxou;;Ͼ.a7UCvU rt_&sP coOSy ]4?߾E4Yo~+yѥa'crE>ZU0rS? j5ǥvaT,iɺ/!d'g՘3F"@%ݸ [UV7+Sێq{'M )k⍫/[NF:;k)wC"xI P(}(B,ECeI"1'a1:mMbIBW$K(~{mM i]V<7՚ +uce`| &LLgP~F\'fQkf@reb x=*7IPxd+oHol߹FZ偋|q\V(w˗PXӳTMB մXrYn^T{޼g]n>Qhi:o0Y[}yl3mL6yK̊,p3গ#1%i0W -w1+຤f, $(x|]( BicEw˔xY>hQ0SDwkjy&_E ಱdހ0k\nX(kSӽbL4|U1ګ STmDIvm6R2w #mۅ+7 QOg<\ܵ%A݄yP-g!Y4MiFOP1{gcu{ EsuީdzKW<ޓ.teu#4/-82AV-eQx=XM쬖aQcf:tqgTZQօ.5ͣghX@yQVtqxsQb*&JaHRz(K2!a:gײHV:Z 9{m3X= Mv!V'FJ-ztB$کߦInd}߶W?{7|Xryѝ?|t}c`u uBc1SZ,+;,b!X喤BiXޖL=UrWtn'ΩͨH,jjji$-_2) 06H_rޑCU#5w\sqV0&(hF$ D'aRL=d6 &jk1iORK- . GJF@)} ""CTBwR(ckB:-(l@B*[&RS9]R ]b̀yL:sLܥyCZ(J!wͩgg*Z6`$$ѢnNFKUoGDR QXNMu3ϲ bd+Mtw-R[ȳlIq[ Fg!WI?vF-=|^e rDe6H X,7sA2G!ˉ[1l%O4~zft )0j0:3x?Z!k$jg,(`Bʙ^2), ѫ8x⅔Ol7V-MԴ$Y2;u*2}LNy"mBAY hnEh+wK`0i_ |z`)F=)&JYi VJ#֖sfȦL,6Xci7C1:k\=x7Z{Pi<J7D GY$ & 61gTq,ݳ8PϦpֈW(TZ5Txuw_qZZzn"c d͕#jOK[::ѻ:ޣ:>Dе8M&YNsU\}K&`1Cs:YGNȲK^Sj;Oެq¸@kLr"!aHtfVr'QXikE\֚?M uAY {;5"4vQ(˕1s4֤$emBNȕ3䵆dX]/AuEΆ7A)#A/#nLdg]" |P"5[LD6DC26i BY: +C`rSR/LX,&{J`|pV{azz11aˠ5ZdϜ Lcp$f}tY8 yȜ,!D+A#hN1 D ȷKY[ebbа{ ÝcѲ?twyzXUQg'K̑RT;d^ҫ!)H2% y PZdH1 {iOMY7wGVj t0dR)5<"8x6!%% uBvC@+ 1CvoZ9E8HVRY[m-ыX9ѩ4x~V" ~JuJDH 7he_ɠ4PN+8t[5'xZ{n>_Ay  I{GOc|?S)Y{1)XL ORcy$CV 60 ۬q>Ν͗]2OlA_mteK^0t+  A?ؽ~V6rZZ. ̵T:hUh-;8š,1s+Xdh,K*J9GuFz-st] cj)ΖBQiȹ[0R·άI>qt3lj*:ڒuysxcG8X|Щ΂eIs{*E6d%èuVJ#\(SB,r)Á4-݇4-9);(dZ V8 RR%sFRYg$d6.s.h($Ƭ I5xzPFji' x:l8q3X7GA$aHx.KjZ.-̠%3AG*J:zo};$O "m"wsT\ s͜iwqF[j2$hTĩlK\*<#!mq̃3CJ#Es 4pOnbiI(e8ؽ!kHOXr(c):ڐ s$jt2!hjIt+jk\ q詙ʻw]^"ru<ˬ_Խr+h0',;xn^?Mgo,8Id^  _y({|H@OȠ=EF uwt*pz<=8Y|ЪVbT\(?_+egխv\,oUt;?OUA 6N4'3_:: 9K/b{tu2=,[=.44]P>ۡ]/o>c1->?Sk,E'>44?>k'?ϭws{/9rpv~PkٟXUVvuOCn澲rv>'`Nn=j"^upjmӞؖ-/痥7}=va|ٿ>nFt?>(o7?w?!޼7?~ <0WRY!~x|__X]e&]u5ńo/9~?2I΂rrgdO?\~{<#X.x;g.bOᏧeL<*^93 1őa~0[xi uY Xm<7# X}1R,x3TB\D*!Fhg R"~J\+ʜ1N6,!oʻ`/qޑlvWf >JcM%U߫itottFY٩K{FӉҞ||}[CZ߶e8xۗtƼ _Υ73]Zvyiwx]Pw*?iE6佱\b#X-$|A]Z]1%Eø'}HqBEjf21 jiR#9y`4q\c.Oftuy3[X=;7,q>]%|EɓMoWr:n\ݞnhzᅷoj py5Bmz^l5dr:Zto]B _ZxgOߟ̟^MՀ^V5_>NJPw|:C], </СG#}lbsޘhwM/vq)/1?ѶKyh;\.rSyi^Py~ dͻRj*M0byelj+*0c*w"H`M'R͊hF[!T[fC3bzqQSu696lK3rLѰvtt:@~劂MőSe22yK+4..]g\oWЇYR'{i׫_Nw)mRbjD b!dXH/{P"ٖCNI$r1&Jb8$@KBCDkm3 FՊJ#[hl<+mlCn?{OtKt)景b3[F~}M*t2[)G6'):aRf| bf'T esJHt.0`[cmvpjMk)q>PVIlHEcc+c] ށ9. ϶[1! ؏"g%JBR&F^ע: V@^,l [UIP05!0lk^2ucP-rp.ZMOxj>+ow};v 7g HVk)k>}@ϛ+\M&b`NŔ2[|RlҀWZ_[w q3rjq4q6#x\e4,M3X5_]wfwWG1%%F fH(*8-:^"_z%irm" A|{Mx4q6,_ ݁qWq4x*"Ƒ1qkWk\ QآTS 뢏N%%Q1R X,y2ճngrY8qe!fo'.:q&fDUgtyoriqѨqqq8E1\mVk ChQl֖W[ Ykb- IXqSάOCX@؈u^2Wuw&u~4Qǀ'p_ą;ǀ'-/s qRj[ǀǀy xW݃WϑF؞]4߭ݵ?Tǃߎ0b 6G1=ǽ% _:{GnK+pD_ܣwDRw-w G=:xgҳեE}|弤Tz[&E;Κp{+TqVq`O(<;4G I~b>O;M JSuz /zY>Fή?鹠%i-d MS3ϒGB"W(G!̊`,?, ./opM=<r{(or|VR\NNDEHZ1YoȖ!b6{'j?ĕSJ1{RwhC짭識tFsRWե_ol`~jf\4U*Hzev6T&ɲԂ1'ƄAi' u`7cAY)-6 ?V(Ր.fmYUDs D&k2 qZI*[&0fƦ#h38fַZͳ#\( rz 6h4A9~ϓ:RLCMZ)ʕ<y9Ye]J'M97* o3 "xB. &ւo%6;٢ Һz/A "luQE~"hwruq|8d1 ow1C3P2d]șcqb_ H}9(u*NT4o-Ծ˙*޶j&՜ȵH9iE%x 6:ljaL<D*i#BK}c ;͙+=#3_;M`U} T0USNOQਥf|,7%>{-R.Aa &`ԙ%EMÅhxGmOBR35&Ph ]d ²- OYnUEE#x4sn 3`5GejD9qjeWTX <*Akjö2ѕ<$, qFYGE1|W SWg8JX P)l< kjXbva'ŽqJa`eFbl|e +H2X @ CSX"Y(B3mؘ\m|錸5gxf#nT4vTr ``.^ L^YIA1,'D jXX3ȁ׃ ؛k(S1*F ) yԙJaVcܣ.P')ՁkUkaPH/gNk2H-!c @ ev:A+au]AԊXD|p.Ci0Aq˞Iv3RU{Ƭqs2|=bLQ|?od<̉0!M@uY; v;\O7p/c[VЌVִ2i!`IxpI u`aPmK`"kJ`"gJW-$@+G4t<" 'aaGţ :C/cRsI >2<.2pȚE0aa{-q{KH@e@Y%cح)}x (ngp6 ^u"2Y)J%BWawV k Z #.Pݩaau1D7[^(dPht:~9؃_fw)qLlnET+)vaæĮʌx#3 XC-D;T3ڄ`roF+}"iDiuDhP:f&.h]"PNb&#E4P] <(BHo2"㬪ZUrf4$Y6 $cs:QbզX5{-6'taҡOI#k4ZIdf%"+5.-MUr,}j,e4CIX5sJ٦}A%E Si|UK Fvc=ރv#7v0,Nry4O.N/ۍu$K%٫PCnNN›ih$tiѣJ$k M("mcME\k Q^$zhx5!7v)oa#ܷ*3bO*Vh :%xKD]ж+9]nDVݮvnD{=S"ʕ,.TP= RQ*1#J[8="2PAzZd}c%Ö+^B|_qE!8ij)ڥA$\s[9,BA ZykȅIf(6+0R܌EEH,{aG= UG o+QYj -S{N낞ٮh" b %_n^& &#m|aMBvܵOi8?v *UUMj-*awx8 "ӆ nqhU8-]Z4kC~"уL<)=I=GwXoCnhM5r7pid":r( :fB*d.XztP%$Di_FY~BufW4Z;S9;:^~q7_\\oploM8)PpL׍CSe=Q1 qIN*e*\ZCl~}IA :N! b>rSZr^Kٹ}7`mB&KSVR MY 7k7}.UHLdXswA!:~\BX轇̎c7]08rӵQ-2/Q^{GQ30\{+}7DA!>uL4 vx h@4Cr4!A;g!hf!hf!hf!hf!hf!hf!hf!hf!hf!hf!hf! A#%H @0׻!-;x- @/^$@L1 $@L1 $@L1 $@L1 $@L1 $@L1 $@L1 $@L1 $@L1 $@L1 $@L1 $@/B8GUǨ!`GCZ^2 H>݇[bv؋3ְۿFTkfoԣu12M!6Im jwH_.r!ߟ}19U[u}n]H\/d+*RjtʎdcS$=ߐ#%KU\.l4SzCR%[_%[ TXvbKn>+f~tT7ZN/w|guKW .?_x>K;7GUsѬB2⻰y@6g؜as6g؜as6g؜as6g؜as6g؜as6g؜as6g؜as6g؜as6g؜as6g؜as6g؜as6gQK! 6Fls摢2 _`I*e'bI &bI &bI &bI &bI &bI &bI &bI &bI &bI &bI &zI$K{Р_'?zNKM7m,_-wi.O7C 7K0)y4局Kd.X%Sj#6Ө*Χ_Q9fP/SUj2 B=;(1Y&\/ e^VHdqzÛ95;T˿qvJ}E%.1*tsH߷*/-]UJN:"sX)t9W [UNsOT,4W,t?NE<,/m/J,_#&_Dֳ/~UoX4¦?>\;XІ]Z% &{\W }5ewBI<U3f嚷5Z?YBJ k%SSf)%u #ȁa9;xytŻ|gçz{]yݧ"|:.?lv:m{kZcڄTr$He7Yj\IʔɘyxwVÞ{<MvS!֭h%iTJj[b'k|6RyHrFI 8^?8!"Ukߋ*I DU:_wlYGa3iHX|roGֹo鬟J-}cp 1Q_$Fgvaq =/0r^ĩ`Z\wƽmIa '-lUi4Xa-XHhR!؍rw&IiJCpLYz{ӌW].]:fMOl@c|&暢/8<ݝUW_|bOϧzzzy߬m{d8>Hf)H?w\h-8VOmgܐ#dr_}zrPO߾x ߃:1d0BUڄW&$등i{Zf_X'щ'6+47`8L?r[!W҅WRWv͖uz;-!sVЛjZ-潭Ȇ#Rz.ZrػqvZ_ns;]~z}\ʺ2.=00eԿ_"mus~?W>X|oӋ#R?6Gտaq݊[g?tzRrKT kvz@JԿ˅XiO,WSMOfӭ6,wf .T;d}Ѷgպlt&.Uo)%lO$O{/e* + f@;̜,!߻kq?bl~;4{5 Y`ǓhuEf֫ϣW#9 hvN!s?S`wN!+76S^)jWc7sl|VMyU7'?m5/nv”9o>ON~O?\]\/xGm;I5iq"#v^p[ #sDn vdzf+6Cwde4_6K+kh\Ґ~8}2“\,~Xͷ=pk0 _5ohb\ϕ\ϕ\ϕ\ϕ\ϕ\jq^]uw1ِJL־TTK)&dP-);6b{}Q:9υJ}jNmD+'$9Sh݊ElPўh2IEtW[*FT(Նv!fw6ªw:oܱz?Wo-0)U}f%PpduL(DÓ#Y[o,5-`3M16Nood{JrjRi/N]Qy8oUר$=jC<)}ym k:Ė#oKwh9m hKi#*[IQ6c5O. I 鬎1xVM,mC?7cw?޵9::{N.( ԓ51|urg߇/>oaۮ?oq_~<1mm>Xr4v)=&}IɞZ,VIO|>LiCg˦[=MRyZahɕs;V5x dS@^z9^" 7'z`ۨzW*4՚*UUTk>+xBE .EMt=wX=gz0׶͕\2gl)|*hstPE 5DO+\ `|VмBRTwp.7-a<7Pi'oЗZ+ T\I _)5RIх ś^u3Xt]$s>|/᧟׵rRA<~SF`ep1DG$hMO-箋cC2TjQq q6##/]$ Q%u;W县Ҡ\HL MBm! 6m܈l{&0/zq(oX9dSPP',V`7ѹд6Zg聱=nT..<1;a_ ֒^-ep4rSvz75 Z^Miх#KJ%%Cx,{ɐ^2d K>V'|V!cqBIYc&# tr V_An:tWmI|Y)K)dN)^;)d U.޸ETLM&#E^{] ٬nj̵T)zYUM*98fiu9{@F uvZ]٦w>e%Bs576XWc*RT){v!b"AgZȑ"bm7\fv3`GږIs^q([q:@DMEvՏUbUBY$&ED'm!\9CVkHAN%YU-QAzIIh9Dv%_ BJ"Q𖀖نn:9$ešA3Ucb:"Р`0`)% mX,&{J`|pj1(2O2Ɯ"2h3g. 'dc4[42DJH4{#c,y{)dN\bU*8Ӌ':,+J-0Ni ӦXUQ6Yq.2Ch2GHR㌒M{K(#)ʌp./|P*L[Hu}gn O][$sk^z"J0֥/pBլ n,ӶESՍt^pt*T&D@*$\0tm S+NUY;1%i0.1+#Zy3#&KW>gFf_a8CB/snv^hγN-Ť$Up<0&L2N;z)m`AYcEqU9+L2gwi}+nVne܃ceҾޱt跽}DdV5Z=>j|)a`8~05|Ď6xIl$0xd1BP7l"y[h Z|H+brHlERA(#&%r.uX̳T"P׽i=V[fR$|q %'.9A>wDiɞSYmFΖ k!{4N|~;&)|*'!fXNj-hC0 1#$k%ďH`IR0hLؠo_ =QÅ?1%j=q[R{ ڟ䁷ok%5ncw=4_s+bZセawi{lhz8bZ_L欽hT3T" FVrfm&#H|5mN^)o (bT֦,`+ HFjlqR qF.(*s^xsu;n\ObWo4|~dV@V6/*9OFE}Ji#Љ{>ODRLd̺L"GY42bW#g3bQeb jW㎇ڶG^p5D%(fdd5iF2_9w $v^H4IǒjRL!H@CXRFM#a}C*i-_;Z5r6ak" 0 "V㏇"Gq$ϖwpnD}q "3e[r%f,r)}zh'JڇgLJ+^KW|QRIE . b%#l*U)!K&ҥr dCf۾8en#71J(IЍ,%E,H/uREgM'G-T7_/ Tg x#eIﳚ-֭b0$ 8KQrIvscPRəvkzJhў|/3H@D0D稸&bg9d9+QhH6AL@JjD3E)̱܅IB KzH%N/w,r?fLXW'>_g~~URCZÏpp+bğ0e2rXy9? N(ћI)~"!P2%;86Z?y8>L 5ݳozI\ #c<+Dfk R+ 7]Ӻ QѴ/>axZ[)^.} "mE/症?j8Z%Iu8=ĭZѐ?g2bjcpEI(ݨQenܙSz0k^\kEIۊbH4ndݬ-`t1jf{CX3Ɉ#I|U8r6'Ze]zr3|8]79'WQ\5ʹ2Q's)+b]cj<(q09= MTOϬhx'Ď"8:o,?߽>O|{wGћzskqD+0X"A&ᷯ#@ 7?Ю4|haCk9nW֜q?Iq[cv[n+@B|(0xrrkE<, WQWlY=d.Zv*G* Bd#%WXN^H{}1ej 6Eͳƣak)"fz`CC2Kd?yH˒A0?ڵр([$ [g$kzZ^v=[Ųeag7qZlf^/>ju:ff0QY j<% 0xOq57{j{?F97/ RK a&쳿zϵp -sj-})dIWM>MDFcJ6`ByH?ݕGX]^$XɌ̀2YzgLV( C4 IRfRY,PU]\ R(*KlHHƑڤ3pH3ѝjFΖq𱡭>8?vD|>vMlzbjd1 48qP' *f4jB I|2&2璵+ B$!0 u G*s< -:=AK!mFp*GV;_u5r6KR=q6U77ϾY]]mk];žT.>nվwLI[y# FX+ȢAsM'笷#wdG=#>|q\~n68Be~  )iJLҢ,cb f:: ϤBo7nn7gtlVt &ďy鰵i'{e}io@IG| Qqa "C KSY yHJ3<811hG $l{D "^ظ$ FYDH`)脉-ٻ5' I R;<ź,4" f ;r0IpE).)e3i(§HZ# #Z#}puL[y'Ǔr Sw{kZl,]MQX3}lU'x\`xBP(>}TEe X=1`KV:s1zVBO< yHdܠb)2e¬LRQzU  TiQ*2%Mz䩂\lsiw1j&X+KGT$]~&K=u +B-nfŗ|%ٙf&60%[cba9dYjRXR,U?VUAj$!mvRok 5[:kAg %.B49ŠNF$e$+M[zr P7JzC>v`BkaŐ6 &li]pjox73Ů{\sK(Q!XIlTE/ f*F[aUg4"uPp 5j0XUqcD;+_hK۔mM:*#k!jdeQ)MΊM2-i+mLEi[@.R,A|2jk3ʘ `1!s5r4m 5r@G]7tH+>=yde3L{Hq&Yjd)#m]YR"yP-6T D*[*DpqɀNgU:U+ӕ1, #rP"BvFep&Α> ;2z)  #F_[LUWs tߞ mS1G4`=r5g@!}qخ; 3:,`(K!WU w vp\sՃ F Wa\=JWp%;tK%U!p_P+v \@0`7pU|_PQUuW/ G Ő2>n:\5έQkq+LFد=&8"?~hB}`c7`q 5ޫ0׹(! b pw'Q=zv )U Mhyx`8gLMLRE$BĀbEYi6&[D+#=b(mu^Eǥ)`@#|8oPz^}ֿWӨ L]wSA1nK*_^^UB_.\t9ZY%Dːf6j *+,VVK1d9:""o|eNh* ҢKr|Ȓ#-W{crKP+wޒ+T*Yr/ВCJ=+"X<r% \j:\*E• ׈켹:4һpuaK!Us:) Ƿ{S{v6Cv@mFk;IМe"?O.\8[tuFpmOvW( Kb> 0ڦRsUphy /;z{F^`{h*ktiNu 9R‚ e%Wjo".uE$!vlHL.>EB]$|{v'.IL0Va4%Yę 2KJD-e&$ll9kWvZRRUveI^z\fF{!Vj|:-N.)>]fL>Hp9@z^4ƓtsIQmNU d;P:pI3ZO H\RHr`e n%N=hlo嫻®#,xG%OCs~5'U.N 7hjnL>  JkQzmy式+xeyi5R1(7}2q5H BJX9;lfZ7[krO`02%e (}"'i#,yI¹T2td"MC1\LxOJRFmd fl&BG)MP"~a\01w[MQ۴ڦCN]Y7HFi/x* LRiS hd9eJh'Fud&G%B& &7AHEo8dak'9acʲ`Dl?6ED2";D\:JmQ)8ʠfV$>\TV[} oH 8I$Z F HbI%Z$Yp0[-#bk܏_N-'\pYgk\!.NVcpb!zg=J'@z.iCƣ)!0p1pqg|?mqǦx-!la[wU#= 7>A6ag~4bfQ(]|YMo(OmU9PBsQG d!ׅ>4$5Ԓp؍pMGd)H:!W,?-l6BdI1kQ0gZU{zRkp), 4q(3x!HKiQ[Ѳض9kF)t"ѭ+~hKiʨ mˮL?_[E٧SYuɒCbɲh]Ѧ3orS0kLmJkK 'TN +JsNh#Z.!BdZ mU\%m{venen*qI5tE$"9>)!}x˝ 01nɑa]+F!yX&| Ir$vZ=aTj*h)XZюKʎ۟jkZ˕<FΓR蜋eM8pPZ2 beN)5OFQ% 6$έ4.`. @xY&"-d\] iҰhƻpV« GӸe0 nPN8? o-(utEbH4}r:[avp7 * >]6I^C8hr0?o7:9-]Ҩ"}X8Ď!8:~/˿9~1ps 8@2$ߚlGڂ[]tmu- a[N:bC^-)4whr6$ħoW_ ~J޷xGIq(\g_!IE^´}G%yl OQqd~8^MMHo+ăl$xd7m$1vIR|b&ɇ I9JDc't lb 4)6*fC&J%_=arC2W/ӴP##xzEAaoLϕ 15 Aw4 w-dm~ G_K4B] UZ=Ku/?a -[`(  -gշ9-(ku[\ݱ6ͶӴI( M5BR EfwZ &F%9>kK+V!;[Ρc3{iwޡ:Hz'VNk!#(oJ(N'Hh %($߁ ܴK[=e]3ii"7mswt^`_zR .\V7]Z=3A! =]t{?s]_=MdEO7>5VhyPJ-oo갎1k%!ixZ#?n+"X9u#enn>ix-uLXnQ3%~8$, R1/??[cS*[*Dpq$5Gt++}9YAFDMcIȇ^%#l"xAE =Qv3|K \ Ȳ^gmmFPBb`1!s5r4lF9!tu+̢P\?y. tmS1ITLj]}.f4l2k5H+>Tٻ8ndWz:"x1`$g78{c}Y x5:#^g~=32HCI=r$wE6WdJt*-!H6lځM;dyo\h~\/' ځ! MQ*3ku!e E,% Z *iF z!e_T2Nd/eVeI(E亙8n~͂lϯ4$~3K8: k6Ny<-ݾ0|FaxvkkO7ݕ!&m0"D<5.gkb0Q&B CD<2TN?2|iq(QjU \}Fp?_kiJGkށ zm{c;wLIazn@~қ}jD]ywwmMYWڮaMzӡS#>\Rn%C7T?OGYǔ"M몔ܯ;UntW[|<ξ@Hs/MONz'_oq{]E}Eܢ9~fҺF*xcSiɩt|x%x{joS@Y]za˥-~.l. >H#k:h3?ˤ͔wQcOdqڝG簗U{]X-qY[Zp`tsBT'kQ(?_袇lQ$lV'qyixy43DJl2C)jcUJtգ%*H*,$JAzM@XA%h66c$PJn UC>բ|E;dv(-YBN0H"J&+ EK\\wRMagӁR&׃"OgMݫX7ѾyF mH=+ ـ8Ǝ40pKeu![Q Gg4pQTOjK- "BV9 :x'"&rGQtj̑2Q`.HPTȦƐKZX6"bur$e 9R3q6XW2] {'e'#9-oAoy}rcy&;'Ǒ=;#,<;itz׍qg)0H;HAD%3e`Q"I2lG K? &t, kC#[EbMz&-( 9<՘?B3A .aʬӥaQɢI9g*i-E[G֝JH\ϓ>CM,&_sROL.?͝c?{uwݧ8~9^sTzx}/Z|ZJ5;,yoX9z*~Nz9..osY|Ḣom o~zD>TsX(2Clw@Frр14OS O䧉4~31?`-Z:eFHQ&*3S3&23M( 9$+x&1cҞ4JDNJ#3: V@?uĹ{bnFak+[NMm./WF*, 5y'Q(l pEfNBt/erѶ(cAǠI2PHQxv4It9X-^JzOW9"9S֣@ZD p *^1p{+--ۊ̀7J<"H,hjL13OٱfHI۔y݊iz?PmփH1 qKAG0kZU "` :0YlC8'c]|v!j-.~:kGpMn.E3g@PXn2ꎲtHMD#sgEjg6'F⯓8u?Kj{5+֟ϟ w!_q:h]~xWo17N\2Oؿ~~;i =KRzKbLl!GF^8Go{l콙{Z8w/T̈QݤJ֗ϙ||-|tSGņ]dm/&,},?dϜ} `O?N?|mѤ0M)<] WYEg ?hJnQ3ޞٶۣ^Eք,iRRȒNms1{'J};FhCTd%>gR$`YN, JQ)Kr֐Hz 8!2M89MK<@`C&sΆx-P1C&%ckI蘅'/G V$BNJkɜj 0>8S%8ؐtL8b2sp >c E p: +T"h Hf9g#{]_ߵ/mխ[݋gڽ F?ZTtyظ>Ban=#blvr"tK"P9CEk$z(󱔏 Ҡ!bԍP6 re'/G ?2^ 2wT:RTFYciBLqDwCHBC7bQ:%5V'!`gઊ +pUUbpKF!);WU`wX\;îWUJGz1p7@Tn ~bC>�:/oe/=!'G0Y$ g9,NsNX\' JXFV gK\cIIO>)$+uMs@O*1';,c[z/(4q ĜjES EsT;F$ʠV h .j\Xύ$!\>$2A3⩣FksN?"<&;a:\|^ո2g۳qsvkG1B^[e&HFtJF\Vڢ2HYq 2uNN*TiF+ JhZ&jrߢ&&e^%5q{ъZЧdkIe#/*6`4+[a -ERT"~!XYm.1ل|*gG>ebT-[@Aypm".xg%R#Q#T@қih03G%Je AJS  څhS}3#@'DLap]=KNtF)ևn3]4-mznx 5t&kgs駲U|-G h5,)O4Ӂֶp $!h'R)F'MRPXФ{pr{  E%t  8ꂓ!KddM 8MA yB+*%*&h`$WUgI dNΤ Q@X֗#g``ߜm mH_fmb{7ۻ)»U;v<\G9G2s+&xPhG[>d,R^<2Ss<=7QSɸ~WAxQVSrCܟ6-<uBg!Co|39FGhocS {/~ị{ 1 {=JWiO;.۱%cJPv~~Ξ/4嘻jo7LSW.A]kRn}24wM\{#ŗ7w$i;A{ i׿RQ^3ۿp1Njl=&[Q7s͗#]sݠh92Nwo~s3 gI1QED0W^sAdƄ$φ7ӌ/;Ffr2;[oy7 >\51ws9[,\/ǖr/TB;9AWB\R1KK<$T᝼ J5Ȝ |7oY|g+9NhvSlQ% /벝;Gw̏ou3w$az@YTەpw~HeGW>ob8ތ3 58*"qs>q}r8&w*[ 16n%_-5۟ 柸_~Ζ/m.ᬉ[eeGZt5+QlBqrƓ?]E-#8NOZY IjT7b#t4Zj&ZJݽcq6{o[з٭6<2IY9T)a82Am{t3lc޶,2uE--HCF7O)Q:2j{{R@B_ bMG~1gkVit]&YD QTG^9 J&^ h5L2Ha$X(KZˣH{q)Gy%oa ɽ 9B\,3jH]Ouqr2Xy+`$Sa[}|J0<8^@YwbnӇĊzUxJ1?mcUArV"ӇW~*@FI@4B"WnzLF9='ߠ|+akt9idI"ڵZDkgf- 5g?/1+չdJQ^A I=1ͧP?"9Ze03kо5,I956aD:υfB)jch" pNTV;E!2a@)MW={AzN9NQYDFx2/S]mF(->&%q+&BB]uQpb GbAQYO9TN /Ldm2,D:#v2iQy0_5 ߏ}cc-W-Kdׅo3w;qHqw⥄}/W +^xfQ*4Ay\Xdp*T^kfE4 )lp{~nїx gqp>u8\O^vP-:i2[N~C%dWqX^H\kJ=kHȸItOZ#U9-Nv֋sit餽6q˙Z)qef.\%;f`kFoW?7oƓqǖk}A NinYU0m˃”V#Y.sk/LmP}Ti^RqοmZN?˕_M8J+W]]Ip=z Zh˗[[e4;N=U\.R8,]cTwUCA0JAPgGA1 Hc@;&d =a};g4>ݎT&W;ow9k(vـo{:;4WVDT"*ZD) yako& ;\}+MRR+'Ih.&gʲg8.,>/|#3>[SJ4]Geˠ#!GCG?J)"xbJL09u>΅}OaF J=>XD"FLS Rpc)Oe\q~{7Ik9cٱAf_?}vV>'PJ/^{Ny2{~2 Ll+Րy^tnlM©dIKG' ǻQ23O#3?|4j=. 44*6ΐ/kmH@_6f~?GnM8p S¡l ߯zfLjZ4lKawuwկ]%R/%f)gE}9:,ok>(" ?P_N2 _m ~m$Ϧ۩j$RfU'.!wtO ;:N1v?e`/w^AU?ez<#5.lw\SeakKy("s\:MXy (vGZoy(ݟ<["<8jc* H Cq:ʕ\(&*ZqETQdTww34C e}΃n!A՞++qNbl:C^>G b-НW^uuH)8 pPfi%M؀ꃠs8( '|Ň42> }.MveJ)(c6''G9)&COE5z ohBRk>lnF9!5z=e7דEl 1 s4_8_TsK|n#p4zR&wBzteO/rU7du7fYށ03f1ipp3bg?Y78VJ^\ʱ׳^m2:>]j(f 6qgj__erMxz_ޞ~7g޼}w:;o@f`\ڰ^E'& A~mu M5Z6G]Ox~YC^=I4tfnrKC7C}ROMi_x5!؂N+@6e/2W7QR;Ti0.} !|H GV>O#ܗ:6l_Se1j ydK$RDP#Lօ SG"@!2̃-–G)aY.x&ȭ`xNq!b쌗)$a *=VtYk)CIA>AF0ɿxMZ9ÓFGSJmΌ6>)`S0&K$mF}o8.icwulps/9,㢲,'FjF'V:wrJam)cgN7}蛈+'lNXGGl"FJ=VZNL)*xvRg4%ݤ.2(:o;00{-#c-ar45rF+lާ|.rĕM۴bdOUMY9fNLP1bj+zޠٝ7$\؍Ee~솦k-ܼդlsefx8ܔȵ#1$Ch}өՏdNov9v^ya;,7sx3y^77^l›a<iot>|Mx~4Ͷ_Wnh؛pY}_Lek ^%U >` ǣ'Qe]L):Ne,"AY+1ʸ0xX $- EJ!H@lQQ8 $PNbXXHKEfA  a|D>ܫjOWhqjw G[b=C5^O6 F)+Q(D Gk#H-#Q=&>yFnPG @OH堹Q[޷W12EcA Ɯ 62FfdlVi [ _:'8N*ÆGջ&'z!7_L9bs [Cm:}]J5xO^tVjLt ɬPMcMe0 3 j-P` dK"#? 9ZFٌv4 .澠vkܱ-jQ[uY88AN i)qQ8B!iKp_ јg6T O=A  X TUYXT;fIwճlzj$[= 6afUʑkzg6[2CT@m+f3}{賔xB.gZ4g {&_-SeQnPVNT1=#IOz;,'ad8K#9ѭ1<^s#]JbUgˣ#X'N>\%6: j%BO|x~Nn1َ%7~07ś~kTwװ5|{;mA&ڈɠ q/rZa-ڊ,ASz'T̄tR\FwGe(w M3hjDKФ\.& #h*дYTJZ%B'0rèUzC@@ P(J/܌^"៚+d Ly!ѹ6&իfs310=IipM4xۏ[rt"xH"L)BӉZ;VJ;~0[*')q˞f̃ufP&Tg1K:2~F$őJ1% F~܉{{o0B#SaTUηYueԦQXzLuAYeӭP"g+([ʸtƶ8TL)r8g\x4ZMRwg^G8 4*zu"ʌĖ$(߽x]%SţO|>39(Z?@J ).g\)s B@:~/x41AwG ͥŲѧ0+{x-d0}}b"qD.( J4'p\$y)%G@5f2%u8<ܟ\dcS\4]}gƘY1ŘaTdu4-XC4aW 9,h#X<7-[]&$Ϋ q9.grG\P5[(W`%!f"@"ǘ+d/jvvfo5c޵#"eqm; !'3`2}8$-ɲmN;/Yfꁪf߉fsI+'13zEƸs rĐ#iE4)1MÞ=w l)m@E D.$J!2)0+ra{T{3Ws=q<)r'i[3ϧ\3΁~ԹЯ>^5֭x«1`GΠZAOm;fXͲQ^j:_Gm]X ,\vvwy@O;_@ixѣ%P1yeLB*w"$O JfRTcpS+^lCnmlUaՙYj 7p9)z88)_>aXH/$%h럦\}ٛR7柣˨Mֶ^dm[h :UZ0%*|?J'*TupqWaTZ ޠoHC'zm<{>B., BCKF5SN*sJUZl"ʲM$\YeS.`Юl>cv,s,=r,Ȏ ݾ3q5bCج#73\0&(hF$MDRW<0Lǐ,: &jk1i_r@ǘpAzmHbZ9}e#Gx#2DQNŊd7z^(=:xe QZZbY3=ZTeGY9tdI)t=ȈG3$Djʃ̦8:+Gqpg*Z6`$$Ѣg*;oGDRk(, &:3ϲ ZV5,I[41cYYl$#:#gG>f n?/Z gKOWV8F#sAX8EA{Ȓqӳ@,F>?, !'n!:n'쳕<9SI`kcNI<$W= ji'zҒy>Nl8ѺB 1XV;=g"1 &%4,r&3GmLJyK1=V؇GY,غ[gu EKæ#]Ĵ$Y2;u*-d3E$ё&EHIۄ$Ƀ*!빭;v1A=lK1jh#(fUGłH弔0LkXb;d%rorWPs"X?= +A:&Uݤ7 dSswik ] $J" {8j^ƮMn\*$ f +$Qy %޺ IQ{".$QH CRZP D\LRRNUhYƘ gJIE|}F]XdCy ADe "Zdġc9qj<( jz= v1y\Re-jR@3ޑ~x\k<ĎtBY"$! 5)%:IXzPr Kk+rJ 9e8%q$9s$Dv%/!؀ʓNF-mLD6Dp)& BBBjSvd2~QygFU6,T=Xg/ Iy{~dE ,J{HD+R]{)kV'f%ٕeVPҪ NJJ8 ~tg}J XDL$ 7heɠ4PD)82)I%HҎ%Fpe\+}fl() y~n=qRRh}^Jlㄶn9|s}f}=v)?SuSR  IiB/ L:6k,Z:N{y%,gSt[*6b27ԡz5{MYo{A؃7 o AxvI8V,W(eHWAK%B# N<#ʒ.3 PξƂޱrNQJA}dY1҃pjؐ^XB Jڵɢo"D˭ٝ* &Y~2AL-֐+uѪM܆xn|<.+GB>|i }S'x98&R__̣KD2gӃS+rqv-3-6?.HN|"+UFpڜMbZ?ko͝w7O oWbJh䒘3u8ʧÏg+ճ]1 F:J--=kXHF$چ8:byOLp4'z))>^.jpirNQlu%ri1lGƓX`e8Gݩ4:մ&t65qx_h;:~,߿9O;~c=~ Ia,m$(~o~o]WCx; Y..2-)[rZƥȭH/"^(+ٔ!ҎjK&XlivuN=A ZWqV}MwSvY%NWΧ\#>/#څͅ}j16/>SƍeX=z"KΑ)>l~TI 08`]w49VT ȬrS~}sw%P0}/:GN{cJEgPq֒,(:,&!G&DNp92,8$/&R"Rr& |D'K⎭qٛԜkz6IFRy/7{bvnj֒:$2ᠥr:ó]TY\nz.wz|>l-E7հv%r#iq<ڼX6w]z="\0^aIj]Qzuմ==tK6۫n|uԲ|{;/;rz['.k~<ʻ_<4|KY0aޘri.h%U~"|>߻2ܸ>џ6h27U Hw7rdEC+twJww{qeޣΘ9sQ@h)QlQ*+r|SݝKDZ %7ReY }8!^rV tdDwfcݽ3rv'|2zO{|%IӮ'=ͮ{Ba|L}jڬDiy/NUhNx 92d(Md%kq7WHB`e+ v7:P9FG09Bی\E-<1UE}m{ϖzq^hTv,Hژsqy ,0'\ǘˠ|U&Ui/t坟(tgz{lW c=O~ wz1ۜzO?܏ZEt+G/jqwp^ߦ.&i4h& ?{Wƍ +%cƛRW{Inv6YW h4%;NCRlEd I69{Ocao~teb8yzy~%j>pR󔗌߾pP0ퟔ,g͏l>xu^v!!Ȍ j3=%ǂ_`N t҂ؔ* )VOXQA+cW6@\a >ͯw8}R>LO$ )ԊBܾ(dv1ZP*s.DPYHFѶ>; FZ [@al)~J9HHTZ=CsԺ>{+qvj>>r?Bh3=c:+i:vcʻ2D& Q+E ;!^:0PRDmo%KEdɵG(SɕvPZ#c3qv#c; iƈ o5_\Vf=#1|_Ѕd~[{F5Yv𞽧%d6"r]V!XI^E$%%L>YvpP.kٍt\̡v3+jƨF6$[t=ϩ9:$\6LӬJ efO6Cff^4lkRddj\KSщ=L CX{~lP~}cD#"6iUہeĤ qeś}R h7v$16ED@fg,K1T(Kř,1lĞD$ufѱu橎U ߐ:iqQވ#.nݏ$cL: lfCOh2,(5 +yt yJ;X ax }ڏ=ڇ+z{rvQ`;M'SDΰ!я!wƻн}g cEL.1[WU*BXoqNlN}lڡgZkt[ Mse:«Hu cҁV()N(%8S hĺ?/]v($q|ND56,ON|mKyWs-N2 _`ÍzQS1RG쮝|o{AsK)D0K4XͫM7pL6o݄he?= ڃ1Q $Tf2L)-Y0䳃aQ% C'Pleӓ!xfybcuubg8J|~BX?{FJ~D -Y >C`oa1i#EGoK/ 7"ƒ B0#e]3BD)k29J)xA6,eW"ڿlJ$]A3>>\3qvÌ|D#~uJΙ=ڭ}j=vlYp?*gāG RʓRYi H&-0 O1Æ~mC}ٍ+U#x=9r]=<;[x?i%vᩋb쫵X7^VM~>Ӹ4r=NnqN% ~Y5>B< MUC _5\NSN2_a~&3~ߧ\dCד.~Bd,^)zećɻ0_ סF~.SIۀe} Gws}>(rQGi  +Yf(!6;Չ ՖG Z%A 4&R9==l-/ ;y0pUWRfpU;s+ㅻǙŕurRhړ)OWO:%N>Hɿ o߾zrL$)O;@LeAڷ|J4x@0kU\0UZC*%94 `XUWUW~gW\`ઊ]Ui:\iW+hzxx?>w}5ēNBo~RhkߝM,w} RW!<'E^g'0K׫m 3b~{9K/x3E=Y޷OҎ= \.N{ƽ|ugӚV^-@C?)oIzr˛X~0k!{>TKzM6y>Y;'^:K5heO*9١2E]TYuQG,@;[cױ@J6ho0ʺEZKIT "(bPm/AΑEh֪(N3)p&`0^KQqd3q)򐪭^9[$/ח)̷0≉oN7I`{ƮH̘-2|"XӜѧj[l?ܗfEu|NТ蜄ЁBhOl&I7}F휩nE7\15UOfi529{![J9۠ed1ٓl6Hh8k!S2A4%>f^J+3z#S8i(QV )vx}9),ځUxMYN/׮ڎn?n~ݗ^`_.WگrvLZgeΒ"sR*(!򲆊WJXh,w=GYhg/59X(ڔ lV ⃣mTIKij| ڶ eQ]qB"YCl8{YoW=3[լۢ}L RWZǯ@&JwDbI*U3*dh  4OE@N?)2sKf*8O3n\QuPT^M2&(HQ-~']K9`N3ޕ EaH).ƢH{[s]@hi&1Rt⎥Lw[2]^^x7glH(1[,6Qffg ːq)z<hmvDrϕZ;'z>k"C3}NXکۀ":D'i!8뀍,y\N:Zi]C77A/ua0}l9?+^>.{q HM)ؤ&u|xFvR#vVu A 18X!)z!)jp!)r@!)r IBldrBDdXYe$*L!l%en$E %*<^#b&d^Og")B.BRhyPErb+kn#9/^?4YG֥?h5c#<~k,& Pc"oV$ HR+f( ݨ33ˬh" ?OkW:G 0ܗ d|V@5ʃ%fy{;&cNoF왳i Ĭ.KgcY042dלkۂ^ wNK8J!/oo -SoOh0d0mZVw][?ߛFQǕkg8MgQlƮɼ7rS`*eWpmEPlQ`ctQCT) խM9jV If7izsuyI^#:z((qxU߮>xf/K2 NDbIL:yX6ZƃSW2(5jD)zM 2)OdqKʀB+/|flW)J=pE{jm"8)Fcʼ4D>lDtLI+VL\[0Gj eV&Ғ^k@e, =d$pZxDnSKѦw_B[ud,bxJgfqN-yp ĺ_GT!X^8o$ͯmO"ݕ_Hi "3wq$w!{T? q~\[R?U!h^:5Y!'-XNV@xR4"_޹Gzā0*:lJ[0e[%e0Vdi-[I>1T?vlH:t*9SiKHI.!F#4sFBU%' )CP!9ؾȑEh(#yr5% r4zR0wྖY&x#{gֽ8bI:q|-`2`DEV%eH=iKg9i+J:ў$ǖ/H\Fn%!P %MEmL \ H 1agtHfZM-{=&wp98ϑx ->R{JfE/:j =U,[NS{²W%Rtj~QŽSmA:ᤝX5p3Lsdvp'3NLK pK@1P{K&NəR ^fl =U1 МDf႞{fMi݆] 9oWkOVck:σk vצW[^̍0PN$ul v-3vj|C98yQenni`z1?[?O~j/>x  ƚ,(_k+zm7#xͬ5![-%uddgO骫ݍcl/Ĵ GV> U,xtyuëlt*^sѪi1GJNJI,MLq0΃"6N e`4&~76$懟~,?y_wݛy^ӉZ5%<h/^2Զc:Vhgnث#9ȍGìВL*SEUn X"C",בذyM=k]k9U@bL5}2%V'+AeT9Zhu}η=U۰wy%Vϐ gͧD"a\f24nO)~>0}7{VzQCs >}/ro=z&R^mir:j<_E5Y'xrvK׳/?[>Xk d/)k߿oDܯs^s+}Oks?. sO}_(+Ėvv Kͻw9o[ގ͜T{֞}J?kǎ^Oi/Ԏ>CuS%=>ռӋekЁ[]&b@f˽q9(-*{R?l"'1R%E5Π%e@MRF擄u3DU`CL ) \d3˜ρ +_WUΞװs|Lc͡E}fϭ;F#h؄ȷU?owU(Wxܵ[hp3j`wSm޿m3ldg{rЦx[s틥`{ӥ ;=a1#Hw`zۛMw4zgGt0l_ͶAz7˻_|أ7d,Ld⤻8TGIycݟ`KwwGs`NCޣΘsQ̣QIy&TV6K%AU$>pje|6P2^OrdwWWYV67;{快l=]I^Ѩ !*Hː!ȜKn'2IV[O;:r;s%xNEO%!(lFύBE-=1UWjobV^,6~bJvfL#4WhekzW6ӳTP(!5l 8 cA^>/Sd>Ȧ)R&0܊̜UJ"DLJ]Dm,̳$gN*U\2Cd6]0x_}/[>$EBΉTN"GiC%t%406å>fAZ8FF([IckLAe+wo <"b#A` vZͨx6i˽,]cGd*%)@P=!ؔDL.R%.@GY4Į6;GiC<]:ڶԶ6%B)b1` " eH}q kъXLXKq !ҢDC&!Zd9ǒ JaZ׾6UE0DFJDWY"^"qkLRC$9mZ)'} D 3N#}0Y! ҍQ%ke2 r3![jܔz4iFz!kT69VeL1:QɁrQrq^/{eT`=(1P FqKʣJIb&zrh~jQǡWfSV1pE\Stۜzg~tBo/~wP7&"uI#u^1!:ke& gБpCnxxHGr |%g՘2Lz늧8c S܁ruZBJtJsΐ5rgY`aa2W-0ޕBi"YS:fJyG-/NSg{FW!N߇`ɾ v&L7fOG/ùIJlӺ([9$$]ׯUߢm ,z'B>a>?~Z0$ܝ~kƤ‰"g \n%DFF*-iK%qhpai}Qc4oAQ"u' XF[  /YPo vC1q>oa ĺp| &L9Ĭj<)P™g`y"jR:pT IjW<;zjo ?;̉;؟^=-OfE#pymsuWOϊWYyNuucvfG-3vW/ Ĵ %=/qg t4'u8]яYOC/2K7G1br0ʕ8g "$xO:XrlqKm0zw 1AkKd&b$# U! ;AHha) ܖG )=c3umM_5|4ws1nQpSK~:/-nƎai:` WJs@GnseRQ GZX$˕5 w`.HBBb,Ep h7 ^[ rE*.L<9SDc 6IY,-`ðqM6hՆsZ-_FbFiW+lXݴ ?yޥh&R JL'0 s#p0W3#8m O3g0|5=yh&p#sIȥ!@Td%;m8aZF,Oܟ.oMu[PS #PAǩj5UpL+"iƴbaUe]5:.h8$0@Qӑ]׮&pGs}"KMՠUsnU ZDjE0fEkJ4AJRa7r,QR!S܌oO1>-SSo;6x&~d8} ;gfc1;&t?>d >atz/W_ q )M<a ?NҷHMRntȹ̥ls~:u01 A/h~▱$Nk\^uO`>d/(}C8+rYKJ~#_Ώ?d ~\MФ$gRN jvq [!kNIbpWfs6uׅ3 0"{(2'5}# Ne>w_zŅ,0%O徣/8:'O? #l3BhRrwvd^u'Ex28ɕ*$iy]0K(L[n^62Ơ|˕_v/%5`F㬺SټDyWO2%Z'ڼm07P1r U ۋ/_?jϧ`ƷP]Wb UWak.ߓƮjIJLP5jx4)ouXƥ0a { 7 Kl,;.,&V;՜mWsyح<}zqpQ#2\4";w´38(1Kø h8{?aem[z@:#0<ݮ꼊wBʠh{t)XZ _\&SU#"F$ӼX;2}oK,.5պ~o- {8D$8h&1XE}WjFbtfԊ;'6R2Wzs[s-1II<\{csUR\=BsE(˜#sޘD\]n1+Vr;P85[~yWr^O;"|xHXx /{lTpS/^whzKye1r\+yw%cѰ7xWK]4@I]et18xr}O1OpߪC3P Wڸ+/ώLrJ͙"<1;8g"-g:3Y{ɱ6$Qs-9˝R 3|'&ը;0y w22)np09sH[ L:F j{C)a𖬹H'rY'O9\1c#tsf}۲kwX&[mnvJdi^p_쩗J7!S[ cMH[A B`Gi[!u#[i( ΍E9lnϘ|o`KTMبZh#x0,bGp4BVyQުDC])v?Zr{1P[4Z[κ߶En*7,A|^( t|~C* \Cb}nБVbMh̢Q:)I-J%ZRB]5Ĵ_8J:ZP_ {v;=-\߇vi}lh&&VQRƌƁZ1b"iZ+Ih6]:Ju^Z]*r7ЖSMC(|fjCQ8˖φI.l{xK޼>k'#ūYS( Gm3AZ9!"-Y@U0ioxƠh$0rIK FT8G"W q+!la=k5Ыo@ A$p6tiuGcT  0("d1R<("GlUI8(_5˿J@NDa&*FE%p\H5Ȃpj';oVcPMzmIs2}ߝ=>kLDMQe0Ή'a ,ȪZP 64a-mP=`,w׃ю$d[ kqrڅ(K{Wƭ; CN8hZm`CV#K%;9ÕHbiirvw3p4mfTCRVۥA)-((뀄!䉂MP~U~BRhJF'BDY{ѻޣ>tH,*,x$XfgJ@CD͐32eTdE/!3y QAr$L܁YV9TSסЧ?^>LvYJAk ~41$3J_GYs!ȯflDؠ2|J@.B["X&Koc!9:em8;KAK鏥`2fz<qc";"/lHʓOR22)ۀmKD6ikJ_\&bl>kiL+C\妜|b!Ef5x>P|]єdƘSd[&YEt `֣IC N2Z `c=$־gVY۰4,>rvQQ7% u+ \iU)}Eݔz=P/aRIzH2z.u)MlLlp#!yWw󖞽kޥμ6Hq*<'KU%dN<2]U}NlFdnch1q}%X9[ㄼңy3,Xɜe'1YęCШTdْY$AԐ7{ %gb%nov{}OfA?_/씽>G|I3ubF'MP@^BF2_XLn_"SV¤сyyI3$N'hɁlۜ<ùB-<UFV$C5q[<l]ۀמ}}{\o[)Zs_1xOY:)d89KOWe905VT8b#ƛ ' ҃CHy!޻B GD/ʤb[ndVd nBJLQK6,m'u֚|JZen \J1H!")bt&|&(Xp#P7AD9@db d$ %l *cڅk .tVN 6ʴ?_g%5c-Jzu{I'm$Ot>@6Uϥ MUT!r Flbx c,e qIϳ+uCrnymM1Jbmm8AOBV]}Gs6I@I j#c5qv8sYTj1`ᗁ{[fܑoy]N~bq]A'xxር](sqH'So?06b=n~|G"j=Ne2bFʗ+UA5&M6Vdab Tw=st>)!f$"$JTS9't\Alȵb^M۷3LL-r7F Rł2[':kJɓI U,C^!gSqY,iZ&_o;( 04UDHa-7I& [jn94j縦WTD$?$V[b y`A"Iq%Lr6sV6pQhl-Y$j0@JÓs$gBĹS֣X@F(D Hhb/,rv߂fLxEWMo?,m尲w/`K d{eO߿i7.V8"= ĶZxɷ[ʚϼ䙔{`( 2);2Z2=?07_쥉kV[͒U]S9B 2_+bƏx.YN颈^>g>8޷SѩoAmѾ֟V(tp89ݿUáLbwjECskiˌsv8yD^ѨroDގnӪhټ,v ǫ/^_'fCq'x'^[Mgmפ~w_IƺF2s$k=c\W7d1Mg->,OAϯnQ9<&FR`ryYwOk)+Rΰt}>OY؇=͖ M۩4:դ?/;OG8>=&u7W_}=|ׇo{E3 ,]"(E w054ZXj2y1I/րx;ëiz<_VzX/;ej$x\Wid'W]RT(ЕIl!eb~#xDc>7e%7G[vHn OYL.S4CUR3q\"HX ֐״>.8@6Ǭ*Ƅq)C`U>RoK[Re4Tәi_CE3eV׾#&s΀ XcF根6f-6G"dl~r~N>ng ɵB30SD&Y'Ji$.YȂf(<hJCUvRz%1 U9bnq9ּO3ɝ/{><6Dme0!*gr4&FK v@GA+5a'I%lJ`1"Z&wkN vya`dF&e! r0IpE).2i('$%xWM0g> @߽[\^Ŝ<];nʎj 75踟E3ʡz4\{T!J2e(}5z.^+1O-=3 KCHMR,f KRV&*Do8x Js)i6 O)tѭ#^'F,B'{G^0ˑHJ)ZM pL9aLB.+"RѯY}/8[cI-:lvWQ82MO"E=q.:|w!+ ^ xntauø,ΐ& ^;3@Zά:P`2?{`m{ޤo]Oğ lٓ{)?ߎ~S2+Lig?-#wJ翝3&p:N_Jq_{#l諯FE<$54H EÊDyOF_{줄t\?%[?LBz9Fhj4 }\wʧ}Ũu0{GwOsQ6)6J-r0WTy>,mWjUAʌR(G޵q$evqa^d$ F?%^(RKvUC"E GS险竪doNƳ4qʾLuJQiZg9}]b8.(OAt?VdjݓIEd;(_o`\Q\E\Ƈhom;)T1RfPziz1 #'`! <⻏E{hQ=I3x[%`߾~wo~ޞb뱓weOBXWsv]izqxrṎ_MpQ#2\Ðxfa;asǥa\u4tLq(zz,L2s )ΫHx' )Mjᢣ T Y7:F&D^FQ%66vL#>qcco8J)|zu=~iL>\w1Ҡf%Y=(RdKJ ,} > oCr}gg GfD! U( R#l5FRK)ǚ>+M471n@x~~l6u_͆, n{PGI?7a˺sЙ_]d;BiO"%C2T 4r"ǘ+dz'(iB" 3&b1 uFJmc肍NGD:SwuOGCǀ7‭iUYbK()hƽuH ÌМ`6q6O1!@|<-Ͼ% G>u챨$"ǮTWQz_uѨ$׽UVcWWWWߠ҉FJ\ɥѠ$+|2ꊡ=C/=R`EW[J+􃨫-TCmvyi#RWI`&J XUVcWWI#[+±T+XǣѨ$-ǮGWߤkgp}ȾphjrYƧq"mw8]3+p:@~}6I 9K  0A_}SxW1KɏyA;bXgޯqזTt>g'{`~9 YW!g0W|_X^eTp”6V,,¥?4:S:U̠ۓ 6ϋË PD.(W<(PJg.ERfB8cw7HYv^R7G\"G)ɔ\1lJel(+^{{0̂z! fc9 8P2ڇl  ].icb"פɦ7Fz=aX_3Gϯ_ï{+~6!Tvy##j `s=<TG61͛٬4<+7aP_P}ҝ=WcwuDL` ?X"5#Q ;AHa&a) ܗ ?Hq/ 2zqcX`؀0έ!#Je^:JBWF8FHcƽb( 8z+|KB4 CVa&4C8}\)w}9cQ<"I¤ҎRD rsđYL,Hƌ'L/|z:UOyh f>jy\z(S,+ i ?s3iFVRܸQ/zu'{ġ8C"@ 鄀,Ոk BVD|?5;I]_znsR%&phܒO\g+P߫.=JJxJJTt__b~`.=F// BȀ1:8zR<4dm]lkI'9Bԍ{(,Lp)^:H \2x"QʕNQpP 8Dr8o4_:zrh&{ c׻wo atզ6}*9jU]u鯊PreKGWr~za$00gQhen*@RqXt.N;\^ q\GC..e2h̨4-ӄJkZUqkAf[xL{gӪ葪=XIIƤ"yu WoSqBCe(?OnLFCyAR#_U}ߛOUӓ4㋷cr}8q1ly@*y^5s"F7̬3Ig^VQtW5 W#:gx"% a^tM5W!w &osdXf0Qk.ԺSl=v.,8I#)e4=Z8<}ԍWjwz&8(Wϝ >=l'L;s4@_ɲ>;.EOEUZ:7x<;FYd++}^.ߞگuv;XZF//@,_6 (bgąG9fZ(>z5f${ʻk(H#aN(EKf-D>p2PCq+m`].W2,lJK#Q`"^m*FA Ne#Q cx&OVZhbt7 yKkop~I??6/ff`F>}8ǔ-?I\K$-;lIJ`e[HS[.rdxL(WdH&H॥[9\!ӻg}>>S%R{.Z0c"&1chPg6Fo1.tDDZ.>way_Fh@ō8`1waX}dqoR0#4'(xAcf58SWzVzy~:n yq>Ŧ]c,t=ו8%2ڂ޽~ Z0}zCyy˿m"K*Y0_)1Q4=w 1{{ǐ@u%L.<3y a92P14Bsč&dNN,=ޑd ־Rnup3~DW9#93n5)بZh#x0)FE0B&+K`{?yL= Qs7Ub~K(hm<#V̩;388T6 xi23ŃN( CMziZMkMޣT[:t$*8ʝ7\ Ø8P=tHbaǎF Ѷgjۢ8"7gapu([g ¸vfN>Tx/bhJ7v=ټ>|wD%ٿv#]&$G&z2t5- CuCntucõ?_1riH+Ր^Ct"$3\i Mf2fJɴΘ+:́bQaNJL(S$2jC4%QaT7c]Q֝z@՝dɿL20a%/76Df2H.0⍗31 K6% )}}pQ8.:8:"f+#TjbdDp%$m&6QbZ:٪jԨk4= :ui.:U;Y )ivOputn@Ɩ`P6%!w;n _!;Ccnccb* HDPP!hp· 'ϑ֘`"y"iJJpp^.V@ib(҄H8,1Ɍ?s7L:&9%/ #H4Im63/ ]3O]ݽ%Ov}6%we/o?Q2W'ڬ4Xx΃I\ !OWF( JMS*̢gnۮx<ϳ)iX2bA8DL` API(a+&-`#;Q\ `ggpUfm 55UuGcT޳WY$#!88^'^\ַX&!ЏjDjIʶ.|XjJ#yآfUUOETIA%PH9uNύ Er4 tO%IsނO?2Fɉ,Brn$qǕI".๕TGc^*&jI'z҂x]A7eMr$$HӐQD,5N!bzIGѩp⎕--%*[YւkL$y|?cXTބ3@cT&Ɗ42驭Erf}0n+W EF|ېq\Z3*P$U4HMv-JƈFFAf"/X/$0VEgܕ:daߢ|N`EhuT$0₀RQ95d΅֬CP[GQ0DJRD&a%Y4AdZ rlt2K KWb`sJ1UYuTbVmNkQ,)959lPjME!rqϹIbQ[%QJe MPJmg|zA:@K9 v@8x&`eBQOg'%ۨQN X-8R$X&s< y-.0x<ݼx&,aq olx+]d!~>\.cJGi/R=HWHu$Y מGa0).5cZ Mm)|IV\ԎJ'۾tخ"Khj90j$#6phrڹ D2=$MQ=DtXНf@ܜv3 VVjkw>ж*Vy1N?MeSQPo$ B>HMɗA$O99lіk([R3X**b**kuS!kId3 ՚1wHT)ҮLYπǣ4/=Y˖1֯)s^kYJal$'U*4,u kj}%vڋ d4)Y2e6P(A(3LS|@R6c_`B@M ]|(l89[/NquY K9)g|狫>!ғ@A3*=hⵤGC6狻)w凴0?w;`a'd+gO.&S8mȭ2w)A6nf'%'qu+bPR3h ,{m[|Mޝ'\> NG~;Bzs28y=0Eh S#dyC@^Y#\gߍga_ >hBACCa6 5xm{0ny,Gˈv<0v:޶ohpoma|<^j.ϨJ*l{=X>دxx 򸚉o&]9kϏYq˹{~8Ҏ,*3[ OT=cr)エw-ȮҫGr^SNJյxfdّ_TPAk%kZW#,gv5S}>;Jq{.:)veJ,Lʈ[&!Jgb <6a Ë [^ ,.zY+&'(NqwDbaG^Aojt V-U [$êe twR_"9[)Ct{m9s r˿ `_ӊ1~eM%\|--,TfQY;rFU&jU|ʷ,Y}DD4?2K;kc`DDƁC Dx%drјBƴag1J,ia G+J*AEVTQymQє /s\ J/%6 $4Ȉ(P ؾ^q&EȪ?x_]N !AE [7?B9agdX+Αdb2Ex'6U)XO] F@ts~P,npG`8S/o#!VWsc=ՙkHӢٶf5\`zzvf4DȜqw?l cj+Cs<'>䀸ѥZʜfyS㟍͛كW˂- i\ũ2vm\%H5 :6W8{U#irl/mUðXredyDa8j`^|'\98LMΊ2~$Wr8.[:7T0 :~޼{y}s#ѫ?:z%8 L*$7un W?}ж8|hnC3Jzی+JNyŸ7'Ÿ2;#! W?y9? &myxDs}o-/OMԈYz8 d~G&Uu@kPY.q/Ɔ>_@L|d|b&c*7:kB<kVHt7#Y-j8 U& DC)i(B"ONDŒ Hml8VXH+$k z9"E/3Abƨh*w:Et&0=?Je1ct˻s]ARvgz]!ZNWҚΐQWWB Q ҕUZ !B;tpmgضAc_]^M;=]qp8-]֞H]=hOWO.z#KMzJеi;{\jwvӿ-L J{sWO̼Ta}\JxW5R P2izvR.[Omz-ã|z*jvM11 r(z5Ors s -`@f5)y3ҭillgh~6Ϣi@K_z4(YoQ%MkK0 V;W>($o[7&sN5/iTXo>nXYNij+եbU #9iwWy u*>A(!Db>秏"0G nM> Ua^ _GZoƲiku3BzdNHGDE.D "3|dS*җr|螏=dg_^c}ɾHm!ۨ% C{Mx.RdL8·k#E5EpEhmh%Es0J!BFt80^ϧ+DIeOWgHWpFm  Bun=]!J{:CRFr%c :CWWȮUt(kЕ6Z.+;h Bs+c0]Z<!;CWVvBTtute-~t.5]+D[=(%ꋡ+v`ѳ b4/=\7T3"VcImUz{݀J[Jw{{yyi[v`AO,Վ+O*8Dn)ǡlۮzԢV B:DWX•+thj;]!J{:Cb, JQ36*JH:;nCeme^EϺuC hr i틏A61 V],D{pAjD [ԛ9ҕ$I!BunDFWWHW(U sB3ȈVDFtutQKSXwuF]!ZzcQ^]#]J$1 ]!\-BWֶ^]JEDOWgHWJY [=9ǡl9=]=E/B< ^}\vǡ'j?l1ȏ+SR]!`;CWWt(-'=]!]1ƤR+|e+th1m+Dٶ}Vz\lA@aܷ+Ʋb8x ]-(2}AZ-%(̦הb\cIQPM~PWxRbp}"Jd1Ta/|eȟx 1?Q=d,הX @ 1dR!o»ߛ: E"[ !+MfB;Wz>i&ZA?Aowq2vɹSai^Q#ΨL]KW2ܦ̃,0^^6; &&qUY_r|nu-W|zZvAL]P}*T=FY=R\ݝ_\:U8JN֕!>zy"IM4%176H ?VWf&V"S ()+^bĦi4mi[͉jT5 ?oO`wXtlfeDd"yмTw%$ARjA*dh`03O*Lk&Q)ML zj|L.- ?93zSdT zC8 QbZd-BL"FKfNLBLIǠy=OTԅ81L {.*NnJP:vVqx8lb5ٓ 0c2sWPzk d*9)}:I4q.?e5n\@qE.S&jo=bH$_e dϟa2N`g" z+魌[);ĝalB =i5ƕu:'^ 2]ce]eIUI |qU֏D e^TQɫ:?o4{Wn}ph~eE"\Lg3ƕ~u/tT@Zlk(dd!×_[vSZW<&Am{VmO̒n|1ysPRoŒ˦>j]Տ[),j殼t:/@C-.;`Uzxa4q"]ݐa%A˜Zȸ6Gv@6OkMGRfmuBCG2h LߔtE@FbaWOpGpVF +8dyiȒRi]z]zZK6JJ8)9Ǔ̎9GOy`3aKŃ5(6;f#5&錧He\k@=2]ᄌ>pJFRT<7,qs"۴ZN=M.gVKAm+_.|h k#简g#bzz**ҩHXpEgqC"!JeST$.@ǙkHN4bU[kO'"0JMASB/imnk-6ECfeƟVq*s ʉ,7,Z]ITC+VR &{՘ qo'vҶ{-4(kc$CD8bb̃: aD^r{h8m#GEȗ6_ \fv3`260b1ٖ<{~֋em+2mNg0&lVb񩝊6']d׈]w/|7dJڐBeoIeb&@i=o*8)kU \;V#rT>tW5D<\,=)7-B iYtJYqB0K',<% Xb+ 50>5C]Y:=Q-: w0y{\ $J2 {8-ՃM4 ; I!`WJ5[%%3d%bpBR􎇤ȝ I;"b#u &!DVX<1Ah%%na X98X)(΀YiE4C $r"%@)Er: 3qj܍S|JkC3guMNm}mK̃dY2ejSDad}E6BJkҊ P7"3vB["'!Ie)࢔MB%<'ϒmΆiϾ 쎤ZO+}|ܘD ]DȦGDi28RLf!0O`_tY.$]XdA[jֺ[":a tEL gs L!m^)@˭媝b풌1I,:kd̻ FOj6+ͳF,c2FpH4{! =}gbruԻhh2ԧvY~4 G4CY.{9J)Plrv+8}d^dQ[pN^{ޗܞ{x=_s:% L]`󹄣t6{jYm)^dOUK /ts`*VeD[Rz sI&PI^GnK Y[\M"M۠+'^ᯥf3*Wǒ1!aC7\}suϿGǣQǖkM ͳo6 meޖ,[)%O2+ &80u2:gd)!A*:CeHux4jJn3QdSՍkK :P>PjmqG]+q'`+bI2cGiӖiБgBCGߞ J  'D7YhAh(ewN绷^.{@AjHA[c3 >$9sLNyƍ;+sw9^~{7^ؿ=2Dqb$vTtS Mv}"ەiЀi]YK2Qp|P&5Yڄ1'#XUZqr8ٓ~u2x@8(.` T AI)+}v(NAz"zDF+m=B>rl9-:DRyŌL'ISID גbp)MAHij%HUk\x2HXZ-3(tF8+ BE2w[;ѹo9?癅{runeV TPʑUo%}= #р^Z'ph?d_FH\ƂIBKsx`9f3> $YCdQNTWgV3TsЬlr(.<qEju-ٔB뎿*XeN ?V~׾xRr_=5t%\Lo',8f(}c,8h 'ĵ\$ͽk\3e}#C8d>%{gl5ROtF tO|AѬ#@hDmvhnq"F J97VcqHFs?ƋYhH[lZ>8؟] n/)]6`R^M=~j։#Z>prx ʻ0N҃yZ.w޵7ޜώޮ:Cq9-N×( ?V˱]s F6-Mb%+;[RNxdyOBp4n`ŲG'ˁ^9<n{9:[V/j绖'VŽK) k>OT}>8{|09ו*{K Gt7??߽O43-݅s|_ MZMS{˦S4mxͷ^ &WSTҪ֘jmHϿ92PQ6Z.WMpu=J|Mb~ZeU˵^y@sNSBP6 +=ҍڍL MT{w֮MpnHsv{եc7ݣnIMAK*ݾ)K^= hzG2wGJ6TYͫmxݕ/ j^+y|ǷԷ{{X;o2 mo-rQq dvax8%e55,iLÂc1EMz6 [ e\hX\zWlG2oƵ 9` LlI2qݭkoHs&~?l{ 1Fq&`j"xԦ7(`)zG ,A:%ń;D [: 0is:Y`)Jμ>~3fp%j_;|{>>G>Q$&Id.B.Bҗ} t1Z. BdY@ĺix I@&ݠ巄AT sS9FVwT8i,]8Zmly t8;)2=Y%ݨ12YȔ@D-ge2BR}0)h$@x6lN)JU\ʒ R=)(N.g,&Í k2)ښZw֌J5]XmfЅWgwW0iz1Ů̓?/Єt5dYR 2c%d=%/WB# PAkcU4Np2Wؒ ^EP1RڔA&>Y2%s>SGG unǣfb֮6;ڮvm3>K$GpF) SiV[2_\MAwFud]rǑ Ґuɺ1B/pwcF8u.7}rfp!ƍ sZH$0U'OvשBj9M萡 5pMzMtYU.m3τgs;nΘg2b00G|X'jVCHɔ4%[HR9Uʈ)j0wG2|ж2y4%Ѽ:`sTMsop7/S]瞘}9[<w'(8uwTETJsEU݃qHFWӕҮG*hvΦVRNNx1xqoC=3>poGËx 2ݴNV?~Տρ{Տs󺟝Kߥ=63Tk1iMEuAy4k.q=zQWZHjR?N'1`.~\NՅwjvOe .>zb)7ݼqR,Œ:2ě_yF2h6*h%(y\VJ99z&Yx8y=b;t2$h:-31ڸ]ppeF]mJ0P&e'˰* bn93\oBW6}+F`Ͳq1tpC\ ]1d2)#t*Z2,ؚ4 .d BWOxDWzطWf+a1'dqa (~3tEwzzl4&|u` _}euu?]eس[t"NV1[' P'?ׯ8vN^ߞ<<\99Mn>]lWGqr}l~A;?~qO>7?zG;ć>hgf"ƫIï#{>;V?V1^zBxv;S@o_rIJ-<LrK?&662Sn>.olǜ'Ş3z+G nF@[Vwߐ{Wߵ?W?lJ=~ʄ+q}jse.ўݼfyfw2|2Rz z1RK~Tcd]1JE=EI$CWb*Q{Y2J턮]}*U]޹8J5;ʷNv,l[;')CUҙ4^w֫W_Ry{[G^_;ɇSWm=9;ėvkHOzpa9Z,?Lq~SVR5Y쭢io.j7KO/{U~PB5{4F\y`><!<,jckعZC۬L,ҿ]r${v~ k-G۲K"oif UgKUXwv?|؝dn}9_2y^lpQ뿞åȺ*-QঠJΚMHYh us):oL׃?6>LB:vZ*Acu֪ܔӍ |ڐi|XTO v"VBC(B%̹I:,0e*:k΄1h71h8 Zd龁ORARMVa8 pE=Pr]SJ-pϭ!nмC2c1^#1> ڶ1ZEA1ZrV>g_ᖈfJd?[g^o[@@TR4pFhSnClt)SQ\-ԎFV91F<ٻ0F>8╤Ӝ`AV5:(*-w@iLb oDXqvQL&j^𦣝W۞dWN#w|эP!ՠw^KCp2PhSz"(0,JLhW4 ~D>u>:LAZ ]{ #AˈzuC.M_b>MzC*jKB*q`(#(v/)ڽ,=kpⶦ3:Z.fX;6@D]C:x XbU30ce@0 9FJP0\qih}0;yM/=כǴGs=w&Y>Ђ ]H77kF֞5`ۿΪ2YUa6f4Yd0-7n(/3Flz6\}YeF٤#ADy ֵؓJC5G?T3ڍ Mt&%\!mE;B '/J;h6TxHuXoުͰc6Ձ'wXQW Ժ) }r7pO6r%֛1TDyô*a‘r'p[ćf&T:9Q܌J#x$z U *cKULmGmAiomEEJ@:XV=Եj@רyWb{Й dXU Tk>?mAOu~{Pq;&-ܙZi.!@6(#< VAP4khAdZ 7bd4zldNҳs|~=bt~֘-jUeviL  XTTGu 5L/BGR!6OSZ,;wQ-* g'~b!hE%蕲7$ ^ǣq|]ٻ;b&N*uC9 h&gmyWDclQk%&^ln;?ofANa9u7g̽Ι=\LvIN ̌q%ӱ'>r''g'tOZ $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@zN t*vQg1`#6pR@}w1 g=I'N.qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 Ӓ@P\Z M'StY$N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'uf$'M8b@_('St4#H@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 t@/p`zꈗUow ?NW@?s\q MZq pmZqg7.ғ0.}:8c.JIiw⭫Kx!v,ٗTM%^!Z)ٗʮQo/:9>۾J{Zqmp[J~eUy{vpP i>jM9~?mi݌*cbޯxw.뷇5A=] >Zן߿K_Po+W߲K\m~8+xͷӿyyP*.~y @E~=Rn]1kGs<3)-Cbt%Zn`n`W膧SJKjsSGٟh}wb!]=A'DW K+Fþ Q ҕw'{ƑB lMu0vg3ٙىdÌnm1Hdk9yQ%6E]a',oUzf+;C * Pn}0* puҊZf}Z1BP.ƻB]0xuf+,W(\pjl;PNWVIcr@xW(e Zێ+T)X? U/DB~q\5,(=2ej߆f*E˺e\WV=UP j Pm i*YNBg+k$)iIڻޢ>Qc Sۗ4ȷ32KȽLQ]SP@?Es2Ck]N4B~}s(q __M V챂 n4ږB8 bYi eV/kP8 XOa B}?dEm,31mXZ}Q I,0'UugTsS˲L.e!Byw$(Cfɡ\Mr@$])FrB+HFBe+<\pji;stĕmFB:Y(d+P QSq*iO~R,k&XlprW֒ T֡tp/JMFBolprje뽫λ:E\Ugypex, P.e ղOB~£EN}W=ȧ lFQKz>Pemoqvz'Rzf=FjuWTҖJ5poSKa  Pfjo;Ppu&l TQ9dHuM|*jk\[s@qw)d740H+p[3f7\arP"mPp~leFTrm.BoRu:E\ ̩W P w"JJ:\ RҌpm>c^ Wăfjy+T):\"4p lƼPm4NW(fs ѣ\r=Tqe .W(X烫Jn6#ֶ~FQޱuPW5, 65J3HK] p;\[2euFZlprWV UqŨTe+L6B>OcHiNN~\im0 J=%H]ӷ./ioJrI|z4Uxm8[MW怷ؤF+u+c<ղ!R";8Uɡ`.PDrVU#9A-e$#\`OǓ6Wŷ|>PemW+h  P% z\JyW+8'9MBBe++U.BʶWRwNWIJLFB,`lATeq*pu28c29 XR P.&DWRvNWZ3#\FQn>##XO8'-8O`z:ɥG Z~f*۶Yi+jߪ\X30 P.g v\J:\"&]`g`fb:A#/L!JןRZLgBt|ŸRg/p]YOF\ %9Q|j>۳ڨ͟<=ā+fǩUMP^8Y8N}{k{>m؛SXelp7VhHYY-~$H/Ɯ*pynQeEg. Vh._VYt(pl 4-%-V% UUo}.^}p"h 򝾇ozfzIZ;\7mvjhghlfg{gcؗ%UC⪡ZCJ:W]5E}yң= [C˭PDbn]kb&'*p1N_k춂P)U%% ɀ]pjWR$w":%{^`WDf`IBZ~tT)E휠#pp5 lpjo;P%ĕ'OW=jFK7YJ0nqlpfp|pN'm^WRx]~GTc˩HpZ+L{Xgxfڇ=-^k2,$|;^xy5\UvϷg.sqـ6j+v W<.)Xj3BJI*h!2 iQ^>ɛ@sBHH,B!Ozտ}ܠHx}G-\ePzѕ/{!򾺌ϰq G0 tZjy1~ѫ`<Q8_׷o'uj?BzmYἏN{QXaU!P/ۗsmJ\z. >h/ogu{7X!ie/r L A7nv ~9?U9~0Cy&x ^2L?dǏָ==\]RPySCƸ1 !TZ&0`d0.a>@S6@R, IiC=q$zN:9lm`-f@e7OEL׃i,E`jSr))Y2C7Ԧ|T $FKH]΂R홺+#AzYϯKN\<ʣ$6[pvhlAVWڬ-}W xRœ e>æj.7mҩ3xjt>|,uv= ɅƖbrvqlw W-^oxjH7{e}L2էt].*mp__PYLp.+7>NVx’Yqmf.-[ҞP7Db7MK +c 𜲎k%ipPИ4-IDRQtIkb"ALP8c)#D$܁yx09O i6ń`l f#bG,L[Gɣb8r9GaUJJ]t.TV[sDDI R*MyԂY'DLh4O90&i"~!ғ@fz*3)&^Kj x4TJnc(AJ" q%ؚCYǾ<!ݏ/LJӚzG>090W1ꦦCU<1If{ࠠO}3Wc{E5Kq^xM?&>~UyPUs!b,FFT&};SY'1˂94]tth8?q205%sLh1S0uT.{$Pr]"| rZvk7ANjC;YN$`G3f?7V{ُ/n{ǝ=ȴGAN}#4ﱮh4-)w5UspSx1.駛W'Q{/43ro>!|&[ n`ZRScr%8ݖ JʬbN Osq$vtmQ?zq?:mU\X</seK{}wj-Px9ׇ aP2D)ciK2bV:)BpT68a$ssnd"Avdߐ a0I9x{l#X}n KQY8X+ U,Jc%thuj:lljKh10f" "xxe Z+ܖi! c03'x"ZSeD.4 3xeR11i5ޥ RQU_* Q" 9X]uH_dm~9NuKzw򴺻Ϫ啿Vt{Wƭ"_pѵb MۓEӠ  m5#N"w%Y,lYQ]gșgΦ^-ԸSoӳoۅ͑b0t9v_wn/!M-iؒ[:oj57cIc3 Nas/U|8uyohlgvټqT K>? Ga*8Gֻ~:OC ۺޠ]:z^?|{߼̼?޿y-8@EI-¯@C[MOwhڶ47kEӊߦ]#7{kd4o2f{aVwOz㐃>\$|rj^(pM.'(6N9|VQI ,BciK_PѯsپXk-;Ht6_dhb4&4|P$' SPDŒ H0Sye5{: RZQ`!I`$!\[HtKO$zŌQp!U1tZ9Pb0Z;`,O}Nv>azs㡳 Xչ<Wy0{)NŽpPVW0)?EZ&Rr})Uu1Xb !tKikeH͗W8oڥ๞l\s7k?ל07<ǓE&W8[U;?|b砼Sek\6R/۫U"j+b%\qx2_1cUg{aj-n^.2Kt6ֆd,ҔGaA[-l1 &Ƭ0rTdS>o0cm: NBSNɴfKi_x 뎫Вv$&uŘE:68 J*e iKԷn\C1U G_9?M4`8h-#j줫/!nPJFgSJAZQzytZTƎH3 au!*7 |;RL~d4_9©<ǥc9jDdR8K<+IOvPu@]ԽTd$c4 mRRh\94Ulbz֚HjB:.%eKu'6 HJ Sk yJچtBwBo鷏F[f-1^;8G¹jz!9{v?W6$YE2YJlin5鄫Jut]:MAb 2 F49ucg}Kn!) *jeC@咆~I9&kke~x\;Na]稼 >:Xӈ=0W_io<}fp%3/ZD +љd9K/TÙpLm**ip.Q9A&ǃ<)DRc 䘍!x G* (1ǜ 'dSV2Q2lMm%*|HR.'3<.B{7*s|cs#T} `Q tx*5|5%%a^snƁKGViY+ڀh\VGg)pos:%*EKZy74\ |A1r _`C3, .8*_c'W{s`qeҌo1'CD>:E^eFi;ŌX<OQm3/}1L酭zlxn[Dx0tPqȧmJ FTB_6PddJ!*- g2Dx2p~pI%燑 RN ?t1~.]2 |:Nw%[I8-PG '<0-s:w$sFi\%M[cQWrpL毈թUjpsz]vʲ:|;} a@WQŅ t/ϗu/~+e®ᥟ|8|/5'{!N._~ ˏOZ+l'~R1; ΋J2V*b_KyUk8sߕ16 _|3B2Fx =J;U3tZtVl.[;y-65.n1 ϟTe1\ԕLRmݢML.Gq|O?9o \p85hxzyTL6|)hH$^b_"*'aԕTjB ڼ &"{^Ȁ8S66Ww3I1 #9 4ײܚ8C6& wc\_z eP2#^ k8PCh#`bǹye\JZFŕps%߳FΖr6@̖]m6"Pn$ jey:g4a&73, .@Ei>EsBO9i5ϒj:{I@㎫ .E-P1'ڎP7F?iV;!"X"[`EV`J$*/r@蜎QyT(A 'v.cGۘiBC!'r{nʯ9ixq0p"=m.j""M-]K_ Ezc 6TnC_{Y)Жj8HVEnp9Z`Zȥb{`#QQƆ$c[-*LԠ]KȜ*e 3uTLl z\]iɝ=z»@odb5R[VOboy:wѿُCV)}Hb#P+6j+ ]vWlŜH] ؃QW\N ]]5* FuՕU#ࡨFFQW~˭?*S{WW"ب};#`ajڏzf`ΠuS|@|0ꪑQQ8tuը4Z)>$u%QW\E]5jQ ]]5*I%+Bgo΂֎3urhzIzIƉy/JM~Yl=d!w-1cWg9Of<[w?? \ ϟ$|LO,3?q-/u׼ұ5I~N ˿|0moB; H ߦn&"ukCdz 'ba'^G-b6=FQ(}<?3wvM뇚FV] F]5rjjT%+oA:m2EM"\[z7?̺ 8[+Й~8nc^3ug  @5:ժͱQ1%CY,XE CUk\Vn 1meޑd;elN$Ijsr'u>|9Ɗ6؉^CR'u8unpp@p\w74\IieA5qx.P(R(XUVC(16,iReRؗ*o6eJax4M݌q޴EK JvUlL模 .xc Fo4"$xcbr: 7&vhBhڵE([#MP5F;@#E݆T;΅+R2*,(7r\}~b7ʟ2#T jey:g4am}猆9k PQiEO\!AN!yM*Ξ~jKcNVoP+]~'=ӫ;FWDjyA*B*mrW0%jT{9BtNG֨<*^Ǭ\x81jOf&jػ8 zjGۘiBC!'r{nʯ9ixq0p"=m.j""Bҵ`P7m#@\ePB[u YxXkEi!DUxF؈b{`/U:la_bΎ$t[-7S9wHTeB˜ld?<e}h :m;2]@ЉHu bbǔ (坶 "RAiA+1wN%BRC&U}*)2K><$oG)YD+FMBwT6g9e*4}Z/(P1& 5D I\YcѮ(^˾ z]6FPbt.V6h "d::(MSܱwaϭc154{psFnn|nj׼TT?6P-?h;px.e ~+vkwG@ 0$1W(PM,Z^}v Se@-F<otz u ~)kMY+X&NhM9\}>~!K;iӏ\wW6oBflܴ;)ۦ`dz~C0s )!՞ ՌA[r|>WE#:cL՝Jt6P1Pjq'?]}河&i@I+ld)'VqrkDixقR D- Q=PZY)N<5zl5@V)[ FRX=egK@CJKt\] WxyLwC-OOE+B<;ɿq bņ-AeX8tѡE =;8Z_<6:XUŔD .5'Dgh(TugB{&~Bz&`iW蝘a^ Fe?k1O .wъqF8=#qh p8S\3Pu>QI~BZ2=;s,})]?m+I~dk-@W@<;LwFUmuL#7%BRP(z.ؔ%X9qbϘtbɌ޵6r+鿢]mYCvd$C20x)Ȓcɞo%25-; K&UW^{gUCR&uj HGT)~sb͍뻒3>Qdꔓ0t\WPք{yPBRtKl2)f%)Pz7Xڡbjy89I2$L l((i:]48WOgB 1J7]%GvwvZnÝY'tk¨bU&(d>T !GE^jPl{8{5m|nZ7džn 8:q8VOP`(]-=}>G舦\n[򠨲<?R"X+qo=* $4*y~?LQfP I,%i-}D)2)%ٗACא*$gEJO$.$,)FPI){vXlsVț^[Jѽ]fbHok4_uu~;Z~[wRy/`K l[ i7Fuz"=ĵxJG#ϼ8 '@ w*QMm/t' =fum3V#$ޜ&aqhq)n6NêA4]T̛˱,dʣ#]KF}t=`qrz#PY6QuVwq?y4)G;?y8ܨlxݣui?KT?.|^|qua4$Od|t[h:ph_]_]=Y{:FvwEg7ו5m':uM'=s|2^lz8JCvy]Ԏ5?gg6}>Ѭ*lf˅ߦ|( ։ˣ4>=fuׇᷯ!Hw^kq#TwDE0_ ']5Z9ܢk}>j|~?1 I8]N)f|I=&}owTNPا@CVʾHS{|q6?CǏ?&gx}I8N=Lҷ-;Te%IvdV &:$ t炚(k; RJ~I],Q 2H'\trDP1PM d9{[@ D#glոryͻbK\86uLǵph+ǜ~ؼSW|:n\ߢ70_ISsmiWMNmNBnm>Ko]^lNCu67][}{,Ji|n,<t~GܑxpYlt'~T[x98]>zw{H{̲R&5)J`|7&F{ qH{|G6>Q#o>KA@9~\z o5:etJnRO<@$_(ٻP !_.%+ TD\&O>v$yɾ[D2(d69CD! ,=A%'tQU 58>&oj0e# (k9P3$%Tx)OE2l{" dnn K.__;o =ɻ:rXl:D.pwsv1F[kM-e  D㜂K4Xͳr6ݎ;n_Ol6 NQyrd",Vߛ_-Fa7]q]B}Swd7w =K}x5M6V_8G*=fc= Kxo*sGuuy;[ p>߇h IIB68)<g'a WIw'sz+27;eMp-<(!R<`>̶Κk(SPSXT$ 3tXAλW٧p<(7K  h^>$)]tOǧƥ7J߂w|+Eϳ}l'->.(k__I~{r|6Z;tQh0ǯ{[(VhON߽*쬒짘<_n!>MjF?.^t‚]q#OOl\a,c 3=IMoo+9A% $TL刋Q . #_-8{r{{'f^~`^FWYEB٫BT-HػWY5\ s_K]3߯^)+2-\X5Nn$qb otaP b+/zmjn֩-0~2Q~ޔ秠;:TfT UַF[ᾴftoV *5y5wUng[;6~5zLyV*sܡ9Eq¢;N;DDaݛQq{Gd6^iDF\p<߇`t}H./p?rzQxp^R,x䕓(_xK X/3 dWí "\H/ Pp: ^ۖ=pAv!/9^T<_lee{՝cqzYw ϜNGR V84K^t7˻;pcv>>.u㯠YWkﵧ^J! ufjcU"MQWZyuTWW/P])`ƨD.ueꪠR+0BlU"QW\7E]%j캺JT+VzvBOF0OH.OH-OHeGm֙z8bU"XQW@.CrSU𮫫D%zH43l>g+V΋^qZ_8O: t?/lZe<&g α٣zJ xzNi8d0,S!3#twA9>LfUi:Ͼ7G;< o|@ DxAzK%" RDgKzlgdU5SUՕ`Ke'.(Eꝝ,4 r˦GGhI.DˑU[j}z3R>u*P3\Lj9[&=B!dGo;ij ~~~;snmr īV)5Cz\ D[PRK9 (ibDbQaNrLS"$^I7AHshcsXP69!/IbC^W֡%9Ԓ݊.(&"A!hp· 'ϑ֘`"UN()ya`"ZEY&D4ǙeFHߙgk܎?D ]X_(_iݺW^^QZb95M;=B3HE$S%8єRn0$x~.wJZ'$&}NQl:%(gsBDڲ0*- FX1i;2XUR#*#+TTp+!le9k嬐En`THҾ#1*  0("dSlNKX"6EsJQTr" ;M$ .Zv0BETs,=iFxZZdI4n״q`**"GȐQ F긖ih C A,"+VGjYkֲldQ˲^v2-3d Z ӸzxcFP>z4s}@#NC΃ag]Xk·x(wN5|)Kx\`GV%-"I؇dҀ(E񘃱܍'睑8d]-K:3ܵ AhWM6ru\)vHMD!!ڏm >zK@ r47rX(\d%w%v%t%.)KP@'Lf L%XE\ "Vߥd VqCQ{0rHRIp%@kY6][tIqAjX9`IhJ{X_V !a(Pu۶l $jͧXőp.ᥰ XEd|Q.!I__A+x#,@lT/yBj? T` M)}VVʁN=P9Ōour^[n_RBz 9ϱ$qEb^j0,bGEH o|&U)DV6<^ nu^ؘ1Oz]q8>,+&S8nV2>/ Sye)6j/&8h>9!9ց)Q9V!y ].'ꄒx}>VQ8s@s\ep `hJe5.|(BjdQ딷i&H _@`R7$4Sf; &E3+PŭD̾⮯;5=hm|#p,֓ZZv[2aϭU+T,^Y<{U^qÔOOp{_fѠ*z[qtIMrV0isro QDbEQV4b SȽ [dOrm=F-C0!B*Caj1LQZFL&Z@iBj=k-7}'ٓwx0 N >` $*Q0eȤR%{)_m$ bA7(8]`DʩJaі{+%R4m$gO/Lysc5>'U{,&9f!DT׼xr7qa @5֬ݰ8$jU+oqN ( # QEpB4sQTpn#Q]aJFvP9-b'}*.}Ԗx1d ,3AI[#g*8ӅpAQ M*I0/\`@h]Iղn5 /Z5V-kmi΂ƒD[m)El 1T`I.}%Dc`6T O}AK s,X۪F #Ɓ3% ĤgO)594(&LRRXFj-N|j zqZ΅k:[{EՋ:Ŧ}cznA03$+pj"H; 7{?mq}!nY3Dx*lq!tUZc@"6Ta9x . Xk;~HhqR H0 ? +T`cZQy ֝j \ G00rdM;}H>:9l,WaҹSb:ּ՞++5XSf eRH{b L9=ZజD[GDJNc 蒥 3âҌ;dʔp0@}' L܎DUT(p5,[YK=0㤍N .pе)S_JI /|JνSh]&}d:,tD{B'߇ 6HPAuT`@ւr-]AY0IPs: (B4WƱq~tN >nҼ*96p}݆4s7xu:6.|+~b%NkR==?Va-%dP^=LofK~ |'.`~;Q- žZӚ4H 2P1:8R33;<@.@0! 8<@.-0n}w#'ljpV!@ؿ|a`>թÏDrپe-ƇJSf? nRd]Z: I]?a8MA<*i^5KT'?g/W XřA[c8R Gßgy%bpa-C%IΚ!h*a8 F0bGe=elS4klͭ֝lj.Y:Mꡖ0~4bPb!>ŨZumzLnTO\UV ; */?xcL/f̂KiCH/ kM;4jګmѴmz>۴rCkTҢbDQC]Z&C$GhYΚ~1?ͫT}* &Вܙ*D1]f oD0})!%?5OIHZ )"(\&B)#f3"@!2̃-–Gڋ +wKHq[9R- n`gL VW&$t{:L|/g0Z{Xݟn }"U y$^VbxR[2EeJ"a"H{##-zY|WԷK{kRly &,dYZdiqЮ3\?N? .]4d2jK-w6 u̩T#{}Qxhi?Et&bcaUŽTRI)vaV`-ȷC:)Q@w,l/:H_$י4{KYxeY} yydm:QM bn>PBJeUzs0-.<߷G3W%K(Ycm_]yY?V編yf~F꫟n']y x+pܡ9}Yw.sCՙ+I&Peʞ[?oY׮on)sݵv8v|w卓SwԁnQDv[!uەw ŀ$ւyeLPFKyiGqNywE%瑫HM7[wm䴁`|%܇IbNzsgyۭٛF|igU;kZ~v1_XxH 4qltHk&ŔzJ2V%ịUh4BkM̨3Z F刊`) NPBj;Ek.7YO@B|u&fr%γ| RpbNtY"72No)JbNHIu%sTG OBPGȏT{G}s]-?˃QA~Y﬉ 5 ,Ip%N#F0ǔD0Dq7oΓ0km>q,yw(TwsmYw@#)J06tjxПEYg>qţ-bDJ)v2e$8n#3%b*p0g)F0fS<(dv;Өz:4a棶ȥ玂yCJ)ws8aiV4O2׍$?vSAk5_q(b)g(RāAB:!qMcA(ӊ菉5K,WMhpHmD(j۳ lOdKRJ< z`~鉫qq1$ (&HI1e +$AiJU M HF1Gq|T~T<ʁ^ 5Xɤ"y3TW͸bND4#W2>DxcauBaYyhw+o$I;a&ʭ9,9ɬJݗy\Py.G6?񜤺g՜{}o'#xyJt^7JO@iMNAYE*k]Z{ӛO%W= +ly yNCl<|"k\1q2[s |HHW˧ׯgX%ѩ/.\؍T|d5`}#]&7v*v)8k6SJs*P 2u\Q sUJ΂4SlE+Wzc>|[O?0GMtkn/6#ao^5ϸ ke.eNF>ʂpko8qb ?|UJӋ_2P~$cZk74/k]HlV)^bw Ǘ;;]o9W8` 8e6 &X`q,mol#9a+vKd[Ȕ-'=@2܃_үGհƤTKdO=y"%uprܾvPG"C6">۸@'+ 6ӬC>#U@3R@'%LA{BՃciFBs1̀1LبBs>qbN9ܕs< SQRA9T0@ &#seZ\9µ~Ȱٍ _]\θ,G4.^^Ƹ˄.ѣL.2!3맡׆3uJϲ0|kPh ߨL 93Ba~`b!]#5 |"ܺآ/Nr) n?˄3Ͼ_v j5 _)XZ^L3s]Y%A.$H<؝Ocg5q.FL+2ȔBK'2%C`R\1$|8?gڂ(!Yf f;β(`λf6̄y6ݦC d DP&5Yڄ1"3uث;o^Yީ-fw&wS+N p@4Y.mg'e& :$ us1jZ,Ie㔌΃h -MoP=ɳ*/WkΖaMh|Wv ͥ%gDF #*g2O/T_/_2Y*\\ײTb@μm(Ex|IPzBK1chH1,'gSLl9N43q[,M AamtWAe,,2#8q_j*8.0.[ԭ~H~͍,U9);"eA \L]k$S; ^9#<0 FUy0Bs&j ̶=UFys5k#',Z-\n~_GVyx= gs5-B/_L?j7”.i`E-G8f?@)M8Kp3Mt5]݇=!Ӹ0v}+M., acx߼sqJ1&+W֤5]FƳ,X$?<e ϯg=jyV:{U|VWkEuHXJ%c(8G|0Vwvy$VjhL+l]) ?;zP~?~ͻ~x ~w4z &UMMa 0ho_PU57Z:EՆ|ioS+c>%m̧ր,و>7LߟM(]4r3kvkX@hNWC#gLǥ{pZoebrL`%>h7^265V2Xˏq$pWs$or2')eh3W0#(a*f.iC\p8pOc GYGӌI1Ple'G#3I7\tg1t ]s\sK=yɄߟlf: 'C39֫6ST9yQ%xY# 4>U S(eCQ:is@k zJB}zc9΁X2cxd<,#ZC>L D:rK@)D8xYN1)}(cquzfO0$Z]\h+L%i ߦNn }Dwѧ DS5/_5~~>>oֺAsa] ;H뉐b\}=P7oDR}9sz)^@H2M"X+u&:Ԟ;-jshLqIcdpjE 3)*.cPMB.2KeNꅧ^%<ǟݺ|_OmEP3bZ8̼znTIsϧ&حϧ.[s:] `MwUCI&uI[!c ;' *#\肷6kDlCmfe N{pc"tq .{JG5h4g ePY`'"YKDVZU-<9AϹ>R^C ג#1pk`d Be ֜|/ *z~W19EG%WY>ԞI]ߐ:Bۥ.SǤ^hhBQ&بȅsu W{J))&܉dNfN 0PI˜4KQrA7W¦S>_`ˇBs|:'A9ɒa-ȉA9Q#+Ȫ**FHHpC7$y-aU$#Bj&e*jl6T< vD!6Nvv|ٻ6$eIn u{oc)Up *$w;V! 7Cv}V5欹5/ @8ue *m1)pRӅR%X|cbY`2Y,Ag*%yF&2]Hڥ0b;Gc-q2#ۼ阞G1Pmtc;seY6y(ㄪ߇# { )|tSP}ҍ뉻!.79ٮJ׋ű8XJQ%A[+˒*1_h,b(dHQ!Rt( $ ݜO#$xmMq{?w!z2 AL0 9k<S;Tp&7™`RyxZ5l_u³g\2峷47?~Qpvv`+uNFxKr=>wN-ίRB8!h)a\0o嬵Y_[Dl=/h;@C}}U:z;];}ȋ|ް5Z%ꆙ;PUozّwm~5?ڛw&'x|޾M볞|^1={qeO΋qcd x4<%<_gWo|R/= I!7o̜38oHX+9h":|c㾱es/s'gP?bHNEY6Zxv(BxtV}jޯXBLceCk <$LM|ɘFKRmD#^q:A((XN82K^s!)`Aj'瑱X< H%H\/iG:IpE)V0JPcejlV zZwV7-tLH>~FA9gkuMz5EYknjp~ͼj+$T. !kg'%A2:z.c+KVS\}fYފÌI~޻p2}O&qkzoȎ_|1|-o?]|u맰?O^Wﯿ'S˛W?ĹAﯹpK[dB`ۿ:2gWӋo~]N)`^B/h/X78_ÏȠ,S^RӞ.,I:ŧ G2a/W2р t= 4rt{KgY2˄ٟ .tׇ͟ {jٌUoafɂW|N̯GNz) ނ%Yx8B 9?_ҥ_[v%|kC_$ :;FZSII yΩ-X 3 W>)5VιC blܛ]ϳve|YVeh~wdW[Qs;!ctf[w(]\i4tjN0k(K7 y~G3mp?x$G˃x@Y{ChNX,x&vک}eͅaSܘw|Q.mi斓z3-mQ3Rf^9c<ƹt2^*BhPxm~K*!|g+ .b9B)jc 4F%t%G%D!XH,NUQzc>h8iM;Qo<>uq].-a֗(\/œ"Sd0n=)deZ,6lkH7YD *Ȭe۰3늙h13n\R:i>X*^Do 9"G 9̸ JhR٭(B\=w Vy!$PEtJ%cXzRj/. ȵ":ZKYzz;<8t|xOJم c]ށ'Ə ~hs«` 5qPDoo>4kX2G\sA42JR \\f<ߡt!nQ:OƵgV26%h53[өzKw]U^Qb*&֖irHPdBALt Y#rZo{>uVx /{#%Ч~j|^X@}ބlB6;mzȭ^^68 ?VD WV丛eX[dX|Al_IP]ӟ/ 2:\\8ʆ~ 6Es׶{ ǵ J2 #]".el±=,ق⍒)nG]18Yip;Wdr'vG1˭hAAU'#eQN0Y#dz(ypN"u%f%v)-m}*hգ.vqw $2I"`2I`"ra&eϽՆ4|6 lWZ)Q:ۆ؅*j[Ь[:|WVCH!i}9~D(.NtfDX$% ȈG3$D2;QR̦j%+Ip'g{*Z6`$$ѢnÄj6z4A$Ź.! r6Ux0\p##Xid$;nDŽge%ْZU#r՟O@[#0jwtKܑAΖ>O+A+ Xf`9T| MΠ=dɸD>)k0踡VL'7mxA xHI:U jI:?i!sAb v6{2 EbLʙ͵^2),uةxbˑ;gތc'Syn!ӟ)"47-BJ&${mLU Y'm""DL/z+WEvM#K1jNAG0QF44Dł-%Nk93i!a4xXb+5qxegl+to1z2.5^8\Qx4EGPe\2)%e%bᴺO]ȉTWc-%IIx`L ٙdv^+Rlƒܩ4 oᝡ 76bM+79P9O |7e3fڶwqQ6Ȓ"baB)s%G2޵cٿ"iLa Mz1A'әmP$e$KʖUUvlAWRzP佇RI38SKIΑd6'ai,}JEf4r<ϝeLpYFƢ\5p g Koݻ.߈vQ]7~z=kd[QivU]vplёԑVL,g-}?-4oŸ$$ $/. `.bppofvCSOz8`NzGl7^â]l@$@xw NEVb p 0ed#8Cc-^0\6˳[8| CA9-M^Τ(;?&xgbk(g:&vɋׯ"iCh/kA/?wIٸY\]çe CWMVSR閮 ] w"JV5=.ǥ#v(OM]- tkcĤ +Hs*)t*|tPjUjT CWn ]:]tut%E+nMn왛r:M^T^- E {a ]N&F_݋gx8-.uК}=8pܠu( )˻wB޳i+'3pK.ɚAX]g)޳O td8;>B#~q؋}*#5zʀO}H|XJ O-:>[T{SzK99a0b.|,!p'\,|F (u4ut%b6BQW.VM+5b'OWeytaH40 ҘV<]JYҕ&4 X7\BW-NW%oٰI::]mcvpɑj; CiقHKWv=& 5k ]\xS*%*3+B At#\u |tPJ9ҕD^G̺Ht &D ¢B#Yr=t({ڏoNʗZE';,:٤%SaS 4ILk9R*hnKK$!d gR4iV+Ԝ0'Ť)tОL@YIo|SMi]4.C)tqtP9ҕ`&WU@s+(M 09yв (n J1N`5V5kNd+3w*zt@tut´I*짎]V]=!v= ȣNc82]mVV():/9-芶tsk* !` ƿp#(?~v6;qS{/Ds- ?Tp4sY*hSjCX5+I>\$z?>ȺI52Nzynz]t' >2ӫ4Pm[b\ L2Z"Mj CƗ\2y;C1>?Hr/ՂM+ bn;aj+OQw"R=P씋?n@[ FFWoI^{~@ {`W<]c.U/V1тRM(I~,thryԽb3?$}`#Xft{㲥:]t|[zAt71Cb'd s0/?g‹B'{0@%53Z:[ '=b9DEAVnh(u7fv&=Se1":}5?fL)F\Ch0WOߏVPm6F2?%.Rw8FܳƱ:M]/D#*%sk:{>]_ iwTX ҠXN\nkv2%g'JN$֒*q0HͨkR{?~(2r!5ϸymF}mU/U3K>7M6KAwMCEEed˳C5Vx] V{X^|$a4#33SS|aqWI츛Z<WTV hAseIu1X؛nI0"̕~`YG )ھuMn{k&0$C CQT4 U,)!Ǟr٤ɥl#jtٽIW>ňCЈs'h% 4sҘW̄Lh@Y&mgBg&H+>TSSP.㝰 A"~rz+B+iAPcMhHy\R1>ƹHcz,}ޛ(HkXLSZFI2?l{ vjsmNQp^YulU)ɡ̇\?Ņ W5ܟe_y4eLբ|ԄIn<2K!v)h$qrӌzcP۳(yY!y wY _UA|3t ߆~I7@w.;|w/\oܬ??cWV(ȯ#gSC "mxs[_]Vw1[n*c, 86V{,Ɏ{{yIv\83&$yn5ey2Ls܂@#2Z9[DS| qےOQz:( kre]V`2q0bqF(wFg]f9A SR+%9/w۝1M a:A #k'ˇ[gM> |p2ZV\Lr.ףlg>*{ܫVOa\hq ~_ҍ_Bn'wp#ci?,uvVo<0;Oz7eobr{pƫ75X`\>w=[_2CE>7Fķo.ZY:K!bUX560BX&C ?xaMTg] XB@=~Oe/Q(eQD7y]M (&j% s&iR-Ə{X7¡Mbz nͥW?Mn`ƚIT*=&)&J {e.]Bԁ+ 9u}F+Z5{_n,zZw׬soʋϽU2!un׵*Cqq:n6Dizx$[ʭ3\ 1vy<ǜ2Tk%iu-=+T{ޠ7*UV%wW2 cCLd'a\˖ކԁȑ޻/`yPB ؿ,nF,dߑ xQUʮ 8rkc\R >:U]UWw?;{a.gQP]Ҥ+05b^-֐R8RO#*kzÛAZ_cl<O%h1f#$1%RMHs5rEP3gIiYRujs;؁,wRLqBd>gYf@LR H.g' g0bZ, CQdXeP2lY$ bFQC"eLtsyxβȓ̌ _.v jw b20_F?9K[R|+fQ (ba湢ZB&F(3\\RCm&6TZKuI#]H;ah CR(*,3'YRa'.UWrӹ{)˵~\kp8)DFY뿵%Ư-mzؕ`q+(K}cRύ()\0̡TlR%B\[^2bC#6tzY&섲LXM[}Fw5;Th"'1q꽢{<:$Cq$9EhLZd { Iy&̫L# 3Aq1#$ Yԣ?k zBb#\&/O$%(6]z+%Wn)k8#béZPF~A,hQ\0c8!$1xO,N;J``+a4rznmxz%JfVH,1L:/"NZ%ϵQxQ ax0*N(Xj'uYȮk~IX_f( KxI:dtRK&D!RQ mɱMԺdjzQwx])) ,%aIS dFcOI'BN(%))kTV0h̘1YHP+ߥd,9VCKq6S}T^r$h)m5_b͌VH#UeKnM@CjfUcITVM>$a&DS0P@ ˸INtvjlSgG5KW yj{:R(*A(jRʰC 'kPznF-Exk|0ԧ"U-Oǧ]+wq +ҹþsQ:s ,~1eνzsQza=wJ[s,n08$obph˄:Ą8X@i oE1?1CuU{RD-UpRTZ޹hɄSV՗ΐnk'N<{뼥^ FmfK7F<1b"A+  ]/'r]q+7=n|:KhK(w:_>PUyȤ`IHZdS>'͡$8Z 2衅3oޭ7ۍR^.ON]k, )RRLیwZp7_J/sӽD$2'SK+Q{oJˠU4hK5/}nM—󎯠R#aD43cˠ!3A0Ⱥo2 2Yډ(ͤ 斪ƗPjZձNm١"f/YinR%k̈́HUHqHyĊuu@mJ8r2dXj1uu>x-7(nVoLZt 2Ҟ1"ji2:?xq_[|G8YYU5GbJbyr!0ݿN-[s; 6|nq $O;u1}m^v=͛˓!kuǓ|ELM}jN;IC>&s߽7שꟳ+?M/pѸZ'鷣z|plMzל#{g?z[>c>UjUO0;{NVuWwn/_Dh M<.~vzu'W}]܌%^Zbz]9uҜ꼐ֱ󷳋ܚ4έۨ k&o^>:?'QG|ן~l~zÏ_~Av?V G៏o]^]K {t>ڞS^7TҲ)=-w*@Ƽo?Mfjbvrk ">jWSקe,D-b~2 x?yQ֨g"LGOT~ &K=>T1˴U ? Z6HK_0F2AJD٫Uh,J")"X ҈ٰPy;X?wYh u)eB zo5TB#6kD_дL˙N&^I!.ҫI/)HoUcH ̣>[K O0CM1:4TŦ6In΅>^#HFtIoE^_/^7vҊ:Zٷq \SlyV"L:S a߹aVNAߏY٤S"`f`R ) 1wVƾ91{W5*$\ȂS 6AWUJ L-*0y>q݆sO/`7Ml||g6W!ڡMG MV?d_!rm6j{ow7/t=hӜ6(wr0_ov嫛77>G\.ִjYg~ǯWs?"&Ev2oz͑ri.G+lIWבۈ͟ hTlsuERy󹆫N:`w,)PLQ%(f< v#v'-wHUmS :fb6P! =[[jشͮ}kF<Jφs_}iO'san kvSZ$>]9_,P#E|+! "tk'1.lMྵrW )EZUM$U+'AkI"T5Y gig9n ^^{5"unL1qň,YZo7IS.d|ȿuQ,(3@vVT8 (`G'`{dWUuԔET*ȫl3gjM1*d@Ebrap*{lӹ`@o_|9crXj5\Ei4f$ `\Ab"htɚTߚ-VhEWOΐْID \5[z gi1KvD.Uu['̫nȫڝ~>r|4|ہ KP2Ԣ!7j=GFSȄlR9,yԮKE+C$#hjdKl*e5sU9`gm856ksWvP6Z{`׭UH} sP٬iZGD4Jߨ*”DXAb 5U5FʍP%*AR3n٬]nqW4bP;k8jQ#eF3[[ FMF*RTR,r=J2QW C1;+c>) K˞=FgIkL `ܛ1p6k/,D/N91Jz)+k#v"zΚGY(9Kb eF 3g?VC!tև$݂ 8"_Uݺ 1!RQ<`ՅNA吣KAz@SlAZjcyЗQԩ |XNs+\*lLFY^&A %FQkkZ@utd#tKm:`۝Wt^dC3[]X\ LjJ".W@!B!%dkUdr$¥@y PKX r(Tw@oQ $)f7Fqʼnfsы~5 dUu[)FaنTVk>v|TR 1.2ERPjdbݲK@&0]̿F<\Wg@v+)qn{ft=cN{3BKv I^xKǧSl~ЈfI@/GcNc}\={'`PTϞ+qZ`Xp ќ0T bA9ƹv)*h૭ZuKbWM>(Su^lBDc4VNie6!=>|q!JNO7Ar&8>dLN/!V}~aL{kq"ɲ%ŧ]\čwVahtϗB\4}O.#EBB̓{nd<HJ#aݔ ƛn^86u/RTIr"QVee*)[r1dE NУS =q[v*QAdq 7 ^v^|$IL~S釃!cum4@l2 )GBLld,Asmۉͯ$6톋Mb.b܍ŵ#BbZ9fYCRIѐ.[* 3 ^b_c6(}+m/c2=ںeGo#qwT¢b.FQqذ&Qeȶ vҌ!/ v)ɘsЮ-AT #T[xG ;bb]6,c,PHA8ex\$Xv2H K8R/i#N\9s] 4Z$ p:1 s YJ>Ƣ7r_z{~Du<90Z.3Ή&!E!p,$)CEb.^+#ag:;zvK=;ޘ>"Ɋ"#ɺpQ(T7<`8 ]"SIl<0O>NݿY3DVH%!#^0#DH pL9aR="Rᴴ~#gmӷzZZu¡@H ٍQ @ ]4G9&-Ͱh$3Yt]ITI{;iV+TN#ҵ#J8ɧz[IUg!pZﯾ\?%^8h(TtͦLJkxVp<5PRsx^.x}fIu#p;Hˋo/j~b6=dN>ɟQWO\<g{ʽ̓wヴ<~N{љӏ'xW޺֌ã~[WG k?<ѓ⡛T=T-z|1Ӽ|7>|//, ium,ŸX]''^kh_J:g?t0\ io^^k:N}>M=g z D#&Uxq&v#>X{zgt\1~SU.V!Ue ={t5 ,Bf%??a|QlZ4Kb|Pg# KOͰvϭ©!8љ?PeUoϓ:\&婤´ddusRuтSـ||7p6o;8}`e:S,\max˗]N2N4+ "ջ4ݠڐ} jBT{wzqЋU(5C-⻋<+zzsE[TZ/{J 4P-7tlw\š쇭hy 6Mp` }ܠyix5uQeϮǽ{Ok*xytrM4m]UVm ]UNn ]cn骢lGW? ]i-[DWl ]Um ]Uo:]U+ʗ;b[CW.gBWNW;jk*`=}WښְM]t [DW mGm<]UvtЕf=fj;unFhћ\ AWt%wtuۦ'Ȉ-+$*\5tUUPJ1Ξ4`-3NmNGLyԎK34~3-&ӝFD#2 #'gIܚ-im MWRn MW6+JS?$M[Mle6{3)L\ {i#h<~lG'0ɃF@[w #Ҿ$чs8E[?+\i%~VZoz(0~J-+I WmVMR]tuF,K.@U+nbZ|O&DٝtE$=IwЧ\U6~2FTgb><>^JE[cnbX>h;}M <"vqe:1srlrNh%RDc82c$g ȗNm׽hQV8'M 9N߿ˆ~곥RwLJ mbDċʀLJ()t*|%`6ɷ3 ƕM]u#B䥤E'IFY=&]}+,`!q/Ȕ Z5hC#6{ńq)0xu F[bFCëkJtxi4Y\yXa buQ9yK~g|5GQ TAz#xYHoS[GXG6| %2߇!T85$XC!Qs4З l*딹cuH" 9d/Xi(bVd:33:ʣ"Lr<{ & +bڧ^$74hB8#LHI%* %q eP4&BZ9֪1rHKPQ.|`fZS !tmEWc jJTzrRPyUN*q3 TypL *n#[:(TŅ W4:KB}Yg60#`2{0eYhR Yl#, hQs䄧 /):k[POPרE%DrX*I0ϊ$$wiBFKlEIInH +Q u-`y*"@irTf6W,r,UqfLzV Ev J E(IfH57]6ǐ?g(c)HSoUĐ+ $8( ,n IȢTHqXx@7 *<{\W /3R9a*2^Io@ES^bY ՆIHyFӈ 2H`B؅ANCJ$BD$"/.9̌6p3 F:VѰgzʍ4WI,07;sY`7ΕՎv也m?Bx0N 4Bo Y{X/3'd㜤\QcY <2:)yh*&^O<Ȱ=˥%E~t;qI8: ,G_\oG۠Dɋ!إ0lvy~\1p}NO]^^_\]NM22 ƅ:f\tɳ̀k0Qn`˻S?NV#?fq⏵k 䠞#ke;q1(绣Fo̠۟mhEpx7%๼'Di%Q?ӜC7iL_ȃ '))7t?a;C`4r0POid/!rM\z/"Nԓ0ӀlO/No+"qѸSnJr?$oq1c^P âĐ)"d0BC7~֋\tm'(0FnGrS\sɵ10ۻxR7c>*ɻqs%?&yg2e$ @a]u >%??&?TَP(Ȕ?1j'/pDdAđ虻TG$a0"Ioqt)#^rջ_0Mp !)% .1.E5|p)=Xpi _4="] 09qMp#hѕbi]WBwj9J<`JpQB? R(—+y'H8ꩮQ6SJj*K t%ѕRѢ+%ZוPuE}4w%EOu%fiCSL LW_ȦǍ# 纪s]ᖙuUGKi]P&-]atئ :E`jt%{EWBbJ(*ʧ'ڶi u OCu嵤EϞ[[w{Fpxp`Yw]}ґ?w8w߃ke# +O.|xPr~{d%ˏ;a۳OD0˫뛋4 ?pb%.{͇OR7ջO^civ1W]X[D?y']?lx|A 5]]x=nQ+uo.3 ueں5QZS7nײaac~RƽZ -;q;Eh=]ѺǞ x90ԇ4K nOvbW?OǛMxO=ʮ~b^K.W//q=s\Obߌr( a=ֱwi!SbJCׯ P_cn\oNc -;|鯩 _}㕏c wܬN7w @_vgnV]7g;n9qe0Ow9B$НvEgGr" y|o߳7_v gM~N/*y?svR0 }~76xgmH7Wػ65ݼ~7/SjwwoیK99Ej듖ZMh{jBj%jrj6.\8piS(su9󙃴[ە-q1 ;xt肢AMI nZMA([;?"iC.x-ژ[וP&uB*t%AW .kѕzj]WBuǐJId{:Zj~}Sj4J)&杗BF+UD4w%GWԬZj~)Sk'^DW%6885ܬfJhKA$ztlq@ %5_U\gVR6\lzl{ d몀 *] /Zt%J(c6]-PW!s^xeusocWG[j*.nU]x?z(^Rs>i.G5#A +{Y4Z6kYյ^8ZYaz?}PX<9ZFrBPBGr|J1ѕঠEWBsʒLW bȤS]pICʽGj9J%ES@7{-_ %9u1銁+jt%sU\GJJuU06zjZj~)YuD]HW N+EWB/^+ztUlq<soP3\\Gfj̍xP*tULWmz&] pѣ+EW>J(3]-PW!EHW!ȺSJp>u%LWKUm{<!l(H U]w{yR 0y 4}R6 n B;>nʔnX`SHt%qZt%sRG 6D]aaNV+zfe-b蚯'u2@:@νrb}1P+1ѕ"i߼uU-~:] 0ꙻJhKl]WBIt@]Q4M Jp1kѕW e_Ȧ+>ɬ0B:\y6ޭN< ]M xJ!ѕν67+,t@]ɺerMTRu1|ËqŠ iË+9Ȅ\~۠{֤7~IU?Ͼ3zcReke(*8Ͼeܛ\εe%d-p$%JS+%EWLK'ۉt]R]ꀣz^65+ZoB%jҕ+vEVAu%LW UrV+/'t%AܕFߺ2u ("]e Zt%!+uU(͠W+ NT+ 6wD](HW SQ܀Zt%Ͻu%1]9?ΏhzwϽYU5p5\KsRKSՔP5]=2+ü qicuŔEꪄ7sRFמVǴA"`U၌\u-_W]A-j .4a`nYͭejnX`]]EC5\Zt%l* gz)]8"]1pFW[0-a:rtD]%c"Eb`tzy /Zt%[וPhZ"] pң+Y:OtR]J6 z;\5Օuљ+%Z)Iϊ"+ tD]ssIU_L.*uu ͎UqRmzg-`J-C%>7s{U}!`j8ZV>xe0͚c9@=!QHNcT3ܹuyrFr PӤ+FWI65?O.LW :M5:%?'B+,%*iZ4 rѕjѕ>OוPjPxEb4{WϢ)m_WBt@]wt%JpQ`Ph+9t]Q @+!ѕBѢ+k*Jt6ztlq1bs+qg[fZ^EZUU0]=}L*] nZt%мvdZB̹hxFW] mMWKU j\pddwqoR'-T!Xį6#Chn`jIK BlVfuCDRR+&=b\Jh}n]WBjB [\?UWzt% JhsՕPrTWJpJhKi]WBIt@]el"]1pTKphѕu%{5j9*ӴH1ѕ mJ(K!A4d] nZt%v6ztG6=l\ZgSui湫:ڹ:JjN>I8E5ܒJh)+ .rl`0XOJWTWBוP#5g]k/`pd4hw1ETȸ=5r㧡U, Ƭe'U%x4ù<&ϰyRMm%#c잪ߏre_x>?}w397owwbB1z_kmFEX\`q7 ,M[ldna,:w~Go[/+DZ*Aj~9$]w49VT *dV\UH^H/8F9$Wy? ScRoojYk=@iSFগ&r!|55^[B/~ f<\_^2y<-oG8k|leYsEye}빤8Un7G`Q9T]g,&!}G&DNp]_?#ɾ*Mwy* O` zD2Q ,8 A)zʗ\,,%g2@Nm׫Y3EIg ~I=0uof]|>&>'VKФyfSFմ"׃/_'/v]?\Nj؛am T7;WƵF Y~s^^}W*M2.Qxëq#ZmcN]ȭE#a(*]uG8͢ԏ>bп֋/e(uE$vԜzy͢nںm@"nSd䥔a4n7{Y̎'yb7~_/&ipDi|SW;-X' *f4jB |2&2璵+!HmAުG*s< -:=Bی\E-<U̷lgY;.=o`w/־Nqs\)&Tr381gbOӄ~38Y௫Cٴ@(*+B$0Qyd$p'B\ C(*KUJq X"[ET,ʈI"l:5Gc'oreC ' S9 1{Ţ(8 c} @y`*JF9@d *[A?+(?%({JӏFE}Ji#ЉܪDx%5C!{>q&B)Ld f]"1hla8HKXvkq׶-{m{'M}d,:6Au ,Yӌ2D_9w y0ZiЛv:2'R !2I,: 4k9r5GQU"i-pkR?Ѩ8cǡѵ#<6'6Ii2 tr)>&Qs͹\dQps=h@;ZoJ{|mggG{OxУ.~Ic$7ukz1Ky.m-7Uգ'^ Q꧛vrz AHdȸ,_iD"pN`>LD.NEoSzޓpu>r z`=d{3mDapNafz_lL%0q Te,0sYYSIrQrS{&{WA;jz쮦vLSA7ΥȹMe ֣' "Fl˥`޶:st#WƀGYPI2xmBNȕ34Bc֊'g{#T5+ц3KIzDZ~r;o}[N+ ?t7tˊDM\>e q. R%#4i@Ƞx LcSonѶK#pu~M7o>$۔sj}Nᛳqlc66`'jҊewy.{msrcmuVMn+eRg6R2V>{k s.Ag8uiB';mI'>L^wbo)o^wo}77߾c4(m*]?3 뿼YfY +ZC6>6!S\<߆|m Hw?{u,f!͊GYhYĈ{pw=H4 r~YhJ{r/SB,W ٌ~FȆDŽyhM<#Ӟ sD0 \ H(SEU qb,2˒AđdUrG~&6ǬݧFEa!dB.+4V'+NxZieMձhM'܃>C'>ge L7ܫt YNa;#4w^1<{ܸ}ڄ&JivlHod3V:zy཰ 2u!Y4<\& S %lxt& 1@u(cqus^`gRu?5>>`,QPF tƔ 4"$lYmj w~7| oM)["׫"o?:|k MmZ3ЮH?lOH뙐=dHk~ibj\4|څеE<2"&bPx2}Y>b-E}HF$5ki$IYd5fB6gͭh ֕ h1Lʵ{RP> 1T08%,RE^Ng`70yPg4}/tL&q6Vikū4jȷܟ]S{_MKh冯 <Q`tLiPw)2n_ڀ:W)]I xiT9AM^Gx59QR*IZ$ ^9\3 V8"Q'- RP`;(N r* p1inl2 (+!VU dkhb'D`9ش 69yTڋjBg`Z9HѢ, R р-dA ZOv3'51oG6>M0X}Wvu(}2MR^CL|['!*N:Adpض4c3#=dLF e%9 #]1×s C,pcB :a`4FD.yɇdGbW?{׶_edaJ}jbmg8ݛaZ?3r[MrFHI"K&Y$[(%PqG$&”\8䴁k§H+c~ĹBRMztΚt*|.wj_hzh:yVnG<.0ERR(0>jϢHZqmyV7 y7Z Oi3 cC,e*UƬmQܓURG\I/udZ٠8)t2?A?@y`|Hd$%QD/H$CKrCtL; JGDi:- Zߙ2IfoV֓b^tA Lk-pԱk_ċ^TqGI3"6P&6٢ΙRZ)tNŽ,z8ap_oT츫t2ʉ%_Nl[umzVh:_&f˃_xWJs|JPi|;>xJB6ϫguX41*PVWk4{=UggMߞЊ?ř7df:xyN>>ͮ/f_s~0&O?|kw-#_3ǿ˻~"+s+964a~2߯K\u*#*}g5=)ثgEƯ߉Œ?WX=mwGY-^Н\{2Or +^6~+)տ7bC_L#PhUW\ Ec-Dku"+mpGQA)jۋcOJ%l?i~mXFme~h"; ~x1jONت>jj—.O^S~0ӫأ_xzt+ ;>s8[k5[I ݭ@I pie $IV2 ֳz>!2:Twk Tz҆RCK (}t< m`^߷iM4Fw#o?,nPI}ΙNf F_l׍ǣTeD}OYzvG4GG2SY Wt[\LF_`@Ռۦ mtWڠ QLu_n.1Vc2w4Oɲ>3HWLlLIDr#Z KO$TȎۘ|t N+Cٯm22I4J\oCH{a!`1s!) A°:z߭*Ţ 2z8,8$0fLyCJ6 cL*HSuyvte铲xZD#%&&9 `M#JցJYtNr2Ŧm?O{A%吼ΡH"Dn!r lvh87%ГQ&W|пM"7Ք}YUlJ#]yl3ԫguZGK玹㪦5}%'xHie2*H:*k50?$~ UzZe@XIFӄ}"cQgOvaR9.sĐLؠ$cd=Tnq 8Y3@MC:<2KN!20^yS0;fvs$C5.)ޚ,|<ϩ+/6RئyQme *_@6g3*8I`k,HWiz땡*O'2/=%-8O,&ͭHJ'cE&FW Sg('}PNeoAwq}zΚ.]K,s웇o]m%TkZ~>n4K}HybiFr<QZF+ y`G%܀{- ]@7;t)*ev*H,1G3sSHmkRL=9 !.j ${h~-*/Z;9g#&f}\Z7_@"4QiW}LXHd 9Q%+;C]!*,NJځk>&{4jkk 0CG)8JH:3*1ntiXe*Ss=.LGs8ֈYW'e}m 3@Yf_?T)+\"XW*Va`yTEJ)0hӑH6 A@t7vGA@7{7a [T&"Y|BT!J "ۈJ$DVf!$8J̅3+$rHDg:vv&1Ld:Oͽ{;4,miJv:q 74]ɟ!He2F*TR\d:mfk`J/<K󶲊lhwg{;Tu> 7m}2خd '$v% ]HEѰ,+*Xk˘vrƇ׿ GiԐ"vMa Ȇ 96h :@6&D f&1$a6hZ8d| qJYv@p9!@8>*:GXW\? )m$m/{$_d%-(!5 IZ8!k'"i.1f(ELLG4BNJ=Wۙ.<O=lrW6vF0Jg!zA&-). JԱƳR8 dl\8>I 4q1%#X:zV՟@5 /t/1wd){ Ltd)S2kp*yA'du!X`г(_gCw9K1ѓrJcֆ5 P8c@f%L6cIƪ8F;!7ݱv z8+k;a%ąKUw.Pp)5([Q&DZT+>rأ-)D ZJq++O"f%P[RtϷmI=ڒ"-)kX $3BNJ Lb8dc&=VBm2, s<5WEfDZ' Dc,' `G;0PFccڙ8Vx/%{Z52Ar,@P,Y&ב}D6BJk<ҌP g)Ƞ)єA_ 7&h~=h5M#\z e2b=\D6+#$"bĂf&"MisAD@&01` L>ĭ+,@z5T硫1I,52{ 0tJl! ƝALAqc=ұ.,鵑a[w3hp>;rgSqPG[o t|ZmDàU|J[g?{wUQ/e<=޹/fۈ+3~߼|q;i[iYn)JZG^Ԓ&IrQH^Y=!g}m6ĭ'clrP&韑p1mדμ}%4:%/_˓E5:v4Z;~릩>i8j.ySZ`GGM&+:g~8=5K7=r'e==ƃs;lh4h6^Z*.OG-so鵕N "0+mc|SHMH)W!\A "WE\WEZW$cjzWE`6jcH{WEJ3WMbW$0aWE\W$3p+RhApUj/Ƹڋwq{")%\=RWjfN=8\]K`,0pu=q~XWדծWj~z(Ap?l7$IJA؇L:.C@,0^4TsDc~͋CS3l$@<*-"<3"8(Mk+FkF ҕګLݪݰoI+߮V|ulcb|!pux-N_=OZs\'>UM~kٜ9:=w{eC/ ~ϫݳfש&WJ1fy4J˳GE7V/4[ua{nCؼq 1 Er%g?Wx u߿&'6ƾ `MmƤm~#C+ ^n`עC 28 OP7ؤN+J6NWҐ+ǧva5tpi57NWRՓ+J{"bq=ϼnk+ *:]1J-)ҕO͚]1`Z]1\BW6C+F`5 خ k+Fӡ ^)GuwXKsJ(72f1Lt?ޥ?s꿽CӾu|KB!I:s K7C(E}HG^oG3.Ogw7 &^ԅW~kKq utsͧ>sus+lcopd?~L鹭'/O6Cv :"Ek6Hݶ0nɪs;gT0/@$(?1ft<<!<,jb7{揙)0@0g797Kٸ8bkmˤ)ڗUbP޳qw~|9}}]}t]o?goۋ]m9%\YW%(PUR&kZֻ 鬐k4%n"ysctJHtVT.u{V妬W \J3Y3,^pr|_x\gk2$n쇅|smͅ>TC@(ާ$ZߞC&[ECY.3:${j%{(*#i6hFKah;ΣdspYhS\o߾?1j+0HikfPdM\]J0'Z[C,} BɌaxhT}BR?vUWUiԒ?{/DF0{beo3M- m*F)]hm\&"LCt(.e`0 3.Mc,c16F 2fSU~b1*BBqx/PR=woq&bTh^u*L%T}?a:Bާ,ͳzΛk#iH%5oރF%k _u}2:%&MIkNֹ[>0Q1c##$63f Ȩ*ϡM(-ɄDK]\[b>iyZmtuE:]^t@0#T],~ xm,mT19cR=ڂom-+5~IzPkFQLzT;*t>(XG%=߳jeAц-ި \K}4 ŢW:ȚPi`EHN֔70hن҂N3 _B\ ?RčtXzGIzӠlCi#8 8%/C)FЭxx@A4TƐ_4fs˥ZpUnw4& XTdǢhBM)EH 8yBeN V3f !KkBE9OXD;ע`|JL0[(Kx!'c{l7 h>AirSRN 4Q}QWD;q1a yIO*KwkX9:ʟy{:zW)iNֳ-իY3ZsN Q@O Ȥ(N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'u159X@z@j@@RF@O D*N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'uQɯ D|z=N j5N @@8zGH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 u ƚ@Wܰ'?QFq=E'P'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N q='ЧKgNx)z\\_P/{;Xw'0$k2.u1.ӫ1.%go\ ƥ'a\h#*^R++jj7mJNWTqRG^tuO_օCWCIFWt"V9Z]1WCW 7Mtt+!ފbBWC+FSb:S??:qj7웳|] n ߾?iQGv86lOxwϮ]>iG}{so&|~>֒?iLT(~kcY[:Tkٜ9:=w?h3qӆ=Wn:wz< Yrnخ7@2 jS o,ߕg~MC5E*΍(MQћ]"z0nWK;ZO]1T$D3r&m~;dvьnWg khϋtvkt.z~YNEɾPvg?Wx x~.:ϣ6nU>њy2dWs^Mt7 FδǍ#I1; >{ÌEnΰ-[67%ulV*10Q|Y#'ƓD8^򒑮XJ>(2\tEڧ+4YWcԕ3 <." RUO鳮F+څPrX;6"\hm`ɬʁ'HOWe+lDkQڬ1ʃ啻B`)4]I>KWDұuS*R3>AµlVCK>N>/|tez;~"C๫8\3h]QFf]=A HWg+UֈuSJJe+JjEWDkjM_z%*vכ)Q]ˢw!:9N% l]ԕ -J–]c.gk\KL3r1P̃K/6 z6 >)}j$a^HWp2j2pYW'ѕQ)NIvFW$eg>_WH̺R 9銀EW VB"JscԕSƜtmsuV6"\\tEҦ+;*j<)tEZUhm҉*`']!|?.IJ^WD>]M 8Ɍ>4haq{+ЕЕ˺zjՃV;F6"\hL]WDi j8+ ;=! RQu5F]y)xYTa@9t(\{פ~;peP_ʠďٽs9'+Q9'^-~P`QI9uB!7[i8 XZ6#9|8ڡg4.F8:8 =Yp%p*H]W=Ϻ/~"Qu DW l+M_WDڤXlzEWDD)st5F]9s]. *2㤲N+o+`p% ѾNQ9͔tOI'gɺzU Odv'8q]᪁sWqC'G[#t峮Z`exp> =uEe]PWj#]!V(+ :u]!5>j`6SJ8)ઓ򪓴i:qR$̴;XZT^nec7l=|ϊ֋㆞2aqHW]n\tNuEYW#ԕq/+QuX^v,]R\tE:$2Q(ueB2+>?fD&Z|tE6GWcԕsF8N3Uh_zuEVg]PWStE>EWHke?RYW#Up^{N ]d+M?"JsWU yDUzd8ڡ&2DR&6!D*d]=C`+I."Z/SQu5B]-Fq9$\77'3Mwyz:Ǔ/*Y|p ˓/%Np&MYK1T3F44uhpMhO}DCG4'ј]ԓ;OJ3)T(@\wxi?nibiĕMhAi&kzZ銀d+5֪uEmruu]O#]ltE hm`tyuee銀bϑtE,u5B]+|䮈\"ʽ#ƣ+HWl ]5\tE.y]wYW#U(HWlz."Z|(} aUb*="+h"qTWqR ]ERBZ@kRR4#]!0H>"܁*u]YWcԕTryrM?OS<HNKe+6z]!5ɑAΓQWGA0k/u\tE+4"Ǩ++1f!sltECOI)uu`+lrWDk ˃1+ဓXIa]u(S[Auu]%M';FWk)D GWDiYWѕlk]':a}19e~9(ɜj` s cPGƕрAiq9+fs6fUފ\ q D =PA05B E ݍvq~3ؼ{a.*7Ɵ7}tm']Oݣr[|W}`Pg#WUm(t+eUF+[~OK{w+݆?HAURt_kuk JvÖ9?7bZ..uIutK&b olP T`VgxljGvٿ~sͻgh V.Io’]f߷ɻr1%]n'}je=eg7߃Ii3/_9VbRv>gmY-MQbc1 X#$gJ][HUKQUYBt7{DIٴ׋7]ݛNz4t`iC@r/S>Y)g;DvJc ]m mU]gwqIG*sNwd)kʶ1[`;YM-E 1~'Y'T' #-@C 4,ң/5>½/oOn"~m??M}=Q8»r}wzڮ붬Jz`Y^ |U:5 >s3L裄^=iBh$Tw#1Y]0—غqRu.C*]ڿJ/Oc/[XM%X@WzAiB+^_k _WT]Q¢qk*rx( ]|`@|3EeMANo]⧣[/N8vCNl}yOĆBL)V 1W/?ϾuPPՅnA+(tWYio- 6}O9ؼ|}މOj_o7UZ7 NTM]kB0P-F8NC֡)0JЊ-oG jZ:m:=iN5.(-1R˶[ _GsuUzK)?-yu{]bu޸ #3ޗrɽVucݱ?K@m$a Bm+jv^5um[9Ql1fŐ04~\o[0BTP5mFmДVC&T Z*pS?E 4hT{ 49*p-hjZJ e}~Ox8ay.>;?M{#;#S7iq+wӕσ3y+ _Fӕ˯ә~>MO_=PN|2j%oᗓ)Qo-,ێ~?7[r'ك=Wr8~˅A^Ǣ 4cq{4O)?gh8Y,M H>XA&vhأlrΆбp3b6,a'~U5ȵ^Z$~y_=o {3A9%FlM?hZ~xSoS,oPb.AY@NE7+A"DQ.t(-fNKOsP^Fo 9ert Ѥ[څ.{6ZmC*-:yR1 T YI2mMlh.ehVg=zr>L+ e!VM'+&Z0g\C G«*)}*Z*ʣfY&`Vp_seS.xW>o|`tl -P&X!XdjVjDg x%a%tc O5cLڗ9hR hc a㜤@H i~J[\( ݥFF42ne\)GtJF;T,5f|2)ިT¶B3" D994Y`2 8ʐ<' *l:mMsw% l_EkBԆLB-(Z *;oGdRBQZAMmўe/B;i #`4wgyܢ1Im()z֚8ԳFlsp@-[v*&fR[в߸Hd+Vgq $S96 :sy&Lmoe 5#4D* Ę} 3 Y: :7IA֠V;?󓾖f/X&H ܅eTd樭%Y G+cRZVj|iڀ4> 3ZB4:r[)ik yJ:mk/"B.ѓ]QֆUzXYQ#wޔMQ` 8 4fu(EX:G;mjjY;=5|QiU/'qHHM)D)drӴB@3P5p#kOVD=VJs))bRRD@b9-Td@>4Rs!ƌR VB4M, #`UYvgiDyIA8ˉeG;0 Zd/"ZƩs5N] ^<]0ӹ%hnlڝAzvYXn~0?r,jԜk'df|Ջ/'荪Lf%^E:x)h6otk1)ɗ:>ApP>KKr7Zh?nYEji#hn"TW!%%8O,&ōH1"dgqylƒ :)o+;Go߂Nm%-;+W<\YCk|7X(o\$_pFX [rp:uР*4GvpH岲,4^B([$FcޱrN@+ H,2F#s "sK T7,RJ{.y4epDi4>48rح⢭c 97tDd8jtcvV4ӑ>l%<T\),D0h;moX_z5{`+ewep Nd,8K&Fԭq}۷Cs5W~@c^ m:"k{L=Y/Z3  Df ad@yA\Z uvʩR֡an-5v[}{⑸, c `f!jx0d&" x 26*fmH q'A̅ƒ@o%rJl޵&lMkn1xE VQQoEm; ,d =786!He2Fd*+B',xeY6 o U}a>.[.C,Q|+q\VWf(-yҶCV4--i-NW (qrl(\85` ˥C޿ kXH_$ I"cv?0Bp]\Z-ljdp]*D0x::v[!ߚAB!bja %la|4x[iֺhZ~mc$iH=Ҡm>Fì29Q`*]|4_Z&&Ymu>ɶQ[皪5-FRx?e$ ́}~]zZb6N1.FM;erCxްy\%w.^x2<~^]<y~8|+0 .e/$bn ٛ/Zw547*w9قo3.r-l-Z(}[~~>/zUE+-7x4s圷0nw,E| &}|O+p$`M=zn{g)PKZ55Ϊ@s!iTcSpx,Cr'Z $It`2 %^ L ? 9:؝sƶ_q1Ekwڪc&7yhSBZJ\sHfWB4f hn:"☁ !fE;$$kHXXS Mgs>yE#vՈcFq#Bs<`4:j!8x@\hEϙcN5V!07-QHLd$&=c=%K.Lp| XTX17unS\~|GVn/_żhR $mqkpOo |L귟pT|vLM 24R7a0_Dp}<y|K2n[n6QGo7*Xf EM@(K68յ9|g3N£gIA $3UuhU"E͟La(S8u$OvEW@ۛ KM RƣTefj_;z-=Ja2%OuYڞDpD"QZKH)L<9cs~8 |4m:@J]}")eɯ!r7CCOMZ.m)>]nz]Zʄ2a OP5;LjnDh:vF.nM9hw0ueA#.q"P,\Fje>Q6.-˔ԥQoUmOV.in϶6>*VB툦 , V`Sdz~ЄN 9VߚBGj\Uxk,yo :4} ~;P- O(hni>i"W>i4[Mꌝ;1c'fČ3vbN؉;1c'f~g8c'f2c'fČ3vbN؉;1c'f|8c'fČuaN؉;1c'fĬ3vbN؉;1kČ3vbNz1c'fČ3vbN؉O'Q*2YN{`d,e0 F2Y#`dI0)S}p-p]H0Gʔ^Q**|Q>{VӃ˷VS>Z\ph!cǜU&@ u++PLR i!RFQFF$\JlB Oc*5(K)µ͂Uw-}6=KU3 a+QhatB!]zHxXxly-k=)l:.9-,JXV{ D`MJLʀPƉ10&{'6`+Wx>(]ZA7eaU)dNF03,*͸Jb,A"ꃠviC X'o>jue\ic6qș~I ,G|*PB |~;04aN~:/.Û|Y'. c I paoxP}NOŖJ+pR@C"r|IA)ڱ0j pp= WE-6`MHtH䦢} r5[]ịAa8iP42> }S'uqej%R_۳oj-tzz2\~T,O{nV0[S#X($^Lm>LaLI㋥ ˷ڜfy uɫr٪av1 KϽa]6kKrmW#p4|3Q3`Cm$I[GbmÐa4jfYAh}j+V1azrޠ749ZG%hI:W,FRP=e$ ́}~]zZb6N1.FM;erCxްy\%w.^x2<~^]<y~8|+0 668ٙ_w#@@<{C뮆CSŶZ.g=[me]Ne[TbY1Ŗ[ o7χErH-]<^2Ͽd^UY5U?ܠJ =7 Qm:z>^"-e:4ubnM|׉>,`iD}ɺr8Py `Ee6H_;0Wye-=:ĀÁcr )p .r+<8BTE!-쌗6 lQJxEZ`BB>L ա'xc|; rsWfʭ;7&x0r`q/̻!,q-!4v+ջ0)?{M:IFGTRJmɌ6>im"b>we)m ukCz*B|1F: lp6H0 vVk/;"[,P2V/N:n3dl L]V"2!ui:ކ"Q@ŧ0,&.m`G6@b۰ȵ^}n~> [,D_2f/˼~;_0b5Ÿc]KfK,- {p1컫9M5lN=KYZmZjnuTx6/ֱk\b{aXc%2"IEu"To@ڌ]`'O埫Tc72uw#YO#&]'9gu8ه 2c䒔~^$DÒ8Ӭ.Y_r|><diH4U$@ ӞPeS]=\y'*u;"D7Ƿٲ̒[ l=5#BPБtP?5~"|i;Emi"}݌<>&PQrd{f)jɾ帏."?;h udk'S+t lʾ>b/v>5ys *A-HxPШ6^γ~и{]S*>P GGWћot>tۼ8‘MݴWf^ y,&zqK#7Pu|͑{fon`=g0qN>zC|k6 6iqL:> a'2\+j DkL>J@rD\/[s*ou4kdEfA1PV涭G@@b{*-e$/0*yj&2,8p4 OQs24ŕ87e AfIB81 L$,,FksN"pl l?䯲zWMHaGD:B>RJor(([9&EktHz n ـ11<ͅ+)}ޯ\v>Huة k<p'͡a+'#B/s\c;5̊S|^>\~KLݗ_!,*O'],xw݌fҧ~{㕕{y=L~Л‹gsuə>˷PT Fz͓[ż]9_;tUX?'K/'Sl8O,wS[|`Cwm4i~ɮ7u5y//^ϧQuG\.=hr/W63c5]V޷bs4#m#uƃd O&XՅ o;?L[*cЖRX/㥿vY&֎I'bdNW^syy^Kو[y)!Lgz"kzz䗫G.=ޢ5JhpyF6t<RoWd 2f[_A!îŬX1峯Fݮ`Cm 17)uyu,Ϯ]c^|.Bm)_o5/sMU&yM:GŴ7,!=03Åom]]8{wl{1{&ildHYRҷ(q0 DŽw6c|6%g3g?mLqStD^]Hk4'͊g߾eKIH csVg 4"Bx,0 ʷޠ"NaE/q + nmA"M6xT<*%J>yQ skR)c+}:Icī-N^:\m9WӶڬږRSXo(J٨u2AxVvJC"deJBZD KRt%N8魰*:$:I,5FY#qo 1m"Ӊą!蠵^trȹ-OYY,ڲ QrN4A jZ|+ af@EcD,`PFș=!`2&H43$9EJ&Jmx2#GEh(cm_~*4-u:7e rhIP6gZC9!jˋnjgbԣ@1ɝY4/xR|dmC=yGao.E3 FF0jcwu>Ż+Fpw7*͒39NrlwLoQ|V+rg_jՓoә ԕ@֏(]Lsql{mة7]ꏛfJAW/f#h5*=H:RQӲOׄ5v#ܘ;cʢ~%([Ɋi+U:V?āikޥz|ѝ | D2&NI2xT;f~IeRDP_e)tHi|7[xHO (1+qcFJRӆ1VBiw<}YbްAU@TNΊr< IP*4-i, S&0MQ$˽Uʸ<70^\F d QV; Tx㩎F&!] "8O`eюAT #g\>$(wCag5d`!M;n3-_OWܣi,55 ūW\UP=Z]>1 (+M ZįJҬ8FiiiiA<2+*BGB3&p8m6H>:` h68s| g gq!Fi >X s\2*(#"ZtJ%[-¼3xkR_7- n¯ةE߶7h"`TB;8WVDT"*ZD)k y`kF[,&jwĹh\+MRR+'I)Rg-ךROA&"I4qUnA3&HD'!2++V "{#1KtC¦W!*{Q(ܨ[t@*ipTP5 gph<SJ)"xbry`NEݢ$oJn޾vcýآwG_ݖDQFGυV()5; D锨ăCx k!m\yB-]Y ߙฬ3?NN=;tVEe^{G?}6?kr2]ە(w(Q~[ŌC:Ig$s(;֐W+˔ ̷Rhtf˶-{ m򁩗J_jcrC}ZiIS*6VZVEY%#$\'bÎX"VwZ=EDKy7wiXP 6(YJż<纩"ALjEĩ &eF2 ,Z#ՁHOnblYe䬯K8=F}3eu5_K"h!e#G'BWYBeVKzPm]u͚|-} 5(ʺt,:;נlٮhAB P(l$ Ǹ 4B"Wж^!P!4I2$LFrBrǨH!Fes$AsF4eK*J胴BLNy~bS,8𿷥K"fe!UE~&|MR%;=:7ʫex*: ó"M DQ;^+祢 :MFx첰0iu(족$Yg9R 1;q2KYNEN Y1r䳚>?6kz nبxP IJߑД4EY@YI(Ij J9:02QiiT%YBN?2qȉLMX0'.uAU`.xLP^k'vAZN 2)bG<:F%bp@d:IhGqP-gCp|3FM+ A2Jʐ%|D&A 4CD} K!yDQ#ɿ2Зf~?{9A`B?%)Rñnq}g$EICƀmieƍ "{WHe 0j-Δ@uTT 7ٵ*e#eZ!`"y̅2 :Rl:K4N` "PntMʪI49wUb2sHMD"n/'|N.G,99?<>\%{=\p}AnR/0A S#dy<ΊΪ8sտ+aj|P$ oO? Un6Mo `G GejvɋGnicUuu͛쟳f5{[At4(^]Q\Qe" zYc:'bcqL~j.|yT4gߎ[ ,?vPUV}gWJe;vٕՆA[eW î2T *CyǮv%zqL %O[Ů̮~f~Z<^PR]{+ѱn=5qz@*l ®ZN %zG|맓^?;'նc7~i~ ߳tq !|p9# f C5(=g3ba( SN(NMuy]*3}aKiP ŗlPyuY"U(g{JRV(?zBj0:?\+d8K%O|ʦ0iz~cA%?%+?ů 'uӈS#j%e%8Sb%Hk",~PB: Rn*'zŦ3Z Fh%{LvjKdWrM î\Cƈ56<]e( dWRYajOxUq$uqF㌖(z\ZdD}}|r[?+O΀1婱eHUyBsj 6 h 9ƞ3e[MCx܄8z?orݺ 9͈G!N!/5u,;=7߸͓0&DAUPQXh s%8IցD 0IB< F`ѡDPdPdWQFR(-CtդTܪ%N:uyn%o1KY*]t+4DWmq{.:˙feJ,L*,J5%Q Ғ${AհE΄YW 7"V ]@ߦ>qπ{0Iiw2>,Sזr3*K'B΅Die:OQ1SI X^D10f" "x! A"qi*drBƴag%}s.7>Ԏٖ=):Jm+ZAP6rD%EiRS.tq(El!%6 O^CߖͻFp* w()/u :V+$sp5[/5g`PS`?N>uq*&R췦 L.N4y2鏏VpYMcev>>s-Ĩ/A/n̩i]ͳi LT?oV/N. i,} Rw~Q-]2餲:F/n!F#}W$. camOy)>g =^ٻM6MΊ2y$׍vT|q6ÌFQv4C1LEf'??T *ĉ_zuzWBvt7o}}~y{wgۯBΧ Ϗ@=φڶ547Њ9fW95~F>%ǭ^qaA7sϪ%V7eqjFWf׃X'6<&4]TPTEX 񜏬*HgCb;בNu&Ց@M SƄ*d0%-EHDɉXz;驍 3WVhpxkzT&xB^ ܼ`$!\[HJ F:WNazŌQp!U1tZdz/[É}+V|N8)Vun;#15z82LJ\}P Eͽ9`3tsef&Rr})Uu1X@B.f(_n5g3Yi3unсjZ>l;yCE9|W!+Nmv~Ǩ0_1*IayP^hdbo^Pcp7N,NH]B C1OiT˗ӂ` [r)H7 -5j (aꈠJ{@q tK>+[L%䔐0 秊`Zx^vqeؘG'QPٰ1=Otop'#?^\3Kadx&E8EdLWrb nRۙ"$`K+o_/t><͠f9H( %\ųQ%://eٜʵ6e9 EưZavEwH6/!U9 &eƠW"ugHو9x^oߢ9ެo|=e+wE5)1,kˮ6r#h!nuY^l: z7>Pb[Wϭ˺q9;:#r7f!,e`nZz^4gZnƇpNZ9oȐZ;:enAvf֬7Qs6YmVwmʛm2'Kp2CFnIeYڞ҆Ri,m.\𮹫ɭ KZ][ęNLGZ MiQ]/'\n»]Ylg{Q^Ҿ%"Khj90j$#6ph(4LR&isGjYbgtÃ;>>6@|>j%6N]B7K7\mlضquIeAR'h H$K9CDX,KQ\8V%y>LGfLU ĵLs7:{Z5wi4:EJ+:4K);pVh\2=r|~-( e>^:EipRu tK?GPQflx([.W$Q 8=ri\;ڱ2x #ٴ!=pdln`(E)!`<00kʵq!E 8OQ5,I B Hd24@DAkMLhP[nBtXP,m}n OԽ,}Ӄ9RsȸkAOi+mQ#ed攈)H&bE@'+"dƸ/5x(G-sDqĎ#n,ca~ Dee&HT*|.[PBP@@X޵m$eOiuu5qn5Œdc~՜FI#ʢʰ4C6"Y_UWW2jYR% X{i$Ft1u]3rֈzq\ӧXgg\^T[Ġ5:lH@( z֖$!DYߋ;Q>4i3ŻЋ;vՇб>ܱR?iojݶ$!5ja#l7۴A&oBc,=o8sc@FCoћ܈$D˟7lPJ3  KQY@/V5!O>&_iӡlIrA,"%HⲆd6(ҊBV@Lti-*D(N9vlt(\.;L:3rn* C~ K "UJ8eaԦ29ehK@vE 뜬i.]նe`BUY/zR|Lo {Ch7Bs7d+Cg^_9'2l`X!/h53d}gӻk4lWgLQ~:~?62+0x4uz/_|Ǜs O-p۰/߿z iyǣ_}Å'xj:9x>YWGfOXU>L?8o9txJНstt_<߿LS:+я?xGbX;k\{-tr`fe9dHCv_O' H'kr2=Z`]0}e2-nw)8i:ZUO7Rf,So k&nD0;Atm5H|0Q6%ŊzszJ!FZmqC^<-W!h~|F׌/|<쥬x Zdݣమg' /q.eWNX?텍LC]g]H}~7YĮ8[!yRhwA!o&Wtإz6bMN2wadr)ii:#fRVyw[ܜa}Sr oh(/ēUjY*{VofA)42>`u>d"&*RM4Q`Â:PՁ$ll[BY*&SJRrUŒ|*auח }3E L y[ JerAp6`#e֖n]s^S 8ZB9X߽ݧ5ӊB--Ձgf ~ZCK3{«|RY oYɘESomhҬ!ХI d}S[ed@) Cٖ<מX:8ܢNtUuF%eRu²Cu@*3LFvT`ޡ*y#^i~YEV^;:MO'jHZ<G7s3QJxgJiojVh|GuG/yIGqA&Ϸ"s_NX>޿'SR|_=vz[jhVT*N:2+SVV2l嬵SgYݧ_Zh=^`#Т"Pa߈WdSLBCϢc綍}WDS?ecQuB۝|S{ָ'_yժ/jw4|p6W|//ROѨ{W{OF̍ϓU|Rzy:Z~_= ~mY>3 :5Ƚ)uBF6FjvlMPX]MܝMDKS,)Z@6:#eFCcL̈́ %' &2R@hݦN!!EwOYƳAWogTe`R>~+P%_J%eIƭ+3/7YƜfZAt<OZ o }8h΀ޕ EaBLd<y ZhmyP(^'v,|W#i C"J]LIƄ7,[H*AxY͆@A2?{WW$Vש,a"LO2c[VM!Y]׏GrKD|3Ũ:PJ%3Lgl$kmlCSɥ9ڋv7ѷd3A:LӁmp_D~\ޜs~jQهwsI#Z7 !o&GMKQa=;82n>cT%3I*vWTuʫ;_kN&$҆F+Wx72wROR; Z 櫫O~Zl|s!6 :^1ჵxU x.[WSx_h1PK6(0g>V3:?!D,xզ7KfKO>I[MyQݛ$̽Iw玟=7Iƴ& ʔޤxoR6.}틽V+Ξ.ЕfJP&+®矌akVʇ&t0>/]= mz[)2+]=뭉[cC+A;ӕ$t@ FWF] {OW@{KW}bwFvdATsQǣߎ߽'}Ɏv\5,뾷| ]mM*sև_o6ʻr >)|G`V>Fʇdw٧twq ?  U`h˶ |{zvu~CGOsTPLAWqGV}/l1,@EDr q~D, .b=n@#MGrv\|xWM4{;f Fn)YcQdB` `ՖN |>]zy;aYgJDW( ./ۻ lb kX[ aΞAEuq|c0Z Ç&]cYq$vAButCrKSH˱ 19q+YWeQ\hZhL=Y)8/* ɮ G$uc6 v1Ĩ6RPѮ!6αOd^F^ Z[);N-`M&wo" d, CC0!$l^;Z _W;dWl1vB!ՠPwjG`(c)(b Dah!$\|FdBEsy f Ly>6]EZ @PS(H󼭮BE*ѤEb2%jɼNwEࠄnMk|'XkR 3j M+,3f2擊Mid- !pB\. v ñ837Ozèȏg][R63eF;(#K<`"qJ_7!J3IJ3Bk] $Et<) 䄸 #/Z,J芸P^ 4I6iy](C=h d "h]%0A%+ԭoH w"3h8*YȩՏ@}ޒqf@5[#e v:"}9g>B^Z@mDebuj W5MtpOk ) a:2πEx8gW\8KAEKշbhǎ@H[^FwHԸ^vyU 9jAkĨyBr΄@HD&e#{ q#s>.EqUIb{tpiUR"찈Jfcr1Z7N ketՖEs4DT52kETJӴ,Ko ^ͨ"`!-3!ZHn|Q}Aώo[YRɹn.Qfa!>?_NO`SܐD>lɹOKa#HNj Q '*UV(bQRF3 ^z1 U @HS]~f$_Qc@sAor0/hZKf='23{J @ N uH@R': N uH@R': N uH@R': N uH@R': N uH@R': N uH@zNd $|8N M`@@;, %:Ϟ N uH@R': N uH@R': N uH@R': N uH@R': N uH@R': N uH@z9NZOnG?|"[M婷 x6F?wN=x Mt@%Y82`KצC1. ZW Jrj\z ƥO>v7n?:/'N''4e8*Ve\g~\ޜ Wׯ_F{9aYUY\}x7xs{;fǬuz~n\?e`C4U 7of́Wc7~?Np#𷫜~=0T0NzK> /?ٶZ7 9l&GMKAg Nv;i'NinzV&~cfnA5?7bj((Q|Z=F>voz 4P67(3Lqd*w^-n~YdJnd_w myٲkM)/e0GgoEwbkʯ1hsD!J%'h{%'(A+9ٮ.C+AiJPFt-YrDW8僡+u`vNK`9C+ ] \{0Jк/WztmLJÙćBWe HWy+g pֺ}-ʬt*;!M `8 \C+A~JPU®?y43>>]= eEWtQ#Ip^p7b=MMFB?-)RC6~3×DR5F">' l*+sL{T~SgW`I \UWj}8\e*:WхqΒ]Lz*0vè_CFQ~&TWBu1W~4?GOJx]H"&M@rI TV9{T) i(onStVf9W RxJ+KQ7V0v?ȒӂFcCH.[{(L[N+,;dGE}őV_f9]3KYQ9.]0rɣӝΙxFZZ&grA2Onz?*AK4z6pe{oTf*=WW/!U&XgW/zAp5{]e*8W| >Tįn;:X\(%ެFՍ_uܰ;ٚ]Ȥ9?++JuXd6@I3k"yTNHg޼BcwsC-v-|Ûo9i$r,s{S9Aha#h{Wfڍ]bJ |bQI˭uwECqFء%aʵ!bX7GEEO7&䋣&tw-zIXtP]1,YhI tn۩:;+IHQv_~K^'8K Ϩ,9'*$#.~MKg`}ˬ{ݳܻd]qApbn*cˮ7)|z$ߍ\&ybC( !f°[N:s4(!IKMN G0]&r YOWwNh{^9Me4]z04O^9^/4/߉+W|5M."] oQSsS˔tCK |?0ri&.ܽ3kϷnz,K- .:&OF7Q6 $4ȈkRn.QcPȸ9uAXa GqɅ-j@"V!ަnXګe ?1_O9@_jiF`@ҁH&N)}ԑI(Yue LLC|vax) f`$ KjPݣBEse7Zm𹋒kR86#s+ξ0lWee.]oNoa\Z) F%whg֢+r'L;8Nxû#~ɏP45PSDTD:3wZI"N:}2堋jq8 P ;$r+c7]гeso(N W`=url\܊x gMY 7u:.Blb2~t>uq*>ۋ鿚*:9/.;|I!Bvƽq)r1r Ά6&NóU&i4~<V.;U7^NͅW bP4lq>wp^-]pAIW~A d]Ouʞ\źnn,Y"CT3g1~z:ѣy݋xm2 v/:0TFz@_*r2b zL<ʪnT *䉿׍u"t/ϑpt^ߝ|2s_%8CEx =FzՏڶ57E׊9f·W95 I~+cfKn%@?N>QW4rYe\> "N_AQ/f1i/7y/\ B ;BGg8^?64XE e4t*zz?b}MFj#1HV ځ&)NcBEq2岛Ѐ"$DD], BK;WV}t84{.yj1U2uLEy^nxi tVXH2ׁ2IR)dP1*.d2=VtF0\>}U{_?Ck!]#Xo(~2-j' {?/?^wGn|\VfیQV,~=hZOiYX"%MITP*KG[PI-Cr:[Kj 6ԩ fzگo./A;ҲxƄ( * B`!y':0IB@ 1P r@Y"1Z(ODrjҞqB$N [2R HK$ qaU<(< E̸2 ٕ7XV rBHKw}קŻI%\]\t1Qun2vF0]9lũ]ա|ura"hL4F(^ #C>kR$QDt 9u"b8Tf`E3-u^b8^Ŭ(|ڱVP}J;EE 'q}2BrHVRuJWkPi#ލ2?rIɍh1j6u^\9AQ}3iMKݎk^݀mߵpf*q?^W{kv= ,?U7W)nu fu{.2ܽ%W7=-^Sۚȍ(fAyYG!eme.j~zs8Ξ(LoͪA|'ˆ[ovjyݍ7gvhy0 679a>mxy`9ׯ4^uJ/ʶ1cnHY|\~f3bhQ{Xkp :dۢi{"OĘ=mkRH޸ \|Ԃ'DtbLGfL U ĵ,sԳr`޳r\mSԦz`Y>eɬc8\ ׫\uQ9~-8e>^:EiB8 z!0d045eE*p)"Al đ9%K!7@馍`Ag/|tˇRd>"KRNsI( 㽦\\(ihiUAA)rBPbL%XG!ьKn&&,s‡Z Û pJ5>V) Hi/ -|GAВ"sD F<)\ʩu*:)6#`cp01qz `zw4) !*T FȹUZ8々_. 'Y3&ehz]Փ_?P^o-)ZJ9F\(ZeWFx\ R*-2Uh l)Ȱɕ2)5Q ̉ ď Mˈ9w#ら/wڦe6>h8Q(9 IjW,h9eJhv: CђT!hE& 3!d8haڵ'9waen Ǯh[FD{@"n ojY Dee& (T*.[EDk(!am R -Iiʣ,eREuNNВfBṡF݈S{L>:[qvqi\\} I(PY>|OԠhƐPABp1pqo~]񐶌;~~kCw~|G.y]. .le0ȵ _?Z\YQގy}wṽhu֚3qOprxk/D}%Qj <;01耴!0@95{+1~if~8#g"d֔#d7^j<e=">:8sLu':w-Dt6G˴ѥT2ghG ԉ9ML6VˬB}A(ԯ΃ߗB?m *)CJ}L2y'XM.eL0-R `3kD\'U\.;]tABjRd^yf'/ԌzW#ۨ:V7ZHB| o 듺x׮r yЭ<[Y ddUJVchlΚ[# Qo] Aga TuGR%CH(IJRd\3DRK@pژyd70yR_ȹ]B]tr`ᝯśq59e4O9TG݉.l7=y(c2gG턔}F}ou7)H YƓd.#h4j[i:eUHrȹ, ?1 pE e8miLGuP i悆 4 *ӌ}֔^EFwPzKC 'CqnGR)ٿ+My# F"ZAsmY-փ;`vl޸&J~γ~e]G krN%&PEY4ZȂf(<~NR@uGPz%aQu9`X]۟2~v߬ ucǎX' 28O޶4cgF|\I801h, ^#܉F(|dJzpA((XN82K^s!)`Aj'瑱X7Y BS$N܁#$"\+IPHKv9רC߼5]s;ѴL!v6EwSDglVff_qIYABhQyyR,CiL2=ƽk%zSGLwW=;UOy6C>s,#6+$`8] _"S ߿\XU6qG@"ˠI#^0# FpL9aDm!TtY꥾wHg֋b^ ]Eԍ@Qmn&pDVK5x/oA'$W#Ϡ3ʐyҤEpouV'\*rneюo~%I~7?~N8OgEg)ʧۣqZJv |c~V |?}8ghBOG+u6MsH,L4`^~Xn6R]}lܳb6}OI|_5]B׃g6X!a??,?W ̲V/Q#&e[Тq)gvLN|4o+Շ1duJBJPy,ƄF7=&;dqhG`fa (Z-UvF&8eͣxl-Ǟe/ \;V5,Ɏ[41cYYld#,gȹrW f=ZKE3M-Ʋ߀'ҕ,3ȜD*>g&N`Qgd"PS>~-TQga9q qC?1fəOC3mxA xHIzU jz?i-NI#q͞LBrBQ9K&<@P^qTQJE_c'Sn!ӟ)"47-BJ&${mLU Y/m""DL=Uc uz`)F)&R=MXR鴶3CFg(xX''gZrtҡ GG-_ވ4ү_{ $J*!O6ydL}B@3 WR6h<ҷ4fɃRRTSRDRRxRRx@d9Ud % 63.(Z ~7IșUqjd%/EfAG ' c) D;0+*_0TƩȹj LFxŸF\Jy}֐+kWry U'[dzB_,vu1P+c#IHC&ED'Ko uB!qcIjs@[Kҕ5<4,M>GnLdg]" ;gԕ-*182%cNlF왳i Ԭ. 'dc43pe"h%h`$uHƦKՒ=}ӛ̮ B[%'ԇ(H)sp1,b %&h|=x )i*3`Y{Qۖ.xEl :W?{uS dzv;s_L~޿,uOx;rRj};*F祔uL3A FR᭜9hٓE$whӁbV7ҹmz1{b mןUY}:s2ínXj:Xj]n,n~8W/V6$ 1Ya6_,OOFKFA y#K3J6Z:.lYv0v[?|-rwq+Kv8u׻W7ZI?,"ןJ"PJ:d!3$u.ifHQlur*)ɂM;\l"̶]ۇ`iItZpP AOp<Ӹ?O&t< .^Msted/xGH}[1K|vJ;q>yM/w@'V#ʪ[pg¤m=|?F:HVRn,ӶESV ֖TmF^A+ r7_'AI&a0 c)hG2(4 !-;Ϋ. I'NK Ky2k|awm|wH_IVe`p$M>meɑbK%ےeeIvOEUU?V1+_yCRpYkQz+PI㱤@ _1h'b3d OU^D#.Z5plYCq'3K WYe1 *)-z! H;s5 a?7,-'\ݰ4K)AC YwZc*d_[9ƯW;׋nv'Hkӎh ziղ1TM*#=)4:o@K92x8J-MTl>:VKz'Kx}c8ewI֤2D'R) t\bh2$Xk4؞oyʾyrUlݵw}}RkK,r?y oUh+?,B^ \?Um$XG 6Mhya6\JKQIUJ++ɥvSaWsh-6Ց؈݄+*oQCt>e>(rAg?uH(ZRgchM(,H&2cɑƩ-P CM,ba7ac䬙b8.TNt#/eәmXϼq{/P׋ QdA D`c 3N58L)U_q50x;k|x5.60I׉8P}ISovݏjq.(M>x+% 63wFkmX8Tߨ7X͒Fq[-]&y *(*k*V`ne+6 |Ϊʺr QoX+;B0NbTTۭygȃlZt:R gJe~3'FZ13b| p+OoހV(yD\!%G+ %Wޡ%U;"*>*˨UW{+)}U?pUp4p%%0& >5 w WlC,S%#:n8*lI7k?<狏X"G>r~\ QYe WD^ }݇m^ ~,/+}Ye='n]PɅ_~n5R;bf7ЙK#TgPpqEpx|v-W$ْҡџ1ȷxdkۯ`p}7 }a}MlyoMKsv&2@SQ{=tG r\)a̚%% E֖hQ' ,l&0 kl3DR|X,P0M"{#֫oE~Ewr>=X r uٖgib?>'q@ey̥HnzV)!m}5mU7U{n>u|[[n*tKخmZ@YͶM_J_J״nzttOg]6M幱UVsXqeisNțЇ]M*F҅RmCZ/& 1!"WAE.dn! w6c%t eh}(KI, 3pR༷ub:sȑLD#5Oi/T#gMr: HNk+oh'ZCڳw IQ&Ob`(gǢ4W]ί7bIZh{\wi)~Q}Lp3aAcESiٜNK$Zq}P\]L!;3ϕA*Ȉ#F!MGBOmέx=(}L%t5wg?Xxj4"e.g0S#ԋV>%CEOpɩMLYZ[T^$5:zO5"*'ِyuH-aC*w3^`.3jhe%' o|hUamiÇ"6v~vuUzˑ`4%1<=v4NWW1S,gd9/齵\Ih@;.U׮5.^~Xl{=]yqM&'ǵB*y^Գހ=/ |*ibbJ.+ZrO8ʼnAzq3FO݀Ai-JBΠrce TA>C;@\>8}pe1"Fhn$2'*^!E 9;҆{XilC;^ڱ0x dR0 faG6T|2G,E36%idiK)U+L YSLNz-@1ȕkLhXP#yAM%Z#ϥ/kObؚ1`Zɣ :Ia}4Ж^(*?Qcde0dCg^ŜY&b1gE[9NE'z .@QdTFdT \`:fYNǹD FyU8 q㛢/iT;fhv&OOzh[r,4;%F֒IaMrRi,FEb(xOL 6j2DC,h0b7F8gb j7iMڭ]`$[D* LRi|r$ )@+IAw$LDKr$ 4CdEH&&HE#a]p>f4&ipcp} -U mkɄ#$o܄}2&pDR}XPʠcxWݥz𰳚6jziVSkv0 ޸H I@2fIFiF 0g3}ZsR05M76w IHi{-M{#gMM(j+ s %]2f~4ef׫WWn /],ˆ6!9-K-潍@$MjX89'-mgC\ ڊx.Tw3*AY2aU6ZBH1 GX͹&HO dJ<+AKf(bܺ[_"rMW+E/vri$hVJ6YUP,_ZJ)c0VeorkmMkEJaI :bqDz& VbP(.Z2 beVrHF%6$!4.`%@x(ybH:F0A?KU)v;uzn;׮z>~WtR%bD CY;Qik^m}?߫QSgᝦOw&Zgr?v3^a^1{{rߺ_L9G:gZH_˽wPf&bg G=A,iՒoV?dYd[n2?+NVe _ͦ yCh]Km-koF,w> G}J4Q4[*A[m^r]k%VI s>&eYQ̒{8U;|LnT?ng?`<},|>~ux0z`\ʃNN- !=xuu[MC{MTK4-po]weemvϩO2Iv)T J?df~d]9exČ8#B'<`v>m/64&//..88'"O%w#D[xicVW&$t{:a:Ėg08uvvvw^:U]݄0>KQx ڈ3dMnP9xGų,?EOƩi>孧$ G$%hDŽGZ8(`!, muwma9 d CWeudS98Θ>>42]bq4NȘX "NTS)`+ S.Wkc\WW{emwuwN9]HA[w2i:Sf6 6|Kn"'Jќdsk,Qle;›t:x[MCfsa2 s~ |! U(IW) Bj)%X3eJqTQ( .cܑDҍREb-L,D4Qk\ .%ہFxz1ҝt;Oh[;-O%Y= v箠ᇩ3'I&>}pDOxP|Z?NE=x W,BWҫp4\O*F~=QMʧ3kn߼xz-i 7+m(r=) zQJ3HZlp`K=+3o%yЫ0ԃ5 % \HS * ]V8望򰉿{s4/mCmN߫.O?*-^}MI)4|nV13dXTƷ d_d:;e^evt{d4Evk__kH랚i _VQs3M L؃I##39s\;ϷUЎV"`A=F ׁ؛eNkEuYLR((cvV K^~G-9zӔ{딭l0rȌ1, A<<;%ƺ}0_@ou7YDZK QNu9 JQhì:yC`S8T 6Lp{Iz}pD;k"Fk )6t dI, G 3sLIX.`F"ֆuɮ-s615]>PSna~:﾿Hxjz_YlY煼|qcX`؀0έ!#qnu1+tQuF,Bϝa UT@Yl>xN!F!k0!CBZ,YZa2D'z1E4`(᠜ANi1k{5q.A}Y_>qH}$Tv@$m_/6x &mm$RJA)$qa.qdS39K1҆1ci&O3=kMOgy~4ށs89|6Q7$rN;ę by\7nō2wl9 CsO)68C"@ 鄀,Ոk BVD|iN{җzZ6CC u%PԶg׷ ؜-6xR*F^0͸# 0ZpLmCx FKUWƗ>`_Qp[dWt'{;ׇZO%βh<^(zͥWfJ+JtPv4Kg: jOs`b*]E-wJ\8\::gokJf鹞{#7oçwz_{/gU쁩M>W͗[|zW=92'ooM+=;OO^>yf5A8Q;=Iq ͦfwwo4iȶtw%iЩo/̫h`Xɴ]qn_I1#L(Ҕ͔v֔UܤejOiNPy6]A#%?.CvPUțb3$.FOD6ȇhffzYU)jkˇIi W哬*>R3f`ʒ_9(K FK' ?l||B J{Tzh!]:u"rZQFWKJo[5{XlSg+Op8稊-"{1jL 9 @e%UVs - _S7x.taP|yv}ASQ2Ѵ? ?p/ !7Ofſ_0*٬GoQJ,%yP:̩O;yg;렗=M񵺥Q^s%knYYef@~r`d4nV] }ď՝Ѥ?R.~Q}N9AJ"vF*wsDss5ikn G-eGV/+0KL,SZ+N>j2 ezt,ڡذγenM0kzɸ.x")m頺{~ʠ5&XTDeiux`N64ԬF U\y촻SVhcq|vKhHr)y(1d<-, j/6$Qs-9˝R 3^y1?[$L/]9h5愒ۤp=qnmBlJa'B ҄jRz\Tkn2-%RrU|vkT{V=46dQ2>) )K0>l{YviO_}߇/5qIU$K2SظeFSʀ /|@-CS+j'2SiRa11Hu쯼Omc5rFf `hsS<%AsA@ƭo}wo~,\ ` ҁr ^xލw%RSᗗ h#n4!<,`'iQj Q-R+5s|J*mW~{qsGs,gjb-QK/m#IЧcKuwp$E.%Ƣ_eeQGJ:TIE).,˜LtOԋ>XC/ K:m+|0&;go/v/n=|:k =`7<;<J-&e- Jt0tт4T388ht V(b$ISr(m ^dSJNQBik[,>酐WV-W`zrv|Ҿn E[I|;rtv˼h8gbG/)-Qo㓳MdB8 X)d(Ȏ$vADqeTL#XxRz[ڇ,8S jr0Z@t2M" h钇O|˾R^Tlt;u}MX[!f=8+m᭱TNjv-+A8΁ x܂塀Ҩ@mٮxM5[K ,o;v]Km)1{#Yoh#R"1Y+dtK*.S_dnۄK[3#TTĘ)0tU$1 PoML5sv&1|2.89p߶buwwOm=M1_[$h$A-tRy\1Z688Q M!X<7p75Wꬿ9ۤw~> xwe+]ȭ |@V2':T2t8!ON*v+P/ |h\*"F*-dph"Ԗ#Ir)`T1@1mZe  5ũDu~J9E%>jpBEyV ݥIE^_N K"d 9P{' `1^l ,H$)ۢ(k1*I6ABq6Ŏ˳MBuҁ 2لdm„LS u6JٺeHcM QvJ+dE&ࢲvu9/MI /hm$-Y3qִޮYJ-1*; Y oReJi H%ER+1+&Hd`-MK9P]9O9R Aez?I:qE{FbNV3.N?i! @D% E|)Ƣc1VEy&1J'v( xeځv! wbD,yv ɜ-f= Lbv;",儺>'z+LEvA%=ߩIdDZEcN+3$3l` ㊌ui|eglju׸ z9 k q)ƉxHCuO$j(" a6y`cX¦[}s׈('}γ _8=ŗ!o~O,(Nɗ i =Q\ ?$iQn.6(Rr]]YYzQ(YR*GhA>TgbċzcwY|̀cgk&Cz,--u䄗 sOohvܴ8:w..'gFhaz9 Faoco/줦ɇoF/^ZK2.OFsǃlǿ_Mᬖp,˳h/(G3*))U?E5κt;p_ҿիXj!VԿk~9Kվxw釓sӮyLZ^Z4>ק*DGWMxaO^˘: ):D-;(* )%nW?ɡ~×{{cbz8DWtVsU-Dז+d Resu;BGoc 7vbL7v1fIyNWu"&"Sڂ2 H,h5ERǜB/GRwEe(,D[)5Nm&ΚTп&K#Ha @"eC$N,pwҭ\3Uw2A#!F0;'g5{/}^t*IN'i@UZ;yOf)ox=%\dM$^`ΐU_\׳MLoY'Iq}F|IOhXfoFy~t:T| ^҇Oy7z>Qu*f+Ox.f^FXk/#wRS]gzJ2Yq_揣3kyMVo{G'kT]<_^ܸŬl ovUxO6>_~v/XgUo\dU &_N;gW4q|^͊X_ù+UrE+ZKƿ.R^Ѧc6ݾ(@.:frY? IfΣ;Fz~gNz5R`><\[?}3_~-5μ4~}oqz3 Wg_}ċ063[tָR/[cE?-;L8 j({9:#0C*#+.\^Y]d{f/>Mj3ȈF8K֛dIlKٺR\jG'/M䥴HעޤR"X0),sJdu`*I.L-N츺eUIɫc{k_xxO1s%WgL̡ ,lM]{0yՅsR.p jJwǡH%A(TNMṷ6&Z Q6Y2 rRxm|0ChC;w {6mo?)HeϔJS F$]! EaH${dȩTlۑ@0RolLA%"ShD6ktjlYqM9&LKir;a/4ӓ /,nF]VZ4 Tkdl&؎4f1`q9;fh4ߚgpz:|z2xm${`A.h޳_.E$BIXѢuiؚ^P16u)ώ]`-?J"ƈL;򒋹+L;6Eml8`Df>&.jM" H2MsƱ^>hg iQؘB]e=df؋NcM&"2qdBd0Ʌ0NM&'*0 "6ӏM7FD?  ilPQE Sgni`%tQ hcѥ8:tO 7м*7pӆdI 6h1̾ e2uflRC=?zֿLu͔P%ֺiOXu0LAR7woSSxXɳ5v*6 1`NpfoR]#v\'C+2|ʔ E d=)"HA)e8Ax>7e@0! 8jeLY 5GBx75>&S|Gg"JT5e{ִWIhq>;|, 1 NU7^PHj~y?]¨*!ٕwjjgudZǺeUrq|1۷/G_|f/`"8mCCay]뮺V]SvZ.ߺ]e]rK7'u2;v-b JW߿(WҥKG+#\J_ l~^'kFk.iϝJ(r ʵiKMsUت]E7&u$-Y*u!3(PyП "C2Hذyy=cr )p .r+,S!*uآ-쌗6 lQJxEZ`BBəN١&y>}99+nwܹɁXջ<2ŏISm^J;XƏu`og+>Nwz3K^|1F: lp6H;Hys KJ*{o6 s53}[ͭg ZN|%mt? E^7maG ,kJ oomK_OY1}_ 4[ Z'ML`^]IO)yJ.r(@,dsk,QO+L!k3w/}:z߬LRG͂{%B$bDI$(@D0Jl c͸6&Pʍ"VFudXDcܑkZ )$tll8wKC&F3ⴾKȷ.aO?KsG;Е _o=y"#݇EyT-H:$C"NT9"ɣ"g \n%aJu05< 2y!`A=F ׁb 2 zF-(Cb℗BAhA +`I%y>c>Gd]nb>ĔL|Z#oMĜbr0ʕ8g "$] `:pls?0>Pit5zgMHpG4jT)%NbtP8RPEcJ"w -ߪM~V1ۖMڰ{v6NƎ6Ƶ^oVTly!?bI8F6m+f%й\ :xD@V襌!X "-_ ^0|< j!O)D2{mfB3dCuK^ ys(SDc 2IHp"SZA>9@-a㺶u6jœv iIk TxWwonoPc{3Gy*޵O._et+?s-bDJ)0v2e$8n#3%̦œiØ1^t#xũEO%&y;PnG>2Y*y\z( hEVLV;ę ^t"y\7n}/n'ġwGH zSr,Ոk B,"|qA<*h]G!u/PԵf7w* ;K]MM@A^NYX|bܑE Fj@8l &xo(D*,l4i,9|^'Dgt玁% ,1^bbR{x5Rk8 WnwGPjк'w _<~2P:țdto\+Ŵ<~=~ٰq z0}i&& ~/zGc_f̪gL'Gg.,n85`]IqgGh?TA6W:hr&RS6,Q}Vt5^!߂[T;+Ƴ]x|$&,,  \(äލgWр`9`68NWBuXx_Eg^sZy[{_QW'CsrVzOU?{ O^2Tѫp>_m=DHQ96qG4)GS3&ּg`y歒Ϝi@G덖ZuXh*E+Q4J"EPʈ sPʩf6 OU:#C~䤼Ӄ-@wn;7xqLt=~0#p>6ܓr *g^qsuñ5u.Tt )⚒7,F$kc#-tR v$ܔ/tp deہl#R{.R4`C uFJmc肍NGD:Sy֭!cFY;@w,Kja[10#4'(xAcpˇ:M쇱糕\ϲ| B:v 8Wtn0}Y߮p+է!Xxu(~lN +W7sW,o¡S&cl;Ky$-E>ג)E 0{0{ JkYw>Bg1i$%3-O"լ h `R;l GjLq)GsQ b.6#Z!oOegZ{6_imL"bdIeAPO[pqPKV[M5%ԕESS. Ɩ}LO4\B9?'WKny;ojy<0;4`pCmu=W*-X|%gExiq+]Pk*lPT=\:m3VI52'Tޣ=Hٍ!noEST&pQ+ᯨWB='Ylxc~G:(rc94y&v>.o;%4%, -,ꨭN0K*9:x{ e8ZԦذz|k9T쭕=oZP%p}r({Ƌֺ.sd"7X"kR4.pg7:Kk1ISІ*,2PRZUY#= w`8[|ݿCH/D,=݆c?cV}%24qq:MKqԎfiXcAͻAA%ɜ@ 9$*0Aj$A0*e|QرkJZbZ?.CtR+e#4qo tṬP%FWoq/{f)yNu%`~+R7.uTIRN$o$5ʚ0("N'GqzG@*&hռq (B;crI#wIsIa#rIa%e `Y$j2\sEM!dj ]GE#.]Umt@4̫`Cm`$ap{b,Rdgutz>H9Rb`sJ1OLeaժM|-o-ˊq`+}Okl#e<#%$mag a~ztO)i R6>82Ń؄P8&٤P,{DDP\T'ϼa#_LCħH 3:;-KevиP>z\< ui+aldkCxQ -&17Ze[,TmⳟĈ!D4V/Þ^H }žb   ~~xC F(jr ~[xk/bб~7+m& 7YXs[BԲLڲ}cRz-[y\1vnOVl4xX#8R4dL8caRi5NnȼP"G5`bٗT(j^ U ]!\Y ]!ZeNWRJW{HWJv] [N+ayVڝ j2vJW{HWjmAt. AZC8-TU/}+eV BCWאbU֎|<+]= ]ByQA˱]!\Q J=vB>ҕ% >`Y'UB嚛銑횞) vT~IWa:4rd`d{tD +t1tp-+$+DI+]#]1vBWG2Tx5. cX$;o5ݧkۡ~E ֿՠGl@2I\&xfjgQ.Gia߁0;ox =M%8^QRl@"d첡EilCSR]]!V9WR *;vBZWC~-I]`x1tpy16dD+FoCFڐƲ  (f1h;]!JUmHW.I]`CL1tp*-3c+Dm=+C%BBWV4vteLDWXb blW֨Ջ+eCM>8]u,):Z>L0\W#[ tE+]=)B˂ K+L)th;](+]#]1&:z#np-(]G#~WP٘IuZ\I}^߭#B.uh)%QX.ւ@k-&PFf*1gXT]j9?p-,ƀ͙EŬy;5Nޢ4u%+9܉T.-$p[$gDWXc'GBO;]!ʵ5*]])-|Knpޒڡ\һKzg+͕%9M!`a++qBz+DY^ҕFZ]P싮.cbt(tte$B - n:Tƒn(ǖtŶlzvH×n4ځmWR2ꆒ%u+VMO,"jR 26vB>Z[R]`QWR 3vBTG2>tRl˨tRNƘNC9/c JŝT$-E7 ڡtCnG $̬%BBWVTҕ\S]싮.壧+D)*]#]x|R7/DH=+-%y!`c+kDhPлd4D.g1pM1ADkG (4+]]Yi,+G]!\YGUjt(tbo(Fp\3#C' drl%]JWmzAxI'b 2Y ]!ڡIuC)I=+b`u1tp*rt(UUWIWjcFOoijb.u2w[+%ZĚ_Ȳ& +3Tm7s~FKhsZ::7//*gЛAA6$8^F^w/D^]˛=۷ .<῏#W'DI5JA]cO 9|?w~>94J -"w7^?^̟gpC7==MEn:y\?Qī95B(ߜ*=>N_-OV'ot~V:G໿|cw}H,uUGWZos[)4GG$ `TH3*S'mb0[EZ%mSþ>^}Lgx|O;y-u˭*=vG"G#\˷Ɲ4P DIQLnϦ_pu~qr~|wr+ͪ *sH:tyvrk>{nZp~4>P]c+ϷuZK]BS\ NN' ItJS{Ƒa3RvKc5~TK(GR}3%sLLOOwOWկO&ӢAN!b]\^"}jr/M ~dΑQ?|yh1\ XT'[Ypf5ܾEΰxQTh{-?21%o0:_L8ۧ -)J0LD9dm+d=0.glb{XdԇR%1"yŔ(Iw͞cl6\JIq>MM ]C?l2OLcqh[0ٰޯA:K(ie"_G$ `ĵCL%JT`y i!@='- fs3DbKvSt6ujͼ t6xY }R+vjEC s܏EӮ]ztx@aQkK飔'T|Jw}wp+@uq(!4}NQMZ d4##\o1ׁ[yW"ZA0'gS yH*ı@H2E "4$i0K-$$gtҝS) jqеƸ=ͽ2sDnRMsw{xL#»e[&52|fa7x'9Gsa\L  ,JAIxʵQ1<:!S3-NbzZYOy5΃>r$1Y$$  "2X?<"ty2\wayZς61g_$De:İ HJ-!@EDH|,T_{o!Тf? 55{0^ꂉg}(p oJ |bQ!Q(,FǙfP GKR.U۰fgñk8}'8D<8RعP~#EW9vo"F򐪛I9Jz}w)\ݥ .AT2B)ΊE>wa{_CZ׃ݧ_-mf&.}Fn)1;G8|g7rt8_X1 =Xs0-&Yp$aet8u`\n.2l^rAGo_9 y^t_&X!'n8Gs$ڣ0\@)*;b ?!֚N'>sXWN⬦b&_a1xSbm>x}~sl׸,o= a<5C'0W.a,^,>%8`l1(2zMbhPp>7eo?no?0~;s7Yb_#i6]iyy0G!/Lf,F0pPԘR,9qTi4%hpڐ|Z 5ɉҵgmnjKS$!jgùgኛ^`ݚ>dJ.UΚnϷ_%vK֪q09ǴUH$@PʒE^'OS$: uK CypP)jcE](M4qL&dѧe <@/\欜=8e5 TfAqN Q"j4F!T)h8Ĉ0*keO, Pr?ݴqo eL>:TY*$ɂ2GQ3Ur2橠=eU:Nnd9xF3egdK3^ۅrmW:Ce<0 sg. +Û i'Y P ]2]o:sDun)s\0)"FZ΀DJԀ ;BEbUЁ#0QJ% x}uiPM~Ս7Z׏"cH!| p=qVxAJ]P:%fM2{ Fg{bTWo JjCݽLJvt5em 3/j8R*PZ:*1;JG3o&P| onGiZYiɅ悢OV̗1Ƀ6))f4͹ :c%CꕶZk"/uM\hEnx{wڊn l=y&)W"2|1\־ dѴ ۲GJAt1Yz^*t(Dj"2OHe*Qϸ9%pBFQN8e%#OCΆs_'Cis,vv7S\~;"x3#_O4W\8:ـ>^r3Sݵrja@,$)82Ґ+gT4K;؃SSؓ2HQԷt@F&$)z+񯨁+Ji,c}Tz'Qg=+KNt$0DQbW)v|\5>{j7gȧˣdFE!e>gTd>YU) N㉰4)͕ []TTQ -֛<]֙?53  .5v|7ϤH"ڽ+ 4^|J)k'ATZgPD+R S+8 7:IxȱI`=pP%QJJ+WŻ֍GdZmSv(egeNj Qp6H)Q)FXBSlh-u2כBBV>"7b=I8qvYf^LPEq+އMPja#u!B gHJVJ¨2Pe@rңxrV+#խҽ0xE5d,j$JkÒ XtQ3Q"h }l.+Supz:G)͏?8K;^X,^$d{@zU X\Ngfop˳3}yO2CMm8;J( ҁ"ƗIiu%i E7oBD>hKAL1`03괗p` P!h''%Q;>ƻh-Jγ˘f ,ET /r,p]{j8w˙SQI /Ä#oq[+ {TtQ*b<X`#!  Q%h<w(9I6o{:;hZ0't T#%d(SꀛLؠHr\3Δ%72P#g ew( d-!,mQ|2 =,A.d'}+~ĒN-wSd5b U1/nom"-X;RZ3eA ʷwUBTuAL,"mrt)υz =hIzmok$[_t`^t0e2PZA-vRo'=jz6%oDaz|.jJh4 UI&EG'cEtOZ ߍYh3gm Y lA$s>>@xI֝Gi;PSWEOwE:c^x(EF$C'C0TVH`WJ8ƴ2S= uzۡxXӋ8a:-:AϦuax7i&͢s(9 Hc)6 y|RzW!) (Ѐt ~D,RBR̎ IQ;5(%&+Xe*&K1&VAB1ZIl*)NEQB  ]-"C f)Jf%ᅑt'vF8uyrt|<ċ+ Q#J&-ޱ)ˉ-*js9۷Af&؈[&s2I6o tL'/' L00edAdڦ늜-,~ AzIIArW~ ]2D&MOD2"uhF*>at+&~C:L>XoE8׺R:J*yA=_HWޕKPkEb N-*-W-c4Ӌ2K2&ar%وFJE , /tJAZ)X4{!փ={ Y[VuowGrN>XF _/i 7Yl %42,+H8l2yXR\-0jGV MaC,KAJBt #+/sJ:P]wF|+zr<ge_}u7Wv~Mژ޲|,˒!޲<;_Ms&78.rL;u` )Mt %Ln 1+Ġw L WNzE"[6&Y.F5֤IDk d&nO9HdMQ^e NY @RspEe,udmk"hgliPnTٻ]cG'-$LR2UAhB|JNj#lAjpSU5`M%IPX756~J$H,$Ni}-B޸\BRl}g_JJba%յf쌜-1uU:ӅqF ]>"57[waӋ).vMny&?gdSԉLA+d4"(9LtJR;lOC5bVSPզ2YN‡„돂T:߱5vsWvgq_;ھ=hqfN!.ibd9|U*H8Da;Fu5kP[Ѩɱ$h8 1b2[ޓ] wF5bW#5b5b7q|*ntF'Uam< >|7jqxW|Ps)iQBAYs %Ju019:DK9>wǷk';A컅%bkKfMG2aT@頻,k.Ԃ5k]TG+mg݅*wk^˵/:$[BqD3O!kf%\`4餆dDMS AYd_.7N-*gI9gtA`AwP y `0xWQԓզJ:ȍ\yg;U)o֭¬)8u3ɰU३i_p,/Jc2ogGa®[H(uBV2ڕR>EI%j *j+G6ˀ7 u+H: y=ðB G8,p]K eQvk&E|_ʫӈol&Z}R@շk] ,[ ,EufpG{,8 INA:C x;|`Oz6CR w*95^N$_MgO84{ \U0$QJ)pPv\F7RWxff_GQ_L¯_F>/~CWCxˡ-˧^L6BfOħո1;Ւ o?}nN՟!Gh\5"VU(+L6Ik)]\tRT@'U!^ӕYڍ 5WxȆQ 1ey4rgk| vz^uivfߪޕDi&Z˒kΘT֖]weȍM,fAsY{6,bJCín>m[os4U; l; \fdܾMIϗjcm󕖛+]樇gކJM<m?z}a?nk0].7 oOSVxqjtө% WNAv**o Qq"OɏI83翜ފ띋Ƶ%c;Uuvwg7*s=Oe1OP&Q'y~?kНjʣ=EevrWOݼۗ/Ad:M!,oZ95hf(\c2CZg(Jg3pRYMWΠ`;&S0F%ѐUL7hs*}9)";9ǵyҾGDΣl=U_V ?0LCL4c4[vɿ|F6YpZOrnr͌uC2RbVMr_ZoiXYn$'w_!JNfrk&qHxjOPr>n-fDW`Fsr,WZePÕgxFrퟎ6oOFh6(o7̧{ ~QaMmV;WS t#mqۿ>ٙ92t1}OGb |KPKAkŮi΁eVQhdI2h[v%˓~\ [kh׸% W˟VMWrӓ\>R0쩸-p*'5S4e5}2wlԖ:zC$G+aB"dP؆JYq`T&}=ؒ Y_hHKU3REF@Kgk,`oA*Q20@0EKI!RB)P儋NR2cNpDE* #{[Ys46*Ӌmko % bV攣VR KYU.,YG().<IQGN"4=C,!F2ÍU)#И񰍎I՚#IJDq06[^{;"Z"[}9]rϐehB,-n@\mB]hRѤ2Cdd.C5Thdku&_m k0fIJR[x8'޴آph'ieq1 |:d I#H Q>6.FSkPf/ٲ*I+JdT)j%xNdEމj Hz29pՑ\aMR"v ESps . QQ~B$BZ֒<$XabXQ+ TS۾brQT&f  $ԃŤSҎDU"lJ;#kf@憠FkE„ V@ZV YJBuP$ZjsAuQVm)$2]( 똌'0`F'X5+(_dXkAshSR3+NJv0pP(PyfJ́d0 G-S%ȭ*7)E"vuP6h %dCaW_gKf3S˺ F@Pj z (JƬA5̹( PV \rlsEΡd(( dJPJ,l[8UZpUTI`PXd!MBPPr/p$I?$X9bv(E P ?i:CU΢I/X*z]|%rzD R }+)f&q2F0[v@4rHBB#29Z$Mj8)dDw.Uѷƨ,,rl{_ͤBL~A+WV: 6ATEpn6{O( RdYxڵT54Ad``ЛAI,~o}AeA,]ADy Xs1JGlnsFmI*BaNTP=`EheDCJp #4D:EP@z[2a^kI Vˉ mpx\d>WI蓋+#cA'yM aŽBlW(jmQBRS,Z3Uk_A l s6" ~8c@sjEAkkN(_+kЪUF(*I2?{ƑBa~!&Y,bL/AhiMZ]9Mj[ÒduA!4zb2-@tպ.T t߬O{*̛ Q)/Zib;o$pWUxt6( 40hj8^K+KׅDw a ''{*tCOQ0TM _q-iVA15A.4Dm=i-5ti?0ˡj8:Vƒf5TmE0"!ܯ.b>(%Z owy( ^]C`xg<.8!~`PAT/b9_ŭVv*v-mm, E.8zLWS>׽C-㊠ {kda j+]6nXnKɪ//4. NW.Moi Apq37ۘM^fsKE1no8pu^5niX7w^?+P7u 1›`߮\>vL9MsQF e^I?Srf䯏HU ZJI |$\%4$J}(s[@@ϓxu@$'8 I Nq@$'8 I Nq@$'8 I Nq@$'8 I Nq@$'8 I Nq@$'8 I&,t@F0Lh> d@cL9 '8 I Nq@$'8 I Nq@$'8 I Nq@$'8 I Nq@$'8 I Nq@$'8 I Nq@M9+|@ @' U$}I @cLy$'8 I Nq@$'8 I Nq@$'8 I Nq@$'8 I Nq@$'8 I Nq@$'8 I NqhI QJ9:O' ҒIM Ԇ@cL'8 I Nq@$'8 I Nq@$'8 I Nq@$'8 I Nq@$'8 I Nq@$'8 I Nq@#JݝZE SFWIJJXΛZAM|nzGկM{^nCǸyضOle̼66oͣlM..vu_;o .^ SXT\*FU N n5|u[X\6|1LUJ:_Wmyc㭬 zk ESDoMQJFeJu은) zfInoF@&3Cܠ 6f (<@ "DB` py4mYtetc$+RR*:JuLx{=$ĵd+u*w]u1[!I͙B`c q=sWHUGɺKB~:B\IfҚL!)cUZDJXAFWK *!rRJźftl0 tXWizSif ]Qf]C^ZE +֞j*BZsRz 1+iVU i]]WHj J}$ fN=нgNr\N*CFc]I) 5[]HzoNCn@H BsQVT<c`A p2ƨ+Vyt Зh (}oj_; 7*ZPޭXWUJdp'$`TW$ zR[>w5F]E-+L\'(Bگ}? ) WWߎM[璀ZWi 4Z;|4JY.&*҃$!]!dtF*BڨrP:!YW#ԕ 2B\ i:wF,j f]渜T<0i]II%uػySItYЙ.Kq9TJJJKq/{0u*L';|uQ:!!n TFr@E}$K%<H!+֒@ē ZuJ9ue(UW\pTWHC(XWcԕ Q JM<J*ZDBJ˳ƨ+)Gi'GGFW녥+6w]!9c*``Kf0> R QW1j=!]!sq= i]W@`]}+Ⱖb*Ҁ $a &Z7)]^ՕjZC^FK"ЃTHEWH;P&ҋѕ 2s,=DhIo|mx,",N Z8)ܾIML݀NQln0B( aS q'o"@ (K=ƣ++׆)2B\/ i]WHpOIB`CZGEWH?-UkXW#ԕ:J׼8ҩ:\2׼IRqu5F]t>2Z'lBnºz]EOHW%]d]!4 )Ovt%lz9R(3ҀObsWir;9%Rt%t%YWmz)^i*]!Tt+4u5B])i5tjOµBQJRYuea3\B~$_ڱX`ѠWIkYܲHkdr#Y](mݙ>ZD(=3fMpQ Z:)5Kf$̉' ٟ'Gy$7‘>T"p4dtedBSϓ#%x*BB q+Z )%_+t.!TtV+t\]QW^)!]Utk ֙u^F K:AUd+wh~Hu5ҹ+v{*BڐO|ѕ:T(J~4ܡ]u*eV]])C^C& 8D2BܡoVDkeBJźZPX2BܡK>w]!en߱GWAީl֓RN=0S'y'!m(sUM6j.ݼ5\5FC[K#RF5MW>HQ ]EW˹ҁN,uN{P}h'оOlݞu0 X-S6.W(?3xu[k`]ݷv}aT}Q5dq"bBGв JQeaSV)ד;ۮwGNNN۔QϺG.Ϛo _ww- T?-wP|{+/g*wrt=z)wb ,g6^mpߓŗ/rq9u{|ݭw&73%?O7=-]wyb ^.;7\w{;o;3nq?Caw湯O_nsosY~4.`VۢiWyqem}Z*fe ŊŤuSZ\эEiZMҚnB:c ^ϛt0/~CgwgKx}pݲپJj߽w~<[Ï7v}ukrt3+A7ra' e[('}\.mzvz7?zl,{woK̺cSx<\}X~)`Ϯޏfr|nOIWw S@BWAef䯏)\a"L˙Sei)Cosώ뻇@/m  :o ̃l c U9%}y]ݶ^J]huS\C~cC<,ݓ8+Vp+'}k\RMYsSUs5] [/%.ԟm#GEX|g\vv瀝/=,Ab,Ė4$zwԒb,jb=U,>D3, iW 3p61>3׈,Ŀֳ$X'/_w@ do;Yۈ|~7-ri'_5⿶dN?8ٴ 8jǸ| ϏH_OLl !A#&S H)IDe 7O(&ṷڐߖFe1uLI,A*i~iBFpE$V$I&P) = yFܚ8xn{\cS~o4:uwwB6jMw)[Dp VC[m{ )*3zvk ܷc:n?.nbXRH.j^oԘGͲMͱr(Ør1}i=JA[Lrp B*wqmVNER ]b̀yL:sLB$8 )/i(CB(2r[eYpgg} 솕KPGNF纤,< oSϲ Rd+M|w-1d6[rV뀂lgsv+0ZQG)*]~[:0wdO4WV8F#sĜAX8EA{Ȓqә@&pZ1Q1OY¸rTD Ę}'g8 qmp :CNrA֠VtqR'ip0{ ` as(`Bʙ^2),uG+gw'N% E'ٮベoeou!qC#cܴ$Y2;u*-d5E$Q&EHIۄ䯭Ƀ*!묭v7Ӌ^"f|ǘR9SL Ҡ!*l)-qZ}0-$&PlcU^u!ղucnZk7f#jRMEn-JY,6@, b3pf7i~m dŕ#*O"G[wu%)JR$@b9Ud % 63tZ ~7IșϭȲK^Ã<4N ,'$,SIP>œUqjkƩRfn}AW1)CZCE: w:5cvc:b6"$! 5)%9YzPr K혮-q-,r,%q9s$Dv%/!؀SLF-MLD6DP*& j7IB?MkW%_#|>bܚ}v*tYR>ڮjŨ [Tgcdc)-h=s604ˢ'ךNZȁ d\0tm S#ʞ(筮~ihN%Q-R(-wVéio8RR  IiB/ L:6k,ŝZ:Nwۊ.kY/6lUl||j8g nE:;mߜ)qvfAlEu? J+-8*URI]8gp5YYc暁W"X;T)Jr2,2F:Y H'W6Bwb]oKL8bϚsI|n5: )%^/"y1iş^KӫŪۻ N:yU޵% d` Jk}袷lpTE@L;7%z쁡L h_.竧:N]ZIQqw/t^z AHde#Xr\kに9aS>&SI\6ǻgAJqĖiLVe2.ʸ8iMH,LV_"AU^ƲU ,L,wG&iAtKv \Q`=z@0*:lJ[i˶1$rn,yq2ɰM@$Hx0Ф!C)^Tdb(]*} ęlam?2ȍFn9J2tS/Y^f뤊Κ=IZbFj†g{qN08ygX:H\&FT4riARr[gQKR7کYԿc04 yhKJKmLl m3v 1,pzE&;e}9r$8`'@@%Nf1Jc6LS拮N>g,5EoIR>==WߚS doBo ]a7 {?X}Ç?~Ï>~< 0& |3_CC +ZCwxnj+ۼq1%Q=n$w_~?sJaGYhY{Mpzziv^5⢋Z<Z^Mb1]#^xSSeÏeg퍏4H_pD0 \ HB5֫ V)b,Qe",% C#=wa9]. ?m4 Y-EaPj-9DL!6dE9Zح鴲sQ Ǭ.M(HaUWs2Q'KYm*K<|׎V=盍_WvY\Msp#:\_3-FlczBm:"ӭ /aK,p@c.[3u}u5o{zNWmlOZVԲZh{n|=oܯhO˫|ӄ|\1zZ@˭ۄ˧mxp{ w728*y4eL'>RG48bIp aX1v>cb6m oпh˻숌4D|"<Χps ֧Aj]#E- 2EN- !7. |DQ&)E/]0 5-:a(K.,6N%JJ쒼As̀rmUA|A/' C1I'tB 8ΓW9+*m3˖ҔZsD /Z}* YmB*͌\όJ1.,62.|\x)NpuZ7 mZ~Xv_Pd~2^r:8\Q %S0E5BSȬM^B4,el/fΞ 0AM&G$yP.0#$?x 3v183MÖ.6:em]ڃ7'-TQwi=QLSBJ2%)ՑИ,2*CE{rMK9%s̙Ku!Qk\χS-!fb #2)̈f`āDry`Gjn3H)a6X<i?G2@* %I,(NS,yQ=zҔ+fpO/]=1],6JEF"#/wvy¹sA8!**HG{Σ! >@W!hJ$V,*$j(EQ&D3cc I2&!$t8=`>Qu쨒DKo<-zy9 u)ov>827ٱu Y%sKp>G> ^i.6O'Hx]B4ٔl-n@g+Nr^LRDAuDKS,@Z&DJy-1_Q3.%58T` ]_b.@uuVe{J-\nTcG,'-.1/ם}6Vb@PZ3pH5''TZd$6uŦ"[Vǫyͷ0^nSFk.U yh"mP^\Pl"1"Q5()@z5$3u0r@}/! vݶ61Dߌg&^`Ϋ_WڗWT8%&Eo[sp $!'ܙpRN$GFFc1 pbx.8bddM"`. 'zB|`5P<1w*& Pt rQ),Es\{<p\O Slf݁K-Q77]};־y3}Wor.r1p1Fe{fSZ4OgX-ѯo"S&ο%2ϐN8N-cK@PP'Pkѫr-tx_7fنuNd{7lTDuz~>ճg՟)sQTsc~KM)#7ɨ|ь|6Gs8w5?_m~̀~&35˟GSwl1˳vrGn ' ǔB͈5wQֹLuOħotZ :yma-Dc3Ʒx/.?{Oyr͓ pܜ_t٤Tq;e6ݼ6g_n,<y7 ϝqT |l6R_PYl.ӺEhW#jz1uoÀ jrWӦky{/Z6ڧoOgHmYM Ԟup3mEr=LZv'Þm 6);5A=fj[nDgatI H!+ πSnBY.td$+P~(?_v+*Ҍ]J*4W O2|)/ED@ˁ'4R`;!X2WTbcp;. ;U]ރQjR14>nlNwc7d~6^ֹi2/L^5'*,FQ;YIM{3dhL2HnmbE[K)8`GI*'$*]5R@k":Q&B斋|M+& ~H.چUڞzY8>w=كg&]H { 7lFՋGͷcG~l:js7;t&ޠzʷMTTUx8/VT8byyQr忞|<Gt3 ]5pM_*tQ#]I!= J)p5 ]eFt&lGHWJ(&I*֬7tBUF bGHWZN2`p%2ZyuQn)ց]I<@*~]eQWF!tQnͷtw7z~B$poؐc#ۡp%J-ⷠ+>ա]3F{DW佡 WUFU*Z @WBWT ML*f7trh:]e tJö3xOr3t{f\Dz;ŁO 6:VQYMޞOpl\T kZ-_x?CiK;GiUUS I6ǿ]VmCxۡ𕎗C:({$D";"k(M!/,Ho`~⢭`v Z]&rkuT=ߊnbMa+`#ªWoEٲ ];T luӷ~n&5(#݀)np7!eiP1#ʀ{d72]@WʩzDWXpA2Z*NW%!]Iň$2ҟ5 z2ZʺNW%#]=BRJ^ye\2\BW"] >F )k  ]esW%HWFƠGtK ]e2Z-NW]}5t%nhWo{{ɭjr%C G d%J-+1ա]2=+L`]evoU t*koxgښƑ_QL,l/lڪdf_vRʊu(9q濟(ɴ%9A*d)!@7q0ͳ8?]' F } *13HtvEYNqPreRc!⁑q5QAGs5)% >9O-G("<0hȑc:_ke(A@r1I"#uLgq7-Ep-S#l]=T04TdV&N&VDaRM( (Q5 Db(by4`:XrGcRޣ=UO;~ϴT{*X_y`bwHM[vꮨiK׺Fj;TtUi ֳJueZ5yQiCu%]QW\,]]Y*EGu%|,;,:4ԶД?4]+)+zʩCz{7~_ aM h9ܝ?AK{D"2RaIq!aG6U$$1'Atd@b= ſݒAO=| gex>4XOV(>JGhiq/U!ɩ/9XYp-W~d+y"w@?AE&~l߃@}z=ޭŪv|wOvoZ⍖ݛ3wDnAoݼKmR*w իm!Bu֎Mq/u(,\L`H/ 3cWgwA{͠%p\"JRbL/3iјk=| V]q S[&#HaBiASK`F|5̀GonJ^g1%HfLTl5#ˆ'yO(CU!u?$Js/H ,ʭ?Cz<aԓɛ|3S 7Xg@f3 %U$Pe<d8w as@-C(# PH6v!i!ؾة*#i74ke%F3,lO+?V6g>YOmH6n~&Es^ĀEp7qpou4vO%^ϳZPRT*P,HIS 26DH${ ߇cy}C WIeljB'>%4X_HxZx`ės餷y lv/dw#FEfMi0hwmO'V1 x۱60GZ+ÍTC̻qRsvvbQ|~;CC𧛨Q9b㮸G(\8^AOjfNk̰ Ww%gĦ=T@U1wP+!G7HMVEg[I@$*ZrK9=TA#?)UoRQ݈"c)y)WbzJHڟ٪@Nej@ w7 ɬHI !Һep ZOR!vx7Jf\S d&l KI:HZC~n9‰zZ%$4MHwP.V#2Bɒ|I[Xkn+o ,NrehTj#r^0hW2)NC5K"0\(=i m$GUV!.?cXʌ1G+U⛊Z*q }dϦa^,]%5o#~:m otnMl=W$;w z&Kfˁg3!4Nu/?]&jTԛ?~5(>5OUF] x*X6/A{{ idtET@h'+.16@$qH(I)|=1}sCˢ(L cRBH*eȢPp"Ij)xAƁ G)Z.yG=v4s"Cؤ9$E{Q@A RmkN@#w [D6(,6mB43P7[~E5j֙ (*AΔN^PQ\3RJ$xKp5!]MVLI@MƭY+1ceS1};ü,[+B,]URdfzw3R%2,;VT,}~'-jZtI% ѣNqgnssJf?>vrwd^-wy3N%S sfT6tӲYhW,jsܻJVfڬdT^۬ds!V۬dfp lL:*JfpY8V)+U5Ȋ>1H5?h7R ~ڬ%K٬%s$andpyXweת_;'E,C}ɾN.U$i'Y JaNUqNdÄU%3sv۬%?g"_;Jb.dco9I̙JRA9SItс.m%d~6}dUkf- )z|2y&Pɔګ,`JkU/ZUU^hT2MҸp6kV@X]M#W|P`YPC'^`u}9< 9 pv~,s݁d2 ^,-Vt_( X~ ϖ*OWzq1Hnj4>P6]\BT@]f ~YO&0/ 33 V |3hoֳѶ L#gh,Ac_Ws3DRa!J"Տ^ҝq9 c'?|m+o8FLDv;S!"c3;2J&:fQIw\JB&mvծx N~ W,}M۬:%Tmw)g{gwyÊUQb||iY؄6r@]ʼnK(uP `Wn6c j~ $ I'1L^}S//d&]k4.ϣ`L&RG0QQ>t 0uOG 2E? ?x B0HӗHx·b?MV=).LH^#µy· ֫k\sy 0 ͧO~ʀD39aD+YJ|9_8I?%zoh5 ˦eeÇ EIֲtDJ2ˊeW4kߓ[r ]ΧdxnNBgrE9do*Me/ۍ(!g/f ki=WB!JH|,>z6DV|8aphมWV95%uٵ et1I{1{cJ ]`,5R> RVzd0؏@"nBF3D&USF%9m̬HiJM,ВKdc0(l^Tɴ$BbMGG~Ԃ|==͗[XW,DF<`%/l<҄4`Ro\0y6NMf[WƏ^S">TMh-X]Qj\yqM yC4]sO.m::ɝs5=K̃C =C?(,ۍ]Jy,U# 7q9,=#'ǽwܮv>5p0ңō!1'^jUJjhzќGFD eBn(s:[\Ey !oo!RFa>ܲWKkK( P $>t.NDwR)ѧ>xiPmӉvm~Y39C l;vEŌĺAhf# `YgB)Ҍrz察)$>Qv˓tc]/64u-$3bD:foMfnڸVQ/FI$ȼ x7\/  "(ˆi8}ښՒ1} {s|ŚcY5)1kp9zo\_43Mh'h;d^3Ӏ{{m| YE~PLfC,(4A8ft1_PsAz5o؁lD dKn,_WY Ƃ׋PIhÄ6IX}ҙBO5Q<l'e Q6]}r˷17>j(N屎oʔᝯ5-JڦafR mb0;GkԇK?Ls]dqƀ4V "' ֹ8qI90&kosta{N1Ae[/K*Y_,۔LJ@в*R !F%VqOLdiy L%$llwBNw9岏ϽdQgD?GQYbgWg' @b].]8T';].#OűI]Xl#НM>La8*KN?$qE tʔ2-Q15ᓍ njpY!IN"I3/w/:R.Q6Hg?3,kf'>] Yf[C4<cJ!`G"Ywvwnk6 Va e*4ew5dH7R9^-NƝ@T M{*an߫}ŴR)F,:+F$WZC(,⠀4̢Aա3QqYJ w}:+k=QYr{dN&ؙ`ӞٜDs0zn"~weq庋?[:.W,dT}GYR,ٻLnzZyI^X$L{\ J- Ĵ|rz$Ue} Ӆ; S36DIJDQ7Z w:>?$VЈ b^Oe+;p.uG4)e-DIr(@|Nt>NWwUAHN$>o??qf^,bh#svCJHܛ^u6+Ä1;,\XU)`q:u@jGm:cԥ|6@ p 2fql14t=~c)_ĥ @ u2ݑ]fZrZrp"H~mA JhdEaDgY;o|/ۖ>h=Ǩq~kK&ỹ߇1h2~S W<*Rw1{p}XdsTRծ[:%ʍ)ǹ~g[>*gٸX>L-Z*cSE9rl%  $y!Lg@ s׏J% eZ!drOLpwp81^O,\ (!HJBW(1Qu7Z[-dZоǝbbu)nbT9/ϝxV#N|MBr i{20hphw7 VU &ј\2dZ,k(\w;{P3jF\LY3Ahia"FA#e%A`!4=ʸ,0_-v\~;Rd)lPe(w#me,wRhw7v@YƬ3bl`FBq5O~mF!K5#im^e+'Pf7>wBh>WIjq{y8vGfel+ta R<fШf֜q#,zZ=1dzm{<[X?C!P)գ{sgʃ+ Pø#FᵹZh.v +aѣ2澪nUo\* /{-Zg4sJ#?WLΑ ɰQt{T3,C35uc6h=!B/t\ ʃȆ9.DdG zl w9<NsOlV}kܥi69ܬy7;@A ]y($- e ?s^eD`ݳ2m+aUH-ƋQYh,,gEPj/ʔW?b UQv`2_[=m1T..mf">텝[hoK.V/:7q3b˻ZiІ9.[G~* F(P=*ZL͜X| K$lEX,R^?+H! 0{ }8uK*l!嶃sG(L|%|zq)zH f?xw9ЈO*f-T7=ʳh|hهYηuHF^Vv[/a$qE9+sBT45Ȭlay8gKWI='H~/›OsӢ|H~iQI&! ױE;$l=A=3 [6̽/a/Di}zg3=Hu ᮆ9.ը}WM=٢6iFFSo}͸Y;̺mz8UHP 6w@;GXYc*|ai 0e>?(1ޢ^'0Džy@AqzCPkn(A}`΅Ao渺7A?2^C y6$jQ3E/zmѾ,!j@YUgx3:-й+iE 2]~Gis ORi}CXR;o[j#W2ȸ蔡o3P 帑ʡ,~EhN^InH(9-8bؚ1H2Wj^az^w^+y^bDF)#DݫڪJHf5ӧ(Ap c8 0ӧ3}mqiEygW5cj!4ʎq 5-~bгrEC"}wQHxß1o4{*~biS -XFyj$cJ;9YQKa";OĢ wUGxY֩2xE)_Ņ匥pU}b0a $UʔMz\ߥLp-S5(b8M\%_ERY>'mۋ=`ԝ+".=U61VOV]8M> ϟc~/&aq/tlGBh&tr`F9g>=Al4k~PE^[q =Ίem?6_ܪPѣBu+ՏȬ",6 1eaoBTZܭQE"Lu `:$g[5r45yC jCi:=MًK/+ sa/MJuy9-ifxˤf ;1F .#SMd L!54w7)IfD &R5kGMuo+n&8Tz7W^HIQm4< V#sՂAL7 Cd XOz IEFBI ر5(UנFAMAiS$re07;!HJAW(ᕄU~vS֧?ĵҎM~}JASIlQ[Lq2:G$+5;j1\]Pُb΂AP~fY% 40)m )+  !pQsZrKUqIqqi)[X0*:Oe!2BfiEIqr.ݮ2ʼn W~G#T >5|~qdTپ.Niaσ_,Gt:6 Tg*aA}aɢmo1q6EBR T;e'ɺ2P`=ՕMwuU&!;2A% ^Q+ǚ:1-Y֧=y !ysyշME֔pA]@PVxKi2"Y=tivxH#Wp{@FR#ô(BdRd~}N;zknUVL[/+PОTK-ƋQj)ah^*o/曈[ޭԒY6oRd7jYxqWש J{ҍ"W(H%lC 9RQ}s&9+a.;?HG+..j D׽!nuOBe_WũZA~ s\I tahWP_߻ u&[~bğl|7):;|KYZ?R0 t;e[;Fe]Y8+cg}0ݽ2Xwo? Ϊtp /P%L.0S.[`#/dʶ TDJUb;&j<yRA^PQ{ZT:AhQf]s=ዾe}  W> Cgik{}!&m&H[کy\ke!YC~]`:SET#q a|&bފ3 *| 02 iXq {U%j_qMk\+h\r]x41.y!C°϶yU;\{VzyX#檲5\]X3Rv% 8SsaZ,<$FZCM&~x䓇Y~4G\?=^q?F_~w`Rƪy,'z:/!n5mlz`8Gzy8N|zQK>I!(&8 qLwk{H,m;H (lpNKPl}Ut@aڛB#WBEu\2!WMGIu62W9 e__~Hbm:1NTfV5pZ# ,աm=p_? 4Kq?##,G'v>:LnI~A^F F.PR$:WAgGrPxZ?en)nώ3J<4> $KnGY!SM&g߽yh֋0i5@faGPY|lpTx}?rRS5]['89:K6=.n. @IVe㛭 ~E|_2Y?A()% Ǜ3$@;t}!0+Z}}~V[0Nv6x[[z\)w\)0[RX,G+"ؒ"r՘A CD@=^PiKPgzY}[X)27TKP-Hl ZW&p*57%_Hbg rl} RJzpILFX -]K@kGVj7$lk :DVv-Fea_yVWwVP*h\4IoNW H'U`F%_}e5?Ƶ2J&RK/כPABDG*;|xCc ϗדkDH& a6$kmNY7M)=jyK 9*l?|\S䮃љU|Xg}:̪U.U~?7 Tߍjmb S좬扽B o<"YOEĪTA:PJ{%Z{_#JE 'ĒVyUg7ݴčהIH6sm0UD55K Wy"Kck-+.K=^ժogxRҬ;qwR5Zٿ!97 ҿDG2RUZl|2R:bM=[m-H8t tS!t(7p}j #x kzU8૿"ԖSyO)ߵN`}`5T1i`h l 05F@Ũ砑r`ນ<fΘ1H{/gxhGw_wAQ&)YEE_DRKƐ,)Op{4B6jA ͛[T%m;*K3R=>a2-EN9sءE/.ލ8r1D-f*ߴ#@׃ ¼ZpZe|舝$8%,<;]0Ģe kZ"HgC$&w&Tq} d!P s:a#;@s̩``LU ,z#?'j 6CC0d[e NlUŞgق%X@{[8XR"ʃ! IbuwIQ 5K Ɠ"٠$l3Ŀ(|_>4pc&);)~_|Gt3Gi>V߼,޾?A:L> |Ԏ-4{#?`4(>20K,N/#tXf IZ-^aHrԢxd z܋D Χ[F_,3xg+FHz5@0 Smν` sa-rjqJz(3Փb}_2X O סI4y`,(A'm&șp`>qg~&uyll˛nc YDiSmro%hq%v)p.B/40X.@*L;\qLlCP FwCN?Kq.aщλk:Zؿh».By ڳqY`b zθE_tt^ ]"S?S-w?~4(R]َ^kP }M͊J'~-Ctd<e{-6bmeC(AkXf-,6&pm(vmMYj]yۮ";O{std[hI<<'ApvpJB׹{dBqsZj'[g{Ty2Ր%to!;lҼnWr6k^939:iI1뵃PV-枌RƲuSK8ߵϚ@06XPyK0H^۫7W2$"X% .O^ ;-gj)i6cJWb8zPpWpݮaivOAoN4}rx%a ^%7NJ_ WXB50T 3?O]c'73)9ʰ^Y]/5k+{p%3*vL_-ԎuVU$y6nEZ4ַ7;ˬїл֒qi:XNxJYoi')'Dz(YS: R}1#3{CeuP&y1 Nbq_m_|Muhn4 aQ )ߌ(Ii:]uuԵF4,BŌS2I,CbF7CqcJB|1?aM㦚ڮff{"ҳf^3*`*̜`] HXsLC*hl9QJBqEq9PvP=˾#tC+!}iFuSLEC 0ۚGqJTXfmnW*L/#M ƒ 1H tto UHPReT[Xψ6#Jnn~φV]JAU ܴE\˄,>J)qa12+8K@`ahpf`02{{bԻ3XF`x{GeuY=HAi<ח4f`Sm:N咇N1$8ZN,HWbiAzCm%I@vK HF 'G# }ph-}TŹo=}0O/5tDq\ $MJG+h1˥qi}->^k.6?%PV`$^iD(&5TPRq8Z4 p8--9T 0Ы2g2T28XC^q.cٷ:STDRMߒ'mm2|nphCo=osc6ڌ S!R͒f Z(a%Wmjf&Bfsyrׂ ߨXIk/ gI )1# d5g%.=sHyDrH_ץ1W 詆1vcw\]̧LC,>v3܌+-U'\+T6V\\2WQ~W9҇#)T'0!483}ͷKhp>'Rc4g4"IO(5ې>.R!4•jz&H k\$P#߉VV 51PX"`~) JDkI6l>7Ӫޕ$Bii}ì4tO YY-I7"YE&YRu4Qw|Kh1Tʂh%$* IlT=iCri"C @N47L4nm*n'P\jIm2Xpº1aV=e pAƁ2A#iE$dUPG6y)pَrMo' 5g㮿J]^+uzRyw=\_ +AaRfIfM#Jf6,cQ/9BDhw*JzUG~Mxߨ&Tjw!KF [!*i Z1WGeJtQ=^J xIxU)(U9A[j2q6sqnwb"Jht*Mz*^vԫЎzoGE,ql `%u|,,NQX[*ho1׋U UҋUC>7G[Kp$\3z 9ezqEz$霑@ǚ%AJ \VA*HUi볼9 +TB[J;OUZw1B*JF0@\M@,(Vvdp=W쪀79Ԝz&s!6V jHAfR.:Rl%a43k_ ŸJ0\OT qC{=͇֎xt(8>v/Z՜Y5^#xt* ëA;ЈkS~>rz^BU%`fY`M KҖDQI\Bkf2 %O[":,QPDUǟlCMLIxuU`ƚW ~E%y#OW2guؚ]Rϋ_\| ן_n7_`XjYR`Xi"4^ B;H7`ʙ~>5"h#(ewh*X?101jtCade =Ɋc/LB1` 74әʨGɧFgvUxy-yJl=fki#i%UvZVv6k}cPzc*b9D;*MSʄQ&SD9o&_$XwUa*1RշLJof jf+k*tg. KA(֎WIXiLoQ6.I6hc##0c6 Uu$6ch½ΔC>%x ?8 t  G9|#c51C`U wHaXf"4v'R,%E*Y\SFZ[Ϩ$uӉVv^hӎ~9bHc"_4q &x½0 mLqqezjwN"uwpiwj}ݕws 樒ҷI^-z !4;CɽnP+f ⷷS ˧֐z ,O+]ţ LѺX憵W#2f;*`5Ѻ1m1[.k#Gl8^TEE4֑EjH ) 9Kڂ=!KJ8z|vwydM@0](ZdÄH G2ep_rG2aB/oyDqn3Hb&!R5Nz JvmG$$Y%4y-ҭ2ECS|-$Qv2;aW\U᪠!`/F3l؄Ć q-s!^+ec0=1?h*rI{+Z벏s_eLF m F{jh!qY+d"3n'i4?z=HP=*8b{L5,Ay8W<3xN  ѱ2LУt"+`[d}pCPýM&.$[k"E[\$J bٲK:ֳ|89Hn\I14 T@2v5v1~6LbޭΙ$X8[1.6#Zt2<E47("=PuEz> ԵXEhj9 ?=|-m,a(FSa%L8TB@^Tb~gܾ)yYN k[$X pLQfaN=:z|<1q%4:Ǔ*nt@C =FHc&˲4L6hli_UHS-Ur36˃E E ¨BsEr-9ū\[5oKA{(Ì+R*"Qi)2U|#*WN۱90kߵcC\Ų]&ѧ,L1h wg6хF=GڠQAjX3Nh ^6jn!49EmQ@ %4:b4?nV'Q)/3*i6u޻1b8;axMHu%4:\8!}g<*(j8Ҝuv88;\۬};oQ(X%E=Պm5xO!/.^ӌyZ݌6tU,@_B4ЮGZxhu 3 !<;*4bVwuòMxc5.YR뤺y7P8(4ua-SEUDqinJpim#uWt8Ge`0EFt MC w*xՌK|Ǒmb'3p;"(}XBO tZۢm;ך6JhtS6q'Ͻ(?C*@閂Wn*S҃%4CPVL饯(}cࣷ=reю5i}Br<'w2ƩF~YgҊ'}e,q~j.`AXal߻%] 2 }=tgx<0FD7fS  ~ s[?rK/>0j\30qėX}~ˡ`͏g8[>-ڦ=Uot煣qc؎Wۃ:`7FNl|;?ruh@ߣ0OG3P"H@ 1^1J˒|z6 +нYcaQZmsLS69 -Wm! DZeo_O]rPӔhi] mFij+045,i@$3&-!R8( ;Xn F&_{į0N歖5wƾ`iZih9i0&J(-]&O&l;`/RTñ~Zh3 SlJG1 #)szǔ|&N]|-OTR[bX?>*(|*YI%LgTBF:t0/[ OP>s<p:0c8 R`pbb,3f ߽~d11ܘnNeÍKz`1sqb펢~XMQ+B9H ,72ڏ$S<~d0{ɻ>Ն\,7FcW|c_?-p0|oK3\(NHWoW:5T&Xתvk/ƭI=,1ʱ .aI׸# k}Tn׷-6(k+ ׻՘ߢ>Q'˔camQq1lٸf5^&[T_KV;dxc~{zȺc-uR פuŁtRD\ th)&jh] iAzU#TOl0,8WZcDl6# z0O=5gԈD@{|rir0xNyҦZFP`@!i3XIIiKeg,͚x&(Uѓog*(lXʭ 7혚@Pa n?.j]" V3ɭd[)F ?dR"$imךAKhtSJ7|dY^Bn'Bj+T(]Y[rT82%4AQ* ?2e w>_/LcH9 :,enbi~&a6Hy艚8xN3o3v6k0I!D!J?qrFtb/ܦӔ?{6 Y:ul`vIcfZe-ɎF=%6%6i@|YsN}umjӹQ}TT<[~X?Uh'srJ]vfCnns:VK<ۈ$ޛ4qg?bw1y*mxnHTsb_ѳ#xO~^!Fh[K蹼enu .{ƣG& ;\(s7Z#KָכcNs'e ԯzSt#nvvpɶMU$|=\/6Q%)>CvW|FRU Q+U o+}}*]9:zyfK> aS#P( >ͅq /y.5gG w~u5+baltޔ҈3'V8E5s7})}{WR{Ur@{ßpH,)Uǻvctk5l?(UJyQw!Gldu]T|vZ#Ϭy?t3njB U{6}C4G왱7yB1Bk|lƳtngBt91 ?bQ>%JLnQT Bzg_a>{Bׁ'>5:]U7l+%7qI 3*;fא]Lj핬qjjSOm5 fŧڢ3nERD]d4G`JV:RiSwc_q#PVwL(p_JbDe+vЀ{$7N?=/.oVƟ=R|*υ6<ը8<.{JxkuPYE̛V3RD)*IAۿoFaQYr%kr5xE8]lBt ^U*3.YWw`ŕ"5!XpG2 #AR鄡 w`&v{"!yA?3{72yeޞ*TuTpgM0_iX,]1%y0Ig\ oH˫̨bJxlϰM[y^2Ł>7QlhRDז}Rb#v FR!)4}l<}ۃq5s˝J8,B/KD@iC `W-.ԋdJTעz9^..`/$Q̽[4@ ƺ{\+ *YN0dq=I~~.=vh{#*D<)@Yҽ7^&Yhrc*`o~ZݓJ2Wz# I/g!\ ǟ֟Xb[8]gai=g&χOGa괃<|&k?= r2IUapgi$NTT>~#X"OE9@ϙ4&7ޓ*9t VO_aK 3+=O5FkjzY8koRUy`#hSG)jGHh~>qШC۽7KsϘˇo ^3Fo糧-iOj-jkb~16py(ΔY+1̣6ى莭|Vwy1E/a2.<]})DڗG* .dή\@`6m :-kTLjIV$mR@y^\9H+n.,Φ&W7gsai> b.M9 q3D>B"pGT\aE-O8p \ۉoВPBEo~nZۥv?lW%CaW̦S̑-Dw|żLri'ّ 4cSPQk3- 79_* X~v9 frs|˧,T}&cLQl4ԾM5ف.v*XgPIFYXKؑ<ٕqǗ٩&#`U(gpH2,J 0SBI5©+!SJ00ҐiBXxrzSU(K.,u"]uKTl?+![5@>BBIܴ@ (Ê+ C}(rIB9zcWNyݛNDk2k5(t⺆V&bUF XA"7^qx͵bid>L Nҥp -YiYs\Ϛl#"1Y’ѽy:iN4 NsO}}'On8 C & aV+((j>;AԺo^{ر"J%Q1dQ=t5QkLƱ@WTƫ'$ե}PY{tvgɆjDz*ۤ#S̯ҒލJ+~Qh[󄶀?+0? Z 8-:lr#G!(a ScroobVj=Y"#AqHPZ8!$(Z &7z}ĉ$miF;DTQ,$R}7]1<(<yz'I"V'd<}Rb);qWJT{(8fJcJrmZ<2G"69 \@s/9e(bH>j pyH̊@>4yobh /VG75' 4O [1m!/e 3}# ȇr`$$*PƄ"Q XbfA~S>60Թ. PJ|~׻=:?y?iE{RXh!S-r>? *y7*$mp@ج=B6Iޠ}w![|95A;@,ju@) |EMJj&lkIQ-@ݤ^fQDŜ[GjE:ILҎ3!pτF2CA @σoDQ!q+nҶ*I'qj%6H쀞* %Hd&pLA=ƦʓTE$zbE`maR.*3bns,S W)BI|t;PHPuhςÌV :ou58Ȍȶ?g'-nHQYn jDS4NhHx3ճcdžtγLɽx 7s=9Cm7#G#nFA!"MTh<u6116s  D.1ґ /`)9p>Lý$koRЍD6Mgp=&0|AJq-7Dl=O@TmGܘN[컨2l>>vX{6r$/w=NXdEL,0s8beٖn9G8Hn5Uz? [E@RΥ0^'(cI; s KXde/xt^JVYMYl/9|RwA.|zXV]*Uɖ)SRxhz9p-[8lΧdtXC6&dlE6+Mj>-0dsٜkݚ{|ٜ,K_PVxشju<ϱjkԐ9ds~g?&m_U`^y0jو xYBzlRdhChlL843#W4M YWQZN4 *@%JTMu!:$67O(=vfcDYͪl 5Ѳ$0u9 L%z>^$Bcl֊r]Gf;6j$0 x[3MaE T9o)jS{Qt`?[0e s@AR` [y~Q. Q:@0<>0Rm3xmD/Wxz&ɬ@%ȑhڧw"?BVX8jc-y [-**? 2s BICr{AQNΙH%c )6)S^9C) X\Q4+nZ#bcDSG1YlMZ&/}\AлrzmY_>5'މuL\ /BAe'j\ʘvK-=(4 ~y$TW,TeO &K~SSNBa:o.6{\tYL2Q>; ȿy7q70'k8>|Kl `b@qp[&Z: CV]ʑcxz,0gSU"yozf1' 7ԇ,"83z z/oJuj;0Nn17{ay9Seq>G~![lhM V +: ̦BkIFjǦY_sֱmbfԪY} 5Vl[^ba Î>_F)= ^ Rv/l+;v(vF|`ⅵ|1^_aj>2NTk(,vm yAD؞$%"Na+Cl5 Ъ5Rc3-fAf I"$7Z)+[LR97J}$i9y2Ufoڄ(\5*϶Td1*x 6%9jY`u. MaSWz'^L_]#B>9TT^gXȖXP\i<+,̣$I l\H.o $!EVGΗD!ZEħ@;YZےشq蒭X *m3@ZE36= ?=@B`VY6u5VKc0fJѓJQJ]'IQgmr*bk`A.I.1 Z.F>hQ)1ʆs'jV*嬪뜃OQ"C*ČlKI ۣT=@̅?Ur-xوl yBɗp[vCb˳LlnnbyUe;ҝnBkɹYXX:g_a^t-jz&&F\M˜5VָX7Dwp5o7te aVSuz})<3>98z_sf_ޝn\/6tAW_3ÔYB{ajǡv>}^L.NO|wnt<>t~Xnj}PMK3 lf%t k 侳(#8E՜?nڤ29?G&0?qRhLIM<|NjG 0w`6o/˞]7q9E>^/ݽt߅vH%߅sq5lU|kϳʱ!W H׎D h}LOc۬v-([Qy$<럷g{3kʚ{2MUw1S"^LXGC(03 deL:4@.gDuк+֒Zk0%%P[v^G-1텬 ,0փècl5V|,:ӟrNsa: Ɋ5ؤ5SGpH6ZRq*,RG&58&6Ml $*)A0 H/K֝Bf N'sڮ]V29R_U0Q$PN?|4yUBrD6RR"SQ̠&bWph9. kMG쾰E59k[Fs'haвp">CvMElXZ^[u* h2tBf[EH;Oc 9̼P] gA@FC8c3+#ebc6|uxI$"hb"; bdN}:6*Pot:}Ue|f䅐!Nn:^n{ p0ڤ MWwR]o]92YJ1GrADA( r~agYtXA>Bكau|>Ne#ztBQǠ B FI9У艓rvH!{Kv*jўքuh->1Z?YN iAsfJ[427_imR*9l_HH]lnJXRS56(w);OƘ H |FAdbmIRX>P6}056?_.z ?9<}n[x<߮zk.}ȪUy/tR/ OTbFbQnPDJ*@_\F ƒ|kOPNIꄜʗ0YN(ڰƋDכet1/bYBhmT1dPvLwW$g-tq@ߪ 4O- i@. `aZU[1*ЫjCb|Hznjh)aO$>vkА`hd7 C%*֨_SR^˷!Y8-|ԶU0_M]} "zͧP)0sYc+Zl$/ ?8_~1u%X~?K!3wgf)2a< S1|q?גi_?,a|~et<>tM@6u؏.b}%~"$|P@ AЁl* 4NGO?ʔ͉gt_y8HCS4^L/x=/ר44W _Yzr̸rx~:|GazM&{#$PZ'VZS5_K5_ rbAxUbBNAi jRK$Za^ uRאեoh:XGTABK*O1suNcQgd$ 2%Ih!!YF+N6H5[lw wjpF]r:p5C`pyRǦvGڐ24,yAt>O1Wcͳ]޻:(+Z4.W^=tܨnޡ9_q>1Afɼ:^<*}懫Fp}1clZh2=Kܟ.cf# b'wazӣa;CfP{wDRyGes;?R:ɋ!2hp >" xHjkͼf 4)B3+=N߱051iɮ$`E3dTv"*K:.vN6xUVz*,vX7hW#j9˶&Ě|)WX>TN-Z{G,ᙓttJm smRW'6,?6-EVr8-q@$!UY!u!r$2CM$[dYUOU-UC4s\w-qH('x >>{F8$@qDҒMh5t-T+lU@e&*OK&R3N5= 6j$8lg"pΪC WA5 tĒ yJ1%jiqFK @ݔsPs Dʄ&MV E54J!F̚j |ax k%B%9$[ VfL ys05D63nr(4p=-)% 6h׋$ȚP pbbI[-7m^r}m%s`V[)2q$'NJurCn iŗa }q5m-3 OPw  cD!&fR81z'bK>T}4"zUK\^wďa`& Il$vhzi'T̘5S$߭T4ȃk֞f\)AACJ{'K^ sNv=!MƖݛQ*pMZIÕqR]dkBT8u01MR 4ض whXJO^{EѸ2ek*`D[ꕤz RbbjA$\j6z<1?N ~c0peZ c#/_7,mb FNn+`k>Trr'- "1=`2<8_>_RM'ӧ_Dd"KnD${;%ը6SlJb=ߛ N?c#hv+cDQ}:<&9d\;m_c-}ѵuku^ۍdk`pf>9 ~nhk1Zxb޽ڼzeon>>^ͤosߗB|z"&3MNh3 |c~+`C|#>>/VLYëP('%)рOCX} 3&6 h ZFJ5*X^s-HK-&yrd FV%mAsryb-fhR/j|l%횖Ȕ#!['|rYwma.6/p9xll<Jr=gp1H9E95[>\"i!E:䟌bϝV!zXjp%0pnfnniQ`~V% (66{CoX YbKKU԰aIYGw6D)cR$)&Ǥoan >|L >&ooH(zrkY̘}jfJ!aoH@ řE؂6 Cw vo6"a&R|&7PskA}E*4t)/(P޹'0F!A׿% S Px,$< irX ' 3cbKKHLT#[ڜF԰DW/Jȗ1j7wR@W)Qc~ڽ۔eb,z0[y~]R(C'9w&k.mͷ.{]Neqt?~8w2mxW~`xHL3چkg͂mKrJ0$T0T@FuOEk<BWЃ5lvjҠȝhdfQ8`W콛"(;UP+H~ &&%@nhd1Tc싦GYd-ΜQN[suB)DPJF{fG+-.1l'[Xi֦sXT]J;V8l1T$6|N{chz[Y1$$ Do6aF)V/SC0Qh ]YZ jp3p$cIAʶ'ahl>}ыɡ5ř6%8isLTNu'iI 0FofͰ8kڬ4Y - ꒘)]e+>Q0pq[_)%d\[ʺr̀TWWY6ȫg0rVˀƋK/njÊzHX/ŐYK}5qdZKu"|Roεܷ[޾|ZϮ5dz}ԷYAr*Rߞo#E o&m} ˳!{[u̫+:_ﱗ7)):Ƹ](#t9 >U~ՅSܖW=H{ƔSKDi(އ ]tVmo^#`6#^]5"2ښ׼4 cf2tb]CjR0<B:`z OV?FuߨnFXǨ*~j̽5=Fu2y-J2ݹ>E#߯sǟ *|tR}4H=\ k(K M_~qqT)qj|I/G=1bgA/` B0(w6 hsI_ZgFCtßosJs+`6+IB+3ynn"WCL-n$Pm(= 7ظ-heTWf it1LA#2a .WU~1UhtAØY&q!Ǹ<~gm$ŭฦO+1OB GSXg1iI)_*6%'ܳ+56UM9Ye^WABcVOr '//`ՆtMN%vʁq!%P@Sa>T)*bpɖʔ<g߄N"`Gڏ۩-T]6?NX96/yUqNl@l=;InYaARvs]*,x7Q f2sv:ƛ͊eB'Ld&S)%o)OjFAz#U6cc7jf;5mɧ%ԙ⋗,f=BjWr]6ElV2[i2}uVG;Jx7"jLo>MCd~܄\":м FZQjQYzѠ< $"W To39(ė䰾gZҩ ǔ?fJ6Ԝ?@J6m%rBd?EM!Yͦ+4.8"d ؁c;{Ϸ7{&/<s[NZNLXx< XW^.:oϰ;BN#'/ϒuer?IlW8ͦ/5lD*璝7'fxxx#;@Jb|(WbT][s[7+*=\KPVC*ˌk WK,ɢd˞")RG%% B@nFV]m!jƈA "9d)Z`Q{Xmu[d$ܑb&H3J.I1?k0aO$|_ʊB{/u f|,[> !Ta'o M"X FD ʵ@[T@HpL|r*ThRD r)Ȃ˱5^rMX ;#dLP,:nNKUrWiK6[|:&upw<%h5*T7 XNR~}~ekAi!CG>F?̻g*YVo:foLfkq'Ŷތ-s dUXLkrֲռŦcŴl/:Z,Rmc3J.iS3w1nUs9omL5GY:`Zc4׬ԇXD2YY 7T[l*PRDsweh2wA+ㅺ2S`w$+-K^% |e9?%'ղt,\nLM@:w.zݯ$u7 ;27WI~,[L׍<`WcN!Pt96#B>hY7&gJ[}?jO][μ'& $,2*Z_807aЯ&CivǩmMA4je:,vf' >X1i%6]4n|VqY;?ȅ-gmjۼfaNeJɖ9QCcoX>c²XhGn{ݏsO5nL[ߋ5z/~?LCny VZ+f^+>NO/3Ӓ/"m~?8|ss+kÓa0 uSjzӫѿ>k>8dpMc\rmtޠ$2@|h RٺjhFȖ9 dށrs 7weVK+Yԭeɘam@ oC ּ=d˜9C.ώL0^Y8,Bٰpgɘ8 c(}Jɖ9 U_}*f ۇo{h$SEKA6}X7~T Ux93JtZ{U+O_֞1{=QMg(A[(2#bڼ@vWx2WM.'F*l 0j{~eJVH!GjqC-2lkcE a z+ř#d#ev%#/j8?ƣCA m[{+ ]32ir% nGw>;wӆi'3XjaktץUk(g:fouAh3"X}j댒-s:pkT=~;Nb7hmB`78Zcvvmt%m["p,։A3BR;l/r`N}* 8ܝG/5zzNQ!cE.]&|2 OVLc[T'~uS y*-D+^gKavݫ-N!Ͼs$~I-ZǾŒ4H^^p 7}JTote1RkXj7<C|0%r҅:7fOеr#=+Fa~%[TJK-X`>[`vӢwy*qXٻdnqMqvhںXIƐٰ[J q* +YK kKsWUK 阽Y!-l f\I =ѠiC߮ۡ6rي2(PI߬1 4ZCJrn{Pǔ%E+1zb~9c:4N,O}4 sL]4|p8!dnJC9uf_vytMyl\Oy|'Pe Vey1 2"/ԙcrGj3fE=: Y oNcSbks'[ u"%z=ø<`SrW׃7U^_뫋7g㫫ׯx$HG$H!+zp/=K;j^r= 4<}{Ӆv RT/͍`oKOѼٌgGTaٛÅs/<>eOpGЁFOҏy":4 DJgF Cՠc13ȈC$b֠})BЂE_p[Sf8yIZ}!"V&٘S Xۉ|bO^Ӣar GjmR,B'!c%dq~#.B>BK =Ӽ=邦܈vw):S.4{N4\얠ը:8+6LIa$E,s)JB Մ9@.b]tlc b5ʉZє4@t&b3 [oj?ø YFD*z$Fh+ 2W6c7#$mfb\%=+ߍ|>Jʭ6kB'l>m:XU.d[a85duT@BXFd H& }lZ`ڛheAآ6860VJZx4&U*(MY(aT> E-z?6]35;2YpUVb Z>n|BOYA_fg'> ̇ӛ`fliuv+-[%w3W,Q%"%mؖ:ޫzJa0 ^!ap"ZHv:tSŠ(çM $/8"%ˌ4@ !p,@spgJ(;C@0(K5uIÎEiڄ 9"u?d6~ؓ&TJcm[>%H=+ru)ln kdŘ 7%-ru$fm(P<]#8$> DF\Mg1!5(j9yެ .`Tp#@\@XG-]r舱0QIhmPQZ=An I2wlB@d%eL*6D<}6aMX3?o*rr v_˝-) f:Hl r>$ KDy4~6`+~B>37s^MR}ɏgL3aEP0o)5ހ6p@@YÉnb5IlU![7Ig !AMd]2 o >z]&sm4N|m΄:*bl~qZeܭ]j)RBK]>h׌#sRR%GM T 'šRpMuuCZY;~[iym5d-\㤍e^>9wr.R2]2׈Kn{㛾dŘVZO$V'ׇcW@@E@1 hc*!IS%-֮qzJRI'%D4MӀF1Ms0l`Oၽi7M#}25«X)ܯիv[Y:trqjB<_ Q 4zt2 tp 8%;)]g3SK`=L0DP7&|gX3^v(~q~7G}KT &FV3Wv[bP4ȫD\OoM  pנpYíU27Ww[?F һERntlV({{4ޙL7}(|釋b+ =aq2`jEQib k톿6`,Nϳ8.!zqO\4' :^+.p uz7#'>1f"WbI\.&n as}ŇEVh7;km.^zv:x.&ԔJN@F"hw5j nh챫5.$G' P(|ż/]*rR7Ub9Hqc遺)3zr)R7%R0͛V`SS)t =HE)ɴM70^@h_-^.l=~~z>z)@A˴@19Blk: ޙ2J`r>lX']mkhQP.BY43C-ٻ75FtATߣwWRTc\$q"5Q)dQf[ 2j^)@}hs F#- \֥!R-Gh` K0(9&zcXb%!/A),hr߄="i4))#/)mHQE4,)۰aP*JFn(U`@!y9ИPrݣ˭JMKɢ;/&Wc,Hվ++1IK8V ĤM S*3 U3Db&RC#}J7&4'EGRw[) lk[?^Mx)';!3q ''mIxYd}f)p ټ&z~qZ^ʺ0Ëb2.fWK}0 I8&Js9.y;3:O-E8>gJ3*hMsJ]"PޖD _j2GS'2<GKoxa]R.{^Q^n]>XB1EvpdeuEHikfom fLt{3fo\:EpoxzraoF@0]~UG" 2y_pO$F  ȭ*)vQzsμBf%9$kQHB)㖴mX4~^HjT!Q?$]ķv߮ʉonJI!hi7b̍Ep8R[')Fy&Qi'j ,QU7궓}ޛXޜ iJKU4 +rO$OIܵ,@/ <9h4Oaլ^C֨Je:sۢyXy.<{ય-yu%~X Rg.~2j"5hh1yLaDoNwXࣙSGWBĿ/v:N'jdӼûxE) b0Ùw^Hߌ˧50["Ls]d %D*)7TЖyN >27 @k>ت6PP"k o7&2@:c1iL‚ck!fS^nb[#ɔG֪I*Cu^RbQ3VQK4RxL ,c[#*OEEtLG}+ ŠQ'\:i.B`2168L447kg=pn d-=V[E`ϨoVTr5R:aS~dj 8gGǏw"4憧gIHښl?x0ˋ FK,TH_?66iMp/w,p67=ZF,@#+3 hW1CdFd+hO"Na+9)z2|½ na|n5XhP׶8jP d&L kSiT\rMYzEޕ Ks}R-TҤ懯}}??[cS@?=xa\tP/=ZmB^6;B_I8r&Sdq|8D*( nŖ9OOʸa,LVVAx7Z3HV޹'5HoE&0w(e8M’'Ģ<%eDs)~R}:~{i+y5XVGw'#!cHBdk1l|DK`d;x8u(ݞzQbu:;rfU𵹭ꍱ2 E',oZ1S*sùSC^jzx>^]aj O.(N V?}vLV/cFP{'Tl'ן:;֧߇v8H_+=qŃ?+kyzw<NJ̚d׃,%SE El aq wh9b&280 X?3eR)(H^*elH,q;P3kwmɔ(16qr%@".4N~ 8X%g,"[OYK$)sR('/d= 5([!N('oH\V5,#i$۩,ot;%@frq4Lf Cj/C%Vن)ȷ?) D.4RK&*U\"5sa =U+wCLubT pCۂVp;9 % [-ö`A)_S"<*8%,=l1Q eFPv3T)`5krI|s#iCK궖/e\̈́˶k9YLI8_BڙyE8cj*LlhQ!AF#:6@+;Onh9{Pnq*ɶXWȶI˶ Amqj*ڮX1fL9z(%~0JAty쎰#y.V8H4cU-ݺdS?l)xuikSrX D@ `)UGm)at}Gظ 3D ge( ar SIcZsWg O}^e.}ǹ-MvNinzNh7_)ƒj1_3u%gJȕb`w}ȇlY< od2a1<=BdY#) o[:bVH-6bdU]Z5}hO֖t`8yϊt24o&H*W9 gr !q<$]Li\fJ"S]j t>vJ*IXWr-&#"g1U_}ӡ^.c!WwZ^d"ZL7x%!ʫˏ&͆N2̊输v/&,>؋9)ue/e>,}C'q0V} -V S&:L2:3=s~@7WEJ՛io}FGG) ]%R("G1<:J(EL-mp0kJ͝/?URD)-X QRY aˏ[)moxYyok=v[$?8H!d'ձĖʘ?ʗY:]u =ots3/(޽a7uM=U  0X'AIy/wZ *m;ե^/G2898 >aOL}ZdwgT]7-xJYfY"G."V$fZE-&<;Q!z"ѓ'YųoݴUKm>]jeջV%2\߼ @akÕ;En 1(!\3"ȕ]H c$_+f:\2=t7@yּ/eP_ p|lӅr.dGM Q<4kmSD+K>ѝ=0#hOMOĨT]93A}ars֮9sYQە 헰Ţe-/hATE ;0{4jWK5a|c?d?>^n>jĝqWYQ1kcHʠvZm;[+,G:K!QٔY NRDtxqt$ƸGRg$=s> z.EGK.9ߢ-2 RǨXm]f |TC8e9P>;Cא?K=(Bo88e֗{ CJ/Kݥ0eb2-\e˴lp^6hLF9H$d=fDHʵDmt(S0SL8$wxBlHO&+}Qd/ߙS%Kp_ YH)TUjd=,DRMC$<,vЙ%)mّR ;V&{m%QVW, !"22R !*@^ -Ø*K%EA:ZŨb{mnGaT $,Dha$.+FJu6Q]&|8ylBka໻K}¢k30 :ya|6-)6g4$sOX{;>VT.ޱ39.fLH sЭVOONSfa4;ѿu6tĮ L8}4^^U03"s/zv=>/o5_z2\̠?,aI>}>}:zW |}i0X|zz;Tsz0 j~ K pik=h}\n0NjhpZ?6.np[[_ߝ܌U/Fc%8]Xӹs\̳k?*}_W;zg&0ňZ@Sgp1mSc՗K@8n|{&׃Qc0;|v{J֓77Áxh|ҧ}V.|zlO:YW;>Hl1(֐3XW7nGsq[|;Ro6L^4\J?Wln -͓ug孑pm&ol<4.FLJqtS.V`TiB  sKn89>laV(`fIcp@0Q2ˆ`SNm2Zyf,bG&Hi6w;J_R{Eu\5 펍WƊc@K@FP`ۀq$ Q3(kWC)鏆gf+ScEkFbj/{挩b^+ Za MmTU= hȦm H+߫YZRIN0ٖ },ڵsf+! {\ JRl[kΎPVN v{#H"HG ""E1Ai(` єGT-GЖڞE6%/ J.s]s@XrEI*= G @"8p) jI#|t2=I>`J֥jG #eAw+i).9=BnDʂfAhUr#tdLcJ A (3I8'`RQ Xq?\R( [S5ݽ*6I2:O#wUhEĴ`ۨ^e#r0-F<V͢ 58VnOl0fF`bYE`p0ޏ1ɉq0T&PIG \Y |VƬ Nk FTE${du`,퇋dpN"Q:-ب6Vw:+ӾgdU&w`8be&),<{j걊0,"ʒ=YFOhӞ|FYFhW {gĂQӡ(VmBHbHG8F0"!xDxK8w2! *P,(Vj|/2Ib"W3Tr`5#F4X\_D 8qa0WRNmElbP0,,˸7cLa cuW6a _92(dLE3 ]'\5Q7fTE)l m+O.8YģlKUR/jcͰ;OkrZ~^mٗHsP'#,p*8 zǍ-Wt;ivbI]|Q$v Rq B qEnhE%D; dxߡk7JrL4@#JR`u7qqP('8h)n$Cb>7*2"E &1idFSp6cPтGw1Erb[_IOņk# `J|0AYL5u8J PR}- >3( T3,a6 X^(+ezj%\Z &gs\\ ZgbŨ2%6֔D.W)`J"dLd)GcS_Jԩ&gۨT֤[YŬH2 /k^2RYXi乭kg:~jj\~}|Czy` Ţpc>ޫuC`TTQ#vq۫&ٜ"/nod: rÄLg̗7 T:\4ԙÌ'l9/iXS\90m.Ԧ6&t0.7sN`uH>) 9]@JVRځC`Y"B ס=p nќ3.ئR(Y1r6[{3! FX{Nw AwkuỴf6 t;DAi$@,ڙnjS w*h|yH4yJ8Oy@hW5t麴@'K'B NV L.:W7$9`+P yKl7a"RVҳ J5;ěl;3Fzx<cēT1w &qBÀAuSM PP]bMnd]YoIr+^6Pͼ0i^{y"O$HJ#b#d̪HdWWƕ_Dٝ=}k7= vty< ȃ,}Lm[(dHSXzOm94@.ɶqAe>'$bvwnyٽ=~t4u*5E= /OAt0iA4Ey@{Gywmy8!Û}KKOL쩆Li)o`*=igþT"OǯRDq[&ؤ?`QE K1ZKy` r@9XdBnFN%Dij hqCl D Et =*-bA&@KM;[Rwyo-M|ᕹS_9WJiYxpB-:s$qjõmc UM f4s J}x G#hB4YRk~!4I-g9k,(Ws; _c)ZVgzzD Yӻ*Mx~[.aL3]O9z0᭷=k->Y6iu-?Bgr m&RY|oz0B)&d퍒O /o96T[HηkB?hO M'fs|oВ%1V-r&])m`nn?Tpm7[ jh?c8~}/mIÛtjٵNcʸH6ŸSfbg$vO)gpz,b o #5=9& Ab!aN*UnƂ '宨Jo;;a9S_xSQAJUJ& Y6a$XOx!=e _(Bq葉Hcפ!5|{Pзp^AJ$^/E9jB9RE.#:}2j0^6Ada>r`gr/y/ ˔A [Z[ l BRIʶ(>k^9a-ocmG[^arpƽ힞]d{Q5&ܵ59;}ugs>\_;cJGbd8{>uBX|F:@K"چƎ=A[kqI谾YɄa'({=u'j ؐ=. `ȍ] PM AqF`p"1Z P;@QjA Pz]J}8Ds<O  W(hR%a/F4Eث?TjK%,U5UEO{ːBX$s굵ӤfaZ%Q3<7浊*C؃ϽTS RܶBxA :b !Bx y&ZbSlwq>nvZ$PYH1z@+e47p-Z5FMR"*|2BJ ̷.t|`RF|#X 7цML$Wdd0_Yc`7t20Q Nfėۿ0!(1) ƕsULXS'zXkߡʋsa84V9,ySSi,B  ^"h-ޯ=_gY{lXͳ,uaE'B[Pzn+yAkE1kaicjLsUE⩳!iVDsQ`ʚ%\lX X XmL1}k!)VVjs FjCXu>" Vmi`0Wm$ѶRp$ kzeܙ iGQ;n^NfLx/6n}< 6~2%WDmO>7 `CDi`eu bVXr1jD‚? BKpX</uJ`" +n>'պ9ӆ]^ H!hD˾pgČjoIvky<~.!?E ov/EKT>BH$<ӳr;cI(Ӕ9D8U"A8& A$GL&TJ?M_)wv;bb:"/?dH{oH *CZ-0<*e-f1ʈc!FM 3)/9:gSV@CdLB%%-"2kI %& mTxSTKJiy B^P6J}bn2+ R\:R_h6~ y&ZdSR>?n"aTY cbviS6w+f4;vB^6l*;SJ O΄/ c:uUR?yz1rpT"QYKH%L\3!UŷxW} xʽsn[3dž)D54ςh*@?[kSd旭v~¦ Wǯ]@/L>ٸTljz~)bzi ?F5Ř_$0++k E&~*H):A3L3}Oo j9B~/xc~wE"]v۔M 5:NmTas}ݏtjg ~P'iB u&I}B_1E%RFF!é/|-kDm\~˖ar@t?[ 謇-jS'OHOϮv2JHQɅ+ iq|iU(h9~Ql$aФBttod{d؃xk6#y=' шdD1s%(肓UQR[znw-;K7-DPN^5g6dT2C4Hr Ȑ;ΙIALF)륳NH ;!I6bc e(Էp8Rlé,eҲO9+L['Bi-w2k8TUI 6 .n"g ']d;3SQ: fyg֨#V&j̒ȣ < PUXEāSe(Za+=UQ.!B@*%IMaΥz`PGR,HևjE*tQI. lKea,vFXΆ-3QcX 7"G^y7^ zRN}ی:=DNkThwB^C8xMG eѡp^MGEE݊n:2pmδ)h2vӑ9o@.>lR$?yuey|>^^WZU+#ԎV]KqpNji YMi~z%눧* AT6bZ1&F4|Ig75HPu-3p1|ܝ ;~_uŘhQKf-RJԺč"R3+JJx$eP;2M41- l91U_qVBHbbĆ+%5)uU3Т<ԷcTR53 -*x5d 8KΓ)#%Hj%/_G, hh*lGUIRkʨ*QjeTh=Fhd5'\%Ϸ`X/ֵB9kʽRL/}LXZ4_ǍM ţGRK-f )bukRVZ !y<}sUF!ᯎw)Wv]y]Jɧ砞3gJ/8fbA.jD2Zo4~MWI?R{O*q |[$UȔql:xmVJ伽38iɾ#k5ej>7UݛV ɉ~WA 규(.\k2_U!%bƛUe)s&X]o[\%jUI2dt'NFW/e) S0vj؈sQ(&ÆH('~.e[:PQ)[IȲԇrPNXBc3W"h SMQ*/Qiz(ơeX{IQ2 )ZbA)O'׶DSJQ9[6%f/iӜq^9;%N fM甎S5'kB5:yzΪ9=E GXetw2E6+ mUy BlWGۿY"Wl^rx)w:͇)Fz{q8 pH&047ZLYhf,fϵۼYIv Zk,k%F$vc[!;((Tq~tzÄB)w 2tqATjĦhG _`LoR!-E:y XGت ]r@edN|%azO%YS1=ol X·zl3 ͒tWQP1!BeSDPVݧ7DTے00 M:h,5w޶.oM)k7W/;[l.3VDf:1)U=g|DAJ vT]ugUꃚ.H$3T: >_ؽc&wQ@yu6fjK}v3g?yu9>gf)5΀&$ԒŜ<{|l/4 `9-8p0L)Cr8SXiasըd刿)|3B BD&4!q!4IYeeށ/{O hʜZ,N (,H +BݦF́DD%-<·Q#I/!1~Z&Qs U_O3w?ƨ!עOSWO=žԘ %#<*{S/5ȢgZVEˬI/FMUѬX=>=3!t֏nj1|~9O5OY0eW 9jVT ˧zSc"im_݊#-5fCyb}{2n+~}{2c'7Mwr![|NV(Q/qTޢ®۫3k?w7&fk;0Mn}ɍ J BoIᨕ=Qm+:l%N[׳&Ffl̦7%i2%/wYds:ՎkV6$6*?hƦV̇nG`aI"S*ɦv1Egɖ@Wꦽڭx-nW5*thI\Dϗ)}~l_ Ot:IiC't|nҷq${X-KeaMO9x'mfBHgy7oQKuۻ#ϯ5ҊvUe"J y j3iM~']$S4xoYk!ԷLZ|4vL&ub4S}\ݜ Nl3MejY1Dҷp@آTbsO5d(.jq~Fۋ- aM 5=ռޗtlOjfD߬ذׂ`æFm5QZ A ~$\IUrpD`Fj?!#Hq4p8h k 199*<>rn48P``AR,F1m,ߒh|P :GBHށb'|2,]){ّ6+O:}n9.( Hi@J#l`y:}LO.ٽVte}wa%m.쉤b 2{f{cQ1ɝNwGP;C_VlINwG} 5h`{hD{ca[=RQ `WD+q}7V%j~{ xe^/ ig.'E]U|^cѻwsYwýwYwƞػ,뻠Ge JT˅KcL}!gƮc(gd Pз@ `4պ,U*"~X0] ,4XҪ͆*~~13돫 3!gyZ_?=qϕin/~)/(T*"~}偫Kh߾>zx"C$[# >P9c6_]MbZOo`uz{Ywruwk}U+Bt篑(W!z'}V]s|狟275w$ky nW `k`JuWz,~+XjYi1.|b勂]ߨZ`I2G3uwpN±LIg6FwU80a(1*%Mjq 9,vc^uz c}X|ѯ}LUO!-f/Q\i}0rxFy4Ks2w.4=yt{5;b,Y^LC4pEH34Htu1_0%(]/}LbJ4Sp:~yH@{IIA A#GrP*j gCa"ї8Ys}pcU={=ě͘H/f3;jf3q0m g>Z! yɎyw {-Z&ܚy߅ +?;y;?>5pMLH/0Ǽ]90]9Jƨ!A?,(gHcvLdBLl+5Q_`ã$$~KVNG;t,[5%wXKҫ{/ ]zNL5=Nٳltn 1e%g?FϏfzyz0ZOϖ jv2ťX3( (Ϳ_tk`QB#}^[0\>׻ہt ӑYĖ7 GBm[ms`x4VCCIgܕ ~p7fuYkpzN t8@ M %6%!@h%ڃ0@֔8i;78&G#)ԯg3ٹƙr*@ pz L NI|<1%dʣ4 4K ^ Znoj{#@׈ Apչ  ZnR̻BCu?W-eBʌg./sXFdnQ>ӛ4T- jjO3o{rOZO6p*ಌc%-5֩Ns☀g 9.T`pNJ6Zgiap}k-zMw ʆ+w+dtDk7lM(85[KEI0b_>L!K2rΙ`LDAaE\S1+em8ce2xoe&`P33\eAl[D|wrBA0V U9L^aPgJS̙U6 "-*m^X:zO4M((%IQ؝tڢ[+ xb-, %B? 9lz)VP s<`Q9[P00X7geH0"vʕnV1ҾbJ o;t;!b;5!Z[+ {*A<̂ pTGTH%s$R[LA #۰Z0o&#U'JaѮJb\()M&+t5;5ZC[+uNaQ`MU,,1aY^(M@Iga]A$&qn2 Ph;z K^?J*זX'ˆPepj3}Z/W'~/t~r;|@MfUPOhWvk2c%\cA6}_~Y~zxͮsm|:?CR;} oy1T=($(ʝi}RIO%JaUe%~H\D[T-J#z^ST߉jkIsFDp0vs hz]s h9"1pǕfٻ6lWPf@F'/ DۥL }N5)Ůf5 R}וּƎd5F&7 ;Wp<魴L} up}&ӵ&(qwКRi{m=k 1 7䀰Asm>68Ϛ&ͻ'}|?=B^߹ogywgaruO. C÷{'_S6&`D{c6{{1.)4v8C6tFϯ+\PGWxp?Zkӻ[+_,T V`|a Z5.jm*Ֆf^">FtU `8 /9K \EbZG ܃]M@DB֌lʛe)"]I9|9mLeB,AN]6elB,_rY!vh %{_k-z-;n2ȁN3xϚKW1ޭ 8D0rsEfV*>waҹiVhw!"L[,Z1%7ӹx+@I7ԓ]N^C4n޴r=@{Y!o1QFx"h+8^ *ΥK0v}@3JYRvTKщ(VA^Yf (4+F"[ _h f {q@lRRĝgIvNZjT6>35!LlF&EӾT9)}Fu¯\BEl3ECw+At ʻ"b{SŌf{r-&kNIQ[jXjR(?Tn鈤s(J p^AV M'UA[Tx$T:Ik*AHJm4rbSoP<QOGE˜QK͌=B㟔(}2q.hvhMl-o͌] !Z){IJegn;fl{}1ޭ 8D0j3E&^xR@m.Ōz.r-_ P|TkAdG/* X╢P%L01;«tDƒ-ݷ52hNnzףi샽txzC7v|v?ݎ/iL։dzg0)33wn c9#t¼]>p21ḄC>|sJk 3vDHC90b@JzVT)xT<ق43!tjHw '*}FwiNҩ[KO7L@ȁCS}ݴދF2ȁN3xݕCSC݊n] !Z){q.x7A w9U(>3ZMEgmϻ3=BsҙGD+#WDd}M޸j86^ُ{"5J҃z>_LWHv,nt28սwޣWŗ;baQ%;UI¦G4ѧA}nL%d,,QI#aٵ[:L,QP;".lc4@--rl~f.Xݩʹk#6cJmfL,7v'twlB-Bl%jre.Jc)]T,]I 6=.rǝ.#J'*zoRNTĸ("X8qk%dDǒNUwsuDž_>5,ʃRjռ][<4U~(Yڹ!v]8f.@NwnRze}X]D?TR{mYEn&>PӣяCSret)>`t+>Y9ݛo:8fd7kBӿ 6 ooԓ~mY|ř;dHvNß_0t}5">'T~Sw]_WфS t1t:ĢwxObQBGfiGYz;-$i " SB0X_皀{4]URQcUhXKԗO0w/4_d&n}ruUă/ϳܤNf./o%?L˫%,[ FOS@8nAqfUAr(>Xmҫ\/oBP]Z"GQOvx*bd^(g١OX-}\`ⶥZ7 #p0{0Oa)Efڀ"Z5N\KhMm$ FS)?k(sP4 a 漓upUP GDFVp}Zit8¹y;>|, 0>:T(k'4(.}Hr^ꈁGATZ{A;gRZBѱN& &OTR3BӅ1R%+CU+je̳@D0@e(3Dӛ|Xfp4\)c+VK BR'1\k)DBvx1 VΚ4hl2]0ecG#\.ݔDYvB—~1u$-`'@ZD۠K(`t B3[Qw'ec!"PyH\sC nO΢}䎤xa5ӤRg \S(N_F &0[kW1Jl\ Y8*ߖ A4k7< AeOm jfP% Ή 짒JBZszlBr`F\r&|a#L v;o-.2h M>S6 $ lL$D쨐g$mNB3S\<2qE{ҝѰ?opFā[6'(%誫3"D[哙fbj5UBmϯNl>勞Jh-w-P]8'[" @$r4G4`aWΛtubtE",b3n:Oι[6E1el >_й>vgg[8ʙ]܃3ɜؚ~o]U#D 琙Yˊ=jNMec,@i@z1- q^m'jxX_lڢD|Z4% KҘX19\F`Pg3]̀c}VfrTlb%h=.i#~>%ZsG]&m}q6dAp{0AӹXM'j׎_G)'KUCC ]cHt4S;qaDqʩe:3Zܯl48|v~}lڻ.;c4azFIJV|Q;;˯P:^b4]gDt 55kѶG⑽@(T2JJ@T:>5@^ PeiH gLgե05p3:4q]d G/qkq}pNy|sqEl+;+Oh+ei z0xgdaVYA6(l\9?ݹ};4ew ;VPѧviCwc*ܫךsCjfo:zs7{;[}qU~DV3cUdr+ _~ Qwdʺ0V UO4ˆ yeNc8D v!q7mzTN{ ZeV]>q0E-̞H.SlnBsZҢZmvm;dP)$q9pHtY9ڈČíK'BCtMt(p E<\,پ3-]@8j:g9P1+Mc=Zkm"ID ?ťF\yc5#?:֚PC5򲶦OHTYJd!cNC _Bj+SNKJj &ˉňJFg9D$~by蔻*Lp4laa4맬'SD\|(^l\GMUQe?Uh%)bO}?<&p.pbMP( deU:ur5&"Vl4"GM${:af )'V0)STDj_`{8r?uy(ro#|ާ&{} X8|奲b|b!it𭌃CA擰t[2r/w!;#Uh7Я+Mu’  E4I̯`R1#:hbN?OC[dBZ6$$cM)bPGtRĺ0"y4Lh]ֆ|""S,jzk<떊AI}Gv("GnɄֵnmH7.I2ٯq[7PzB떊AI}Gvuk`GnɄֵnmH7.I25^<׽-α9uKŠ褾u;D4:7-к֭ Eto;!U]zT]^p=4OʹW^b&ޔa~2zY2"k&Oz 7!ȃڀcT5zz{DI[ ʍ'%dff!uLveeV9t;oC8qׁa0 ,1 /8b`Rb\+#iH4?#5P롯BbiFS_ Clm <0붻'| 0BK(^u0˂\|g~!|k>W83,̬ǀw3'S JJy }~3[XiFYmM!p?"~U_oz՛ϫao?O/aW?ޡ6nB.ݼ7}W?/j.]y(Mu_+|LCoW?[:[$s9L0H[$%p)Hѝ<|4!\tAD{RIN d(s 3 wZI@`{K&B3? uk ja`Jɪ~jt;?^_&C܎{"UDh$6/ۥZrA)WmZy̫P-Wq45ó{;?iDOp3L_׻ooBz-uDGjNܤK! `E@5պ]ڴ&]7eYPNhar" `dgT9Mg +-bpafdb,cLjQOK }[):fQWy\sq$/ő8OƑ\3g0bͯS.{NIk:>KX_gZ1 /Q`  `1v%`T9DiE92wwťr;Q4?(T|3m$W;OVH =+SLu]r e["໤0Fc9x׊"+5^* a׶H 8a4udKls`꨻ԗ-4dc=\% uR{W%E+v"Vv4a繃oS1W ;A4RjV ϸ~Rҳ|+$ܽf$gVDWU+i"2Bjf{Ѓ>&yUn+G)Ӗʂ;9v2F#yF }⌬߽/_`&VpgfЧ14# 맭NwSIrg)8YI!kWlL3I F99xNpA Ris'R ,/Prr 0]ȱM/r$|9qB(QlZZl( g%jɸR3B,lD6 >:RzP~h!ĘXaIAN! / L %zq )/: OU~ =%Q>Z .U24|xcfФ6;HX[X0ixUXxաs|.AB0p.^X KIڳK~= mqMQ&ɢFŅ^ur s3W92#R1\)Αe  'G5 8ʂRiADLND D GPm>f05l68EaKf ]aVjº ,7Gq̧>ϲ4giG @aaPu·0OfX3SE2W8-#ǁ} Xؽss^@`2EGRA; [(I;đ5fHt"˻q'㱢 ʡ46s7fP) @Yd zɱq4Q۴HN7E'֜jFzd"\n5>PJcibASOG:2Υ--ؔ5J)Sx^|]Y̸\4*g:#3,$Tg:v>>a܊)I2xeL(`%5K`+f0'Jb'7cf3A/` K>{a%%-ٔDnݒZY,VPnU4VqV=\ 6i(,ٖIhPQk:*ӂ-i)8:mӃ`FDK/}Ks_4f~e(creʖon7#%xlb ݨBKए0 ^pRKsKMƣWL1OGWe)[$Qhܞ =xydE> P0:;@Ӆ ^n_;M-q>LaunvHg:`Lj%bW~2%q<ۊbyv*49ڳBVۓOpcmZ(o{BKM(UѶZd~d$؞ 𣁕PVD``B\GO$Dy^YLUᅭtjn%OPF( r4+q9Mh?Ze~cL=3%;F;gB`NNbkP7%6hv?''j_ &GP%TGB&GwQc晞[Sɕ bCu8y{{py=diQ%K2Y(:VyB1Tߍa_bi&| 6. _ du1[,驒2% nvJgQ#:[<7W {C<_W b<O-"@Ba\$?+š:gΘOD@NQ.t͐ӱ܁esr3jGy([Uzs. nh$EkxzԷI%#|=>ZC`J^5ڑ'EG_8_jjB@FA`.?Ζ+X{3?1A::Vܪ^A!Se t.1c$_6LA6JPv&B ¨W ܯU4J:P4տ,{K>S zv_@(_gqWf?;#%JரjzuA-9=a<ӕ,aB%DOSd=b r3Ѳ+g{ B{}MvǵNW4h7^)BH}l^u#ѷjO8;q\Hg atZj4 S1yQzŊ3Rȷ ޤ,LV[+{+W_[f<|._;9̟Rzj_(1LoD0(Uu_jj{r(@q:"--5N6wO :y&꺷qԏKpAp({8؎3Cv$|8yS0  JܮeI q_*sGXҡka/W&Cܩi@ D"P@f=^8GPs7{lUe<[SȒo㔰'dwk1 =%3VRGlG1Xo߶1 .Ifz:^^3 &LGyl{H>w>މL c?Y=V!~^^< Of(fQ闷DzgvNj ]" ̓7@ ;4<ݬQN!bJ1>|$(m XѤx4U+Cx\ 35T8W`9}2du^4mL`0ؓ+z-8ܿz!WrIs|r$Y$\g=eRm8TY4y( l݊ӠrZ4>l3X/{{'5GEfwO>r:ϕVN݅Q_wrmj7͐,S f> `eo>XNkJ?ٿs;d!LY{B yqluBRlw{M2+ENjĠ*d$Yc{=_ (vXiWS 0`"H3 c kz*FT %QbOE{⤖ig5]TFR*ҁ$I)&T[( ZnyUrZvD.$C<6pg 1:z#v7[ ?>ڲ)v9.}x؋{#]g=q㾛UK(XHb hS<ٷIܾ,)8El E"HML7a&P b6"vWsA{z4ܙx!d{q s 3΃ٹUˊ$9HPq/"@P0."أ5_{ е>\;J7'rE0h}#2{⣩iiLH'I݄ C :gL<#fQTʣ8gZ $E/ !5~A-۴Z;VU6ѧzв}Hz<6%VkJG=θ`ee+_u+ZC'$!纂P,ڵaIi2}{Bgi9=jos;I {޵9$8WkvI u 4Uܱr(i .Qt2.b=Lޯ zS{ڟ'[{b(=iNOoQ: '}cN8K7zWe{L KJ(S٠H#@xK/r]˩҃<څW_&vB Iqh5$+'eCKy;51J.yJ;{S^iKe ~xxgwN+16J6) \-NkvNO|zK79]} Xns(υ]oX_ŋ봷\'R:VIb)?p/q=rW e({`PKoY4f<7[yW "x[w7)BԃX72,-qȗ=oWt[YvC4 TŋQU,Z͞i{RN/Xb˞> ;qS?vsi}Er!KʇY]3 Ѵ K<0βHilY$SⓦuA}`C q2T2J0 xB0eq614VJ.' J[fa: :̄!'uM1@L"LU'Hzf<Fʎ%%`!0 g|jTs3)TJJ%$0L0PfL$diƳZi*c"P2g10s5}=Av5eқXW72/S$ۈp;?7B? k4`᠛ŒIl?6*2޻Z3C˰z8Imrigj1K톹<n(5+o^zӃ[ oܲTߒT |Cu9$ӜLI,3œjfx*)ha1^rXG_WYSUix(Vi5Un(01ͭ%WrVG_9=M3@g{7~=m6٧wr斠 lfW`x"U=Pܗ,̪}/"`Os\Nӟ1+.%kGoͲ0򅋨LU2kY7gR֙$hϞ&L^'!v*p5kЂ>,P߆W@.V-;@":$Hh!S\P1!Lt*bN(]f\B}E CR<^xTGe=T^hUQ^X%MPoVI*b?|7-҃(FPz֛¤dkzL\wv7ߪ/|֣@"H-R֔4vw*a)Ϣdk-CIa f =TꖤbUT].tݑ̉]ҚiJ5T/:l{! Ythqh!T=݈.޿&Z& 1 w>NgI<߀'"cVRYg3y72N<2ux1:'<,0I&9-񩚓hW99XsjfZIXED{Hok0"B_IJ8-P5!jO .g]@2N{ε`XuuWH'HEbZT#Q3Pu(%G8:8;4t"Q'v5 %FM6 cd{XyTqAmȡ~X `Kb+M¦ұ!҈ვB5iHڔ !F1xH`n MxB;H.fj a*04s. kn$.4tM *=P 84,êьu Ȱy"q5g5CXtհBB#3DU[T-JU% PFHAf. Ν@Ƙ$Fީ\j"Hq7Y/bwƽ.Co/qy4cQjaR2AOӄkDN3g1UfJtvZ_"Wo?]?P> a/470Gt2B$чޜ|\ |3O7xvU#7G|zxwͭ3MoDѱ@c2{W+6zDsy =K<|_6 ke^*`,ook/[dye gj/<'8"u?oySu?WHYus18%lhI(D $竡ϭF8u:_CdX&ޭkĻN{p|wxwAd \rp$Jh{ X[Ar{ɲ^yRǖ:C9?eܫwCwLQo wRxJ~FY0S'euķ쵠["oZ1h@ kucX& ΜV6Р6E0 i璣N Ī9e^|(tz V"^s"aSySii:dJ\ ut [..K5ڄ<4D䆃,V%Yk# HYdy5wJ87DR^ _ VR咛,~'#m_JȔ K5MIdc*Ir[lQ׍'T,tL%3U4 =kBPZMxW󙵪EgEaWqr\̝1N D삣X"ZCP8JJ6xT`x<ƽW\Q=`NZQ~xsω"h朗 d¡:Zq8{.l!Ȓ% +Z=tڹs*b'̧&](76*i>>#)"~?ď|>nѫW.>8 m鯽[?~ !NczfKf+pt{3_;Wܒpv9ZQ6~eFԇ*l#Suuagn$zN^a,ZxhrVmmm2͉qX3sG%H$9|:Xv9x'8aoY'6O.V*.X/3#Iu=Y7N9Qᚰ * >eڀG:.0!g>8XԱ<5`g<&,Z F-4HRq`̪hPv USi0'xWNMD5sc}G]Jkyku+=zFnA-Yt珰dj^'i|cJMUWHY$d Lrm1JT"Ą8<2SDdY&xCXj L./FW윭ybt]1`|둍L1_Ř-?ɼ?^)"[?_)],ˉ|_n|` ,)WE}²VNG]okW7I햼c7ٲKƿ_NIZ^]ņi^fK@=usxZ66c6uƒr&Dד{`l47rH,`j4K~_.ԜЪ59Xq yk -ya9]BB<B7p%1R#?S*Av?2JѦ?~L#% t@",Q(:]L gdbi}e8Y&x\w9Zw9)\'i4NxP.sTa}'^D!dխ).kݞ1N=!;;,gDBC/;<mᧉc-%i\:݆!(05)F]?#|zD 70D!(f8 kKn/nÖ8TM7=iz% }6~]28sGHhMs)]1ʡ0 8ehh8kxA)G܉aٟ{%j5q*{>I퍕+7VjoR[ͷtjgk HcD$id J"4ΰ`f:[o;b-)*`?bbќ:&<5ej'UinbƱ134Syf8a[Iu8键g姠+~N2m-#^ U4+TWNv7\%op*(4GI 6ol,Ss*j=]q3%<Ά0&LZ[^z"t`Xv~'⛋f<1#f0KSvS˜gtFhwbIKBbLyJ^YJPrO!cN6el8{$ͻ^(IBEҊ]$-.wߕ_Yeҋ\a|=mpG)Ձf띧,><܇bhݹvSɡL +r;u;xf`q{%Oݪ/:ZMcmެk|T>2U냕BoL<k|Hׁgc^?DI䬣]ƣ6Ǯ=>NP||wxwR-pz(]=tF8=Sko{+IZrn&}[#!HA勉}\lvOnFm{@C|چ WLj(e0ze@FyklO%G䷱o uk|]ُv:]unl}ݚA'+o>m-um[j͹lcn{ZuF i4`!{}wr4^+u-n]/wOD%n˨~Q˨]U$;Nȍ~amA(Iز4.Vk4]YoH+_۳> a vzP5ҳB^,G:\`FRDIM,\D1#⋌fF cgAJI_ H0iJ«6H#)/ǀBqa:HUv|4,YҖ~=8Tq!/u8yÉ 83kN$qXMѹw܄Q8oD`z+t:t{i\~<(0/_w-s<\#Yb<ljrVd:da!:<Ԃo ER UstiRo,B*(s* I)Pz71qc饓Molc(y餉E8Vp|FH*K d`mkEYe&F|(N;{ΕFLm4:8Yu2Z3:75!\~LӀp&)?]H?L`,fX\Z4Sګ.yxCڽiPblOuj1#~+i4f ^%)Xr'J(E"s{ר2\am JF5bWsP Ғ1Gl9 {lgI Cpac<*ݠGVJɔv fFVs )mO7Z&coB@:Axmg%Ķm ~'j_1~OnU߻ɱ`E#]~NfRi֕Füj+l|1M2]3)PW O9͎0;|. ĺ' Vk;ny2v '}M聲="x\S!$>e7]MS uӳ'a Nnq{2j0)[dIR?prvJ>}NOwpi+>VK !Jq_9I%u&XX=B(#0IѶ,z/KJm/u@̒A#[h%D; $ D3g=pȊX x,1 (&)׌qʑΰphMXXK d͸tы5 ^<ll45ͮvo<8GХv{XE6T;U/Nϲl$1*!T@dtBV4WdtZ5:QƬ" W[Ϥ&ˋ.s3d;Qoz =fX9{Y7S©K4u4aƀQoKr v([dA#@SrIfN{(P;dp⩂3N/)[ 'vf+꬧hZ^%J6oyWs]ٽL$|:K—O9;&~8&Y΄=G{{bsҴj_*d#;EDvoE0O'9 ^Y|w0zݽKٰL)րaKw4λ-6.Y,lp۲޾Q Y΄km$Y G:Lju]lq"3oo$?~ovq"6{ۇҮKzW zW'xn:~)xmT]04_}pxDV5ఖkZ%MqmۖuA1Vh!dbO֚}Z,S aXB~ԙ84ߙK5tw7QelNrpd 4+DŽ/H5Pi0R%!Fz3̟Dj,4fy*L}y#) <QAgD Ĝ(f$xSrY ;?}XQjB5SU5B\c:ׄ{4;1VKz.b Wbdry6s2 q Sx>L-K,W T7 D7<2toa+tv]ɽH0g4 )|NoofH8ow(Vn$F+ivY?LuUNq(ܬEZU3?9u.Ǔy(g8aȸ=6z{͊YGXq%8K5a? L]:5F Z ȌU| >nM-p'֋F ʔďc*~JoQ>_e OQ}.8Zɬ ake Ƒ4cLqZݸ{׫ݢ`&zоx1҄tPgdJCSOmd/a2L^ &y@Uub*FXlL=O D 2 q .)cռHsokrйB'7E=q0 Q$i!a DC`5'TtR2F!zKTNQ+ׅ݈{ V;I0xTB&ILJ(Dp{VeJ?a3:PWߓilbIz';\u2zu]#c56rL36t2n5 W!BxyFb'`yȯ/!$(,_*~WۺI^g^cx`yE7t;>f>jg{)%%H{m̥d;l2OvqA\qy|;PGL,t*BΫ H\YÏ&Hsa@Ԩh(J8!;%G :L 31;yURMR,*a5 'k%ϐwpPWV5R妃oʤO)񞴽V_ @hMVdBdYElJJ[æ%4hmF;@%7> m⍈x-i9\ҍ$IrnJJSEMgyE >H/AG)HS`:ά!(4FZy!֥$h™F f= wP"VZ#PQg5vXJcL;P(ZdytX~*zMkS-|P9L02!icB 3fLDX(*j)0庸 $HGm Sm ӗ |sJV{FUo3*UI۱` Rtv3%hR qg۷u[m#jɶ@h=k0! UǦta"bVQ~oA~0(/R? @R? E' R?ɔ&D NᖵAEdqՏWL$uʱ"X F0$ i0? 2$XB`G̷OԳLHi%pukL!h]ugsZShlka>UQTdh^fb tvʭz+xmXdT'غ B +s'R.r(SM##sIB:!6<Աl Bb$N,I/[93E\@p"j=9D*̄fȐY-ik\ kvYI18:F4)VG?{ƭU%@nҢ@w[tw?ItH$' F.hXC ! ?r-.b+1ābSi$ǔpP(^goB@NЎ`3X?mY`A%ՙ/8& 8kXRn9!-'¢rk 6\Lb F0*=E.eH1k ?h}6_ 1L Z[J2Va,7Ug0e0QaX͇DZn1!89)XV[Q^dw( Xov\Q揓{+[pKwKS_+uhn 䬂`bRܲ~-oodR!s&# ;y"%HaH*%oxR7ZpvEEa't*}iʖV1~26`3&lM;8 ̈́>53a+V6ɨRJ2 4hf? ƇǷ;j9`}[vJly=~*-zoUL?KnXӭ!+޸2/㲎[+X-*zVnq6 0#Fc{^ÊU>z_}ڊ.?.|݃7ήt)mP91F?L=9ܑg_oa~6ߌ~foI?VS=<.̻H8b~m%v΅up9f>d嵶ӏO}xfBg)Ykpxm`ӿ=7{tl~h_'$3"5SM=oמHms4M8DK;/# <`kV}!Q_|^,݇{ hONje`t/5 mmg%BB95="7O߉C*3ZXy&0?.u|޻h_ƛv,޵7/5(.8m-xyyAָ=Y)oehknj,m{;o5-ҍU5wc.|r,7xZe6v/#5vnyAwEɕ _ H⬴C`Uݳ/=L~,^Ѻ8V)W#mCqwzf|{SVOZ5Er>JnM֏đٺ8,!Am?׹; ܷ[:e8㢏5,o8-ӧoΥc» &n\l;bKhnBZU֚,ǀ]&8҇wOUz?,N.MBsD/47Hi0ݝ4|ӵr$0=Lmzr?x{(6kݴԲG вHz7QI)؃BIIRq-YK*4p w0 YwrhATK2):sNN?줏Q"$0߼HR&w;|,skv$[{~(x[7Tݶ{zRaӫAԒn5AJK8~({rXGMfy qr:4q}&LD+9&)!7>{[0m-hc2uϔ9-;3D(r̂4l0Z=B7J0bh8x-|cgkAV!8VWkA 6j 554BAU16ДyiUV4 P+.p ZDiV8Vdzhz )#U.y8+BAOף 퉣/MGOUv҇ߺՒ +=&=Oǐn<&5$aeQ5Ѣ%Y(xyP\8Q ,QY/G=@]ƕIX ) BZ?5 D%mp4=: E)Luj@6NS=.~qSbTNT}!0I,qg_p[W0S֭c}gi: .%6d`2rt9U79lxsqaٵ;1|U)RW"޼8r(ÌAXd-Ln23.1ÔU0:5L+~>z54Lߎ[/,jY.F 1[e>r@)gjs|*‚h{7vgpZIY9])8e߯}ˏy08Jx^|yA'#4NY&rq'xw3cL*,#")3#d-X idm Ǘןil,-_⯋^^<7 rRVιxiFo/~/ڠ /s ڂ6f61E`{K,}?^7ZQ_UȰc.L|umX__gTWn2q@=2Y,ZB7?\R Z9߳.#=oK=Rz+֪H*i̎jy=MNeJn(LS"$/tq) sZL3afGh8+Ey &W[>'L[c̹0/f84"m8t,ͅ0,G!,,Y0Ea:SB0[l !cA3kR -4D io,T*a a6iØ13ʝCDX$0-v`jBts.LT2-\х; ʑK63Fk Ď:_8%Vb0pA$CcJ XpX V#2LY]%C #؀>d`ʭ4# f%̨:)XT`S k(RĂ ya eT\1Ciۜ@`C.{VhkS cFo~D?_~yO?.7O/'ï{pX6Q|rL[چ?nbY~9wf+)_w(d<ѯ0Q8$SP2LrI(CX4+"}W%ϪG8sSq ^p ~#qދncS}ڞc1NU- )9LlC-@ `\}ntQ0!+W Fc*ż'rn3~s|s~Xł//|yq}ȼX =Yw<_^u(./-/|gˋ>bAbAbAbAbK $I"J RXI GwD`m3,0@Y2_u%./QP"bqƼX1/|yTˋ9_^u'2/G!8[^,+C w=ֲlʿWvX)OU&]G;;&G!eޔУ׊ VvWR/_WŪdP@_q~WUCm&ɍ{67mk` i'ڛ65 'ڛR65E %{̀hfq{͙R &`{-4koԼ׺ {Ϝ^mal&\SV*t9):pfXߋj׋]*0nbl]ɯF[Xlgh:۔r7o.o ΕUUV_y9[rNްRu+V/_s[7;k)Roݲ"ZhșN{6nR7wŠqw”ՙwn!4UK*_nybPkbl$od*ew;ZА3WQ/ܮwLѻbPGu|8ŻEgu%jzr*Sݾx7A(wŠqwefg?#w !g^:qc|ލIAm1n 2Eskzr*SjM}nR;&wĠI-^~u2jhșNa]D//uTǷS[<0by7s 5[ 9s)z[$Šl㤕i|QFݭLO&]А3WQ¨9ȻލV*nԚ=8E+Kw+Su[ 9s)BvzͅHy)=%pEvz-H9):%HwXK%uʗMA-Ai23Z!Rneʭ i q>ZR^ j  ˔rZlp[8Q B9\%>7)o jJ)ƜbA-A#!cքcN1栖@^YcN1^Y DS9ŘZjx1f-M1cj q=1fJ1cj Z>ݒs1aMc^SS9ŘZmw/ 愤s1Nb̢-ŘS9%H<3ZIb)o/,Pt?W1fŋ1C))ƜbA-23uX/Ř \C1C)ƜbA-AŘ׊cN1栖^ H1栖@Pl1fBx ŘZizR'f0S9ŘZkNc&cN1栖 c̤S,bDa43s1::U>1FėZ'7ңןf/?#M˞K({q3]t\Tm4LƟbйaK5x;YQ,80$w! 7/oLA U: aGTA:"!vXZj BX|E 2*bwa4,ՐE+qK4ey @b e?\DFE{IdwsfdJgLvRu f/ϭa۳TL ˜V D{0)0[Q:NO!\h< !$_2j#E`%H|Vx")L*$j4CR!L)%܂p9ކD\Fjdځ ~  RXq`耨$6PXK \~a1K$4OA \!. !х|*/${n>n&w@|{ηy PFL)3 BO]惙{f 42)Ͳms<04 D1!'!O(B<Ş$\ ͉"Z yR02\]-JB:LBApG"yZDE !h0E戱epk+AHihBȆo= lҽkh3\5,xgro>, lb{16ͷT˧[~>1hxfD>4߿"^oM(W?@`|&`gd(j_LgK}=م!Ztް7/h Oے ΐ\)HHl>|+wĘ(t7.O},KL=.il`ގDhd{i"HAG]Q<Ɉ#XX0c"Q8pJLS !MnnaҭA@G~ X <+qY||wJRLg,GUӖRf\V)TcpWqZA)L,0A6ImKt8 Eо}Dm48>3xMq߸p5U\=hՈ F  q+naJhp@0%Bچz[nԚO+lXDF;"Ҝ-02GS0S 7vNB჌za=xrx O>Λ&awd oe'׾i @G q]aUT9U乷_~gvJBɪIu@uNn4KEӃ/lW'z&9)ɨ|;%rwxK-QIS 4N|Ynόs`dȇO2J|9ZJc/G^,Ջ&$,UJ#vX%Z]DuCT$#n)љuwa~]/Jގ[9U>"tԻ R! C~}2p_q+9)+Tyama;,OE>j#oQ+<ƞڙ{0m?"] C7 C7 C7 C( ܤ7 Xx!tJ23hciB6[e`Ss͛P֫+7s2mU;mYl fnnNB>}ޛ;⎒YB: Bҝ<8CN)UFme|8ҩZO3A*A)mSLoDN !h.|^|CU DOb *neU}Ha(;`̅ց%% ̿8^7$yd%T7jNGi~kl\2?=vR m]塠F7L/irLT]ɩ|U<zRG.9zjj:1yl&طNs-N ~=`:jn!c0z8 ff?3нF:o}xbRw)Ua>;ũKPɷ^~&Lw} ztS%գ_ |c_}z|`ģ"=gTO"f6_ *Ckp l{/3>t"m9͙F̙T]0tZ: vo=Y{'佯0v"mjҪPG>O?@5 TXd:Q6ف>VqpeۻW h$Љ+ΦE++}ΦB|gkTT=1ۅjxŦĦĦĦ"wS}@1&" m%7MT!mItTX$- .Räs^~L#KQU;Z\"Ee}1.l v ROкy?g0:%{3 (scy;ȢZCXU(&B6\qm=HYR$Fx>k_,(ŷJjusLv fn|"߫;#`&OӦ\Q2-CYûΧ(̯ܿ$u^'w/fS8<8cRL2" Iv{[^0g2Kf%OiP9eL53٤WNq'&IeL QaKMUIV=^L dy_|z̘4Ud`)WM85⃂+^Rx< zMH{ONg`LXg.s5T/\X;Giphi,* Z uʼҢ>* :f?v1рz}ϤEԹ G )c[Z|%YE˔LʨlcJfJRD? h`X2*32+6CpؾL{پQhrB#3ݟ8?4:x$Ҫ!ɨ7!cܬ8S4QV ⴧg&Mv ǿ u8a߮ i[W/fnT;x}=i#X9z$c4xYpn"-!xd6) +F$:IlMP{.hry:d?I.QTQ~~\^}!-iKW4־j5#¶ɧ]ͱ+Le!yصWYnuS֮vb֮X##2L֚†(D;ČFi L{0N8tmU \9[:MιzgD 0An)Q S^Dy`T/?il 5=(nj5?|:5R /פq:F>>s5!D 5D _Yw,NQꕂ}!tu5K\2pNC7{2JRCs$! ͞΅ vT*uSn]1"zM4=Gg(@D[fZ;:^:PJIoq-д.6 ǜ[Ip< ˺tBY ZrJC~jevtMfG7ݢQ\|PRs#({콱Rĩ &xA܋+TՎ*PU;?9]=SkVW"K$*fވV~Lj*._~6`FxsUonQۼLyRWTWsE)QX}5w6,( @4sbhCQ!a5QiXoyЌ)T,r02K$VDP?)dBD`ph2wAjL TBг s{q>2D$/N(xO^|ؖ=3C-tHA[2#qq49|k^4‚]s-%U EaZyYScq&$f5/4бډj'FBy,F ).4/ Q6n4>N ii:T qjT [*AԏH>Gyl]BQBc 8P:.LK\B0JQrq'`X6q%Tx'!<:X-{&ĒT"q4Yl.6w:^(fw,Iϡ{3T?J~\jiMeYă3+T;$*Ϭbda;߾]T{ 8ď/xEJ9QMըTJ[0I>{a-=lFUX ؀ě0{LR)ʹqG̫c&Ǔ-Tzj^߆j5D÷Jju?1wF[~!]/n2K;" ^p\AOgM7:=p ^Kɞގ'S"O=NbۨDD{*{Iƕb$̰ɂWNB-`j> m4@dD$|4R+%.Sֆk8qyzhj]6TAWNy0Ć lQE^*_чz=w*& o>:񴷅K?&ՒKwT Q!hT,'wd>6埃#kp1wѬ9skh0/f멇usoosBwGn,,K0i,WW"v+Qv%ʓ7%z+,Sbyvs&Wm[di W[/fIF:Oڙ9kͯ:0㘧0& ;T"Ʊph5_Ʊ0Vyi+%^{ꬒG/P-Sh?S[Y$3jd*Q}ă$RBB௻ ԠQʃ_o.ƷhTV9-0w&vEpF(w1B]Pn1%J`#SG1!@R)ҀW Z9cQSj¤hGn>dbE}.Svx<퇏~П6;Wx77~5\Rd&%)z~eũS$WSbB_>}j #'_'eUR8N#GT$:I붚|5.j]仢&:c@cRRk(Ŗ/KK VZ1NDΈ@Mm`r\E⺦/z븭i pcs[Wm@;t:ʲoh!%4`^ăqL!%k!P˝ y Z&`"R:hLA[l4^3r-L6|3d?øY(y?պZ ITX/yIMTr0PHc b.May$$`ʔFj"'I-/fŕG.qP@ͨlui_I?|c<˜Q%ߑwoY<<~00 ""A$)NlF KY,W}{NK*DҜr=xpJS\m[ϟ:\Ap8F)N|8mal?N hO P3sd˞(aPnbb7M|z+!A< I̵8UCM KQG RG-X BŨY?” -Hm-B[IVBH) '"#iw/UXcy0|ě0{LR) y] $  X:Ʀ)d` =3~vi~8`o/]Ӟ_ { k`wu3菣ЈkwJZwP'oB kۛ>o./;L3NYq blwq4YxI@Tv/Po4_í|[&gO_ڶжRI gE >S0}B㚫9۳53k,D&O>(*s{fPnVӒS2v/yBRf`G 0fr|?K(yvR prϡnN>fB[} ɃկzZz( ӵ>_A5mi["û)_ArSZ &q{%ZkQ-D~eEqȃQ;57 Ԑo 4cz)M4@= ՌMXV$!RFlp{{74,8myHAћ|tn] {N.Wer0FtĻ;⌃Qv`Be|״1[3~||.LA v1m#2rBZkBTr0*|E0VK]ɺ^Sw=s1``5-XQ> <4-߈Uxh?3`}Vx,HWPڎUf~(l'~nߖޒ=Gs#ܠ&v[m^o׳$1u<:GyL4AF?f "BY{7(Q8]xxbR |Ʒ07|ƣ>l KTۗ&9bKeXFk^j)VZ odf9빗S>j W@.bRIrMP͝1*DJ *(ZX奪4GAh@AF ڷEj Sx _܂L*!'{A #HITGmŃ 6M(Ϙ=xeBĨ12-!ꃈyȮ<`x w&֜ꄠ[dgwW0+:C\YRNGGRdjy:)pBV%%\Ƈ^Ni޻6vV0+ڛ+LiQ4\(&"쐧!EG1}&N;bH"p~O43E{ k幎)-fc[g~SҼ{F ʢm&>F0Ѣy#g0nRHz%liŧ!j 2뀎TNN X,=#coqϜ0Fkf1#Gk_*:a}-!F2E:c.`A='63bCg+V^tqQUuQqdZrs1$0VA'F"Q4 O(Y4`y4,j2ob[Qڪ6gEr-p]yEԪi3F^)]qØBk!'2 ڼ1sVe˘p%Ĉ<\$*IHF%X?۠U;ڬ!g+{i _ڭN(4JSkH`qI=c A+OqnRn[F +Զ-!;CׁKg#2;zeA$ؙzZ[#K R9)8 -%lh +zZNOD%%ى4Dj,6ckԺ9pT9Z-J2 Q[SZ̫I|Rx. "$~U2Zy~:T9Kj)5-'aR a<#"GKzZa"(H8ҰA{J!_'[)R^?3 0x3t a `vGt@-#KWk+OK0sJ-iGh)@u ;@@SKE,3<'%)%CM<`IҺ 7-1[ B6D)q&&E`[CGy@=@OǼ'*HX'kIvܠK񓙌H-9aDU{ X7U<@ (Bi'‰*]④64IXE͈؁{( MӍD UJY8C^QFHx%omADG -)`UDd$nىdlL!b%eeGfX!%*C,4b,EԦ9ŝXmvØ[MvWgh+tO7kRɛ|Pq'1>FwqˠƃX!E/./.{`kQsK}XUBr>B:pbŒ]^SȜgW!> VVӵwu E@nֹ/Rsç*O,yDZw?6q~] s ʽJ* ~4|\~X7j\E8$06V«ʚ?X.()yҁ}j&|Owi$>NQ!X^ ,AZ}*͢`x` GJZ[) Z '0Z#O04W\"Ix3H:渕 ,܂Ð< 9VBU.L6:y_ĒRZIͤrl7TBIIVa@my%Ӄ!>'8AsT͵(M>0Bplb7R LJf.DMhz\9{nHQئ̧ׅ'Js[Ro(߾"0MwgdŬla^vE>M`/^ZOr(PiL"$~~q AQ'\wx=\rK{ܑӄ4ƬnNMciS&y6D?RowITh|y'w~s]|Wi,fgw/r6ONO;;P/;hםG؍F%ˑ6F."j鵆\\Ffڅ//h܍:)Q|uyʲDWXoh@hG7jj tn_|{wldXL(Fȫ;ݬWB>nV#z7I=uid0~cl/Qx‡R/ہWS%^Ij+k6H"~ ?Oێqyz$h ~>yH|<\+9,[Ok/?YvALJ7kk~$Va~}w{{~vuI+ZʅH~YPrUDzVciǮ: 0|!&wq|7.LZaњkBc _Ue6ݗ1s9$q%E5&ǎW|(E9 P@ʥڊ?SrvdH]Ğ 0葴MҚ6ůuE# o9lޣ^ ,_9/ Jxt*?jܗ\jOyQvj`a?;j*;uP^ 1NX|#! 8j"c{-aF2"NZ*'9iN*lb"n{l;Cw%~EvF.8a%y8OZy/l:101 )*],p-RJ>k]dRO /%mI`N9C6d]r"YJhW)R +Έ7Yp^"ly9Vs s3M ИP("7V1 UJ 4 @J Vkka7YBb'-z6WkY.&x DOe1OUCA>x5ʖl?{H=/5y8 &@^ "/f-[eIc:8=,[mr_dzz?YU ă-*PZ+)תzߧ4mBdXk BuO'i Ti2󃉝?Oerv84.JV|u@jQwq"oL<86i`n۴Z`ɮSY_6o |(I}($Muj(0cu> #Mn{)!LZ`N~gkcc-~Hn_ݧ(X}[#Xy3;XX 3j!Bt0;tYz}G=y^qjmUU ,^G)&h rMͻMh;Q vءed+6Ynt"9ג:ɓVlp|܁͍t_l T3;2 TM=[(4wvB[uo;gBR]ꗍWeG(ރkᷝ3|VNw 6y}uD`PxH(*6áٯgo ׏l[@T.1dD1 kMӇoկڑE^_j@`N RV`&J~E(xǂ1_\>;֭<1OdcLnvBK|p;2$mҦT$&#e$&d00B]TUu:FP5|0] :|6?_!+顠2)5i#oGĺ+UFq,E$hG͢JJfzi͂6/?BV1yͫYrsieJ̵WlH`9AP79=^[rm?V9 T-; 9y6]y Ybu%N뫹HW#"*.RN9N>1FO nS(C}SpG|G>fNSi{=G=}rp4KWFվi|Q(r@b'ծdgw^ #u%~AJ∖#}Sd nvZX j{0]#eN_Nϙ“.4<[s :FjNk(̱#_`hjN!v>eT8ҳ އpt8 YS? @'dqs[A:D__ 51not;ڮ?!-hВ%|6 :RWxZ:Bh·/cKG9P,%{[ƶtyw뀝)5q䛏3b-A]hN f+oIi ;7fUaI)p@ \c h 65 BEbw6A-:w(Sq" #+qf G:T%Jv $8F/Ϥw4TWIja8ш WUdQY2e2E0KEF+rbg)l5L=Afԗ$eq+49ω( yYINLs3^# r@0 [⦋ 3\U Ljb(&m Vb 5(@p3-TjnO0ʝ:cʝ*#Y3R `JxZ(N($35ur?~ BJr.B{O ]8cYok'ZɤFfZghQ\Y+Xo0-7=@D~hS+.U!*)2&2^]F')R<3;5 oc.$HVWY\oAnUYJ FZ[@~wbgAQ׼[̠AFx z= ˕'$h8XV4Uc<'¸p)406Sj< cݥ'B+rb4Dɨfeʷ~g{V֛AIXEnv|uۚ#Aj3ndr'@2[KNq {i]m1FSU{OGOmhH3 A-Y Y%==;]K왓# EoRJ - ɸDBRNrbz'ս%-foNOaW=?B{zt]Ȋj%X%K|=xm3~.ڑG꾃Ie!GERHgz`Y!KDrb?S uݸ\MI+G=sx ZW 187RV(}Kgu0Hr@+_<T='e[rV8IyEKRY4TG$S§ms!rëkA0գ!Ԓ+a҅Dw`""hs8buG#.$HTRQ* 쥏#uY-1i|O70RouLw\)J@%1D3QGo`ձ1Ƥ1 OACy:oMSBSxjl"&4ŢINOӮoշvvvQn{=ן6Udv/AjWܸfnˋ==?mų{ڌvs{w_o>})ٿ7 b)WrrΩ@q3+U~JsjRQ=K/$7:_e6}_\&gOt@Ug/=gluaoC8XF!0]|q/>aOӅ1BmXbA+.+)X&\|(Į4=tp<8ErŏBj8E Ď@Í|Xd4U)3%IfT) %bgRCvN) S<3EĦگ,k]=À:D`t.5@,kCZ /!Bm!Bw8߷!Bp!D=JsP_ %٭цƎXFcOBz6Ci q6tIǘ#CR~ٴQ]UH '8 B@0`"t|}u9m(#ikX1(x(F䩞'$9\i]9/E"0s+y*h!8qRc 'sC kovBt^ Z^a( _ߞ^o`֫W~4&0x[hhQW]u90}`8kGW51Ao½[>6kp\3Jtu:Gm5tzzZ#+64i&&;RT:q\2]#n9\_ WX9L$ptљͯ`O^D'|8 &yly6{p OېMcp3tyKo4;wYi):́! = _Gg ƯMv=j=+~oy6Hkko#r0h)I쵻g_yD/L49=F_1S:+sx|b [NeS;Cg:^eb: zgtEP=̙/3Юwڇb{[Txej]7O6w鷩GvݸM% 84J'ǩq3j2l3Can)n#]9=/d7jwMt2;Kl6?Gu݅6_^A BXhPfhrŇ o /s#Y ~oJ@/UcJHcCx#F5Cb4*05gpaqܙ _r˖K-'bLLܾc[NiG4[n'> ,(em7Oj:ۡc.R̢<;]<v>p|pNZ:C f Xmwvk쵧DwfoMv~"Y]uJT #pio Y0_1#QPF{U l?0Gf$!h2$DeCM8gh @Y5KX5KXc ^K!X-2L#Ϙ>$vL{n-FaY&Qj u?_J.Pm'2ػ8nLfB AW{ڤ6{w)&ce٫f?p$=ӖbK3@oLO9:Vʉ( `hOʉnM9&`2b<$ yg%G9ǼeaGqZ$yKKL9ce?-japԔhJJ&6Z<|p^먗o(el(}Gf7aDv% ⧼8퍰]#4,qKcހfQg}e qiiȓ7;=7ri쮿c0>;婏'(ݪ7wro(Z¾zLzuN:\Qh݆J^ֽe^ֽ^争;P K[pjVQۥi^3Rb_Q)e(%v-k Vlt+E4N5 Ahxsʓ2y4p i<6}qyn-8;3f+ߘSeOyO,ObAo$d})yn-8xvm& Ny涡/1Mbye >IlȞJGVkNe yCr]Ǡk٢t :ǰs{M5?S9vgW9K\c~TDZ: +ބ1bR]\ UlT]PRG. &!➪EeOг\$J#ks !3k@:o"X^8M7z،x>""aWJSQ, DG% /(DG@ȱS%<BL fyˣf'EVp o']ZwmhS9,c1e:.1[$C101Z F|')l]V!B(%ؗ-:wCH$ ]DʹgWJ(q)wނr&鮶md ͨIp*LB o^w;Z\oib컥 &}sQ9L6Bs((t%l;AOz02NLJH}O Π6P!S ֬+l4 1%RVAɥ s"ՉIDejgd3 RowoARԖC5VľPT¤AW*A%eXO7#=^._?C~m]$Ne@P; iK3SPT(L]l٦Y"Qڞ n^= j~qg+|nThsy~pOQBe%aN%S/4FŮ$!6IF`HP#< =1s4`5K,F1ʞ {J]&I.:y!G:iM6g_])bnj\w(J FIƓf'eQC][yb\}7ۨSZp0.Be;)A9UΈ tS2YҨzR$jIBg8{Byr;KBzŊ @10 .g@,au4 t fK%Q:ں2*^wI$~י%:ӋÞzXk |oZPF^?FN:T.?d&zsݫS٢_]ܣjl}߁J]b2Ngp}w '{~UKVtx?X !wvGZ #z~;x)[d;{TrlߏMq,ۚ%em,v*Rœkу=ԈJPjP! IڌBϘhq r!l?$*ݑWI!W:2GD2fZd6H*FqT(Ok$#U}fe3$t7|Ή.$"<{ӏڶv~*dڛ, NvQ)TUӓcɋj0cfUOصD;vG,8wXՃEhw8 $Hֽ ɕzcBҮd Y9e[db@u, XݮoSʹ;MϱnS5'Qimb4qc'ΑOQ_L}Ы볁D&y4"g.n升z~nqM~hģ~O0Z oq!k:Kst"}?  ^{,fIɘjMFuRCD>pXhmz7'pRL[ GiWIpv6J{ɝ>#?ݐr$1JvxvBmY/۵&5cWI EeJp~P@Yܮ3Ŭ՜R&:R䅥KPkjY r=s8MPk]DGYc~ 1}bXjtdI ވ VFb`w {nd9^_lwUQ~>֢(|nllXTAU %`GA̮PBl V9Se8}GP4Su'ń?MuUE0WBMW`cbɇ,`*}ܬk$yx[:AS= Y 8؀U=h#NA!ox֍w7t3&k#5 dn ߎ\AZDdޚEOIh;pkNHfdȩ䪕-Iam \mzñH1$EZWa.ѦWn{sI6?k\K4h!:7үA8/veƓ6&XiM$c] !̵&#(QZz\o~2]^lG{!}g<-?}| "btuK^&SxҠYE)ۤPEۘri :Ql9;-Z =c`}=?I$9ƐZ!Qmg-X]զj3t<쥂&u RH8.PRLG5P-)aBy+sNZgࣝw]&t!qk璖q[I<.ney&rX{^h:wc{f%ȶdS(K#"Aɸb]'2 %¢O,L.IskNy*~Lݼ7"AIA%0)sgzRIe4؝KKY#-Ch*=)m\#`JvnHoq$evk[(7̻̖^W?/9\ݺntmBQfMa_9렐f(%aBiFGӬI F2Q% 59 ]QIߤЩF#=!UITmTRXsST-Uۺ5]9F$ pN[o$ ()84bR؏ J$) [p*4֨dRRI:сQR*Z+5k*fO J*kiڶ;=L;JfRPd27±dn~re5ZCtc4UmSɭSPg޴u{Sdb[a2vg+zXud6KZ CNvOo u#Qj8En",-Q =v=Tn'coQȷ#К`rDA3JL\4m6R4\KawswXزV䠐=(^担 m"B£QO^IKDƴΕgԢD{,˄BWM\BTUb쁲(BBџaQd)-\,h&LZQ aeZ4e޲[ZF>jKuQ#]ScU?k)#ۿo(;3qjh4 GB̍DǾe8֬ja,znCyˍwC%[ f' 2I ?Rܭ$Ӭ[Gf/Q+]b>|&[⣓6o?Y璤_wxuǎ~]7q[TudٮaO><ɔ0IsFR(PR ]7&ݡӇڕ_bmaxap= ƑŻ<s`Vr>gIiu=zm[59qG_nݞō=Yoz(Շ+CxpyYB>~+6II"1*AۙS]~Lo4C!ɿRF-L6e\t0"}EAt:=/9q\=?Zw]w$։${kzȊzf,@:yӇ/v}O;Bdg?ˍE@U#%ˬ4"ڐ7wQ id vM=)uNq8_{oę;'wN_|_6x~r@sݧ9R$hH#ZHMPLyŒH뢯Ss%v]զێ=R>$sI d?sX7׽,EL3&O9;A& Q%"(Sb2XrM{#e~4̐guh}*D<,QvvDU* i[7/jV>zi8L r($gR|Ɵ |Z{%_ܣ{9+@a V ggIq>7pue+ {6f#@lQ=l ѕL_tIQ(] 1%r1˛!&UMdY{gqŽZ 2cRd%k4592R"T8O8D',AAfu}~O56@:>{Wg{Da;U0 #n֗/*?=yI2:ޫ\/^NxKg*7[_@Y8Qjӛ`7mpWH|2#bυ+ۈ)fYۿ6SчHfGe}An*\ ~)(' O_rkq#rof\GM`)%)KNs4<:@*vQ篓{{,QV^Ԝ$ڌ7"$P l8wpR D VdӖҚJ!.Q_962'/$ {5ʽÔH2~<md*mw{wj૚m9]:T,1L %x`x* dƾL9M&Bxgu}PlX5ԁ+ƐH_轳Ύ;;z_U527Zȱ%˲EidL$<)U)s\Ʋe yquRXl>6oۆtEhn7c/W14DS7!L-6䛕&H/סCzOR@oo".>?X{nZַ.*Eي$)࠾'RoX~_wYd=ޫ7)R\{Z\y\ Fdi$*b +q*+.z<rYyv7kET^D{+O ~5`f9Kdmݞ(+9v['v~Cx\n"êy[~BDv "|Pt{3hyp%-m~?2G}`- =3g 1#UGV5:?7C1L(cC 㨇 -8FR9\MI0^IEY3 #5מ//rGu)5c =u|N i/uNQ/1GwAO4yo%+  B״n%FB#9;Rd6WuzXn]R ><9ڗrQ\.@yUW; C*2aZr԰sD_c{Tq|֮urnւTWy3O!^] ƋgLA[Hrl@~2y*=EpěŚ8tpx0:KĦZoߕgLV/:~kpF<|MT Htvk*:O.ދg&PUd[πA.(KF31p xlQ " P;F A P1FOݯdCԔlL}7&g]ע)9xw-nb2Mu4x&CQ6-Ql[ .f.+AJR *zOue6收m*4r_.6P w/m%.es'{Z.3p/YDZ -DLhk4L$ǑTD)1B\LDiC>qB9e -9%M>4b }m7מ'(س(98 ZM sL)*H@NkU\Q:Q0gAb>>Sz <ٖn'@S&RJv":K$*OySEsJE2T]j=j=BY@Jkdr|thF(RcP`F=0L1*KBr 4q$v \#>X^Q_qWIJp%$e4# CA@Q4`,)y\Tj?q rzi{#A<)f*t;NЁiyޅniz_gLp21}`Ս"nƣ1$AqPS8Q8ztJmy56m #Xn~νz@_R\*5!Sluc=m)ܬuMQCx62w'Rx?J^~]|"BNjY82@^zL?[c%9$&'"su 02GH5 xuV̲#jr.1" M,H#yƁ&AiMY%2kߟ!żǡo%B#}M֫Ov0\d Hr[ʄucI|wu'aZZ`ȴZϖ@#=n!.S*5/}ڇ L21`+9ڂfP)Z暤D0fP 5 HU|&AH y!JÏU'K,MU'KWjChGO'ҰHGOf`!̌i|1N80ni#H#q'C@ kWr4å`Bq%TIK׼i,SlvA+q*IӇ/=) u !>&Ob?{ƍ /{RpJO7ur*8ټ$բ-Q:$KRRҐ"EP"ѩؒ84n4@k/ld/f;s;M:N=jv$E߅oQ ",oEr{RcRќ">x*G!e0I$ؐ@|cFފ]FB,э9bRoנO;oCC'YW;6ty?6Q!k uk`mvō?~Bw 7olrZ{"jdO#{8( Ҵ\HB:Z@j r<0B"VF Ar-[niCE[-3 xb چOnnt Dy2{+!=^I_qĜP=1H/qs>RAE)=A`L$"Pf|pJ#gQxWiCE㫴`Q6dq[ʩJ9R)^*krjڇ6L/#%Ѩ5̲ŭ 68Sرl(m>JsFVX OJA&kbGhzhzMZA,fqk;ntK6\o%1J8f#81?)mlP6d odk LAXp?!X&sr gW P;L|tm㸚{i> 4bjyVh,€f|!;wӓL+SГ`.wO;o.:cwݙ}4I<n|Ex3 ;}bea|=X{vɧ#uohz<cߜu7NF"+nbRJ*`Oe^+uy*atsj b g9wNu/1`ҋoQPI4u1Ûzzkz?^߹ɫz&ni}oL\*)Z̔ QAHƑY˽9ߋ, F NiϺ_&cs1〾"Q&oXL:7iM4G/4<;|/@3}wO8&WL N2>nz̡z̡V?:҂X \vLnºT1H* LC5.1"h l"wFJay|L0+ 'zɼ:|__pJ XzѥV"JWG3,Л~zA,̈́ha-p(UJDR"JQf"J΀+M( &`ɀ#1Rb`둸R &?>Xoffn a>`%B c"\Z<d"*2:乑`K΀S;8h!U!0AA<۝E*iPTiHPA0JxD&S\`#1.,od3+G+ƽR&Ae *H.熚' ;,K|>+gX: @z|N(F 9S誡_d!jcX`<st3sjl9Q 5WүlbA| x+pڞYVYlPI@jX6vҟ$*%aV) JIU3 3k0mHI|q-,C8R ¤, &-? \|X|簦m՛;9h_pԝr0oӅCU P\[qg#h%+WhMwa?2$QC\G'ן*4vUvDg<;3qc$8Dm J/+"W>.$B is%&z$ǕW͋R+ǁOikD4?nΣN㛮WAM>m>fFUn{RHFWDuP<%6_o!9e6t:*XoO:Ƴ %1%K1QفjJb|)'Gyf<3x6@0|kGΒ6[QBH9Z֊FM>'I6/(&Nk !/9Z2޷&epp mWOqN^$z/j.ք7K7~sdU\չ@/d~ܝx6$qvSa` {K-vpS> s`JN$jj sprR⮿|CgZ-i弣_T9Q5 s895BP骫V&66[Vj?}t܏*A}n.icշGtf~K07W_&.@姛qNDa)TU߃xywq?f*;17dF4aSUj4\N؁FU:l[Okt I LON4xZ%i8ֈ2)mgJҒ2* C'i$UlT|?nrI2pVgVw ZbpZ+hwU}0u^4B3z%ʹGjµp iOKN;G"mSCc/ׁ.ŕJƈݑPJ+u5 *H%&* QZ*B \*(ݻӬ0I2GwǷ/>*b%bw~TIxQ415'47(&'-|Ju|zwt1SSp~Sӥa8R/qY}`U/fΧ{OQIW lC7^Iތѽt3nF7opUeYL9 G-r<1#pJf6JP^o?kOߧU]bJ:ͨ?wB6;hK;/n|>XWB7Yhn͕qʸS~>b qsUA32 o!z½X)5D!_pl"LYLD.aY Yz 1$7ƿ}olE˅r7d-?֢͋7RK'zn ]ם]ݣ}Ih[zbѨap &]F_o= -2A&_t]hwa'Cƺтu0t[wqc-iH\ጰ vvMal%Jd#Kܧ0tCq)^jYF B4p'^*PÙߴJ-)D#)IG_Wץfg9\ۻo/.n?<5:lmzx4ig klrxƔele8%gYrJEymϘ[7,(8` ;|g^|[ϫ[ϻnnB>KF)t(()JEjJ %A]@Ce/l}QS֗ЇHnVm_$p'{򼪔U<*yW쒱 %Ԫ:LUu( 5ĺM 5Rb-='d!Fˬ@1R$M4afӒcsUQ(%3VA%|+&ҲvnJEA$),Cc=X#DBMS 3S Z {7b= b- ꯴[ SVSϮMqЭ'B&UdCXH*XqY3Æ)<;0k8eͶV4uAsA;6kB IaT0s %e`[7$ -IU`قf,7i-$9=cV]rbtiBe4ekIu!M% V( VճQ`>Wլ9: Lv2VRLtdw#`I vz{sՊ׳ =rw} Ow/ ,]Z ꫛO,<8,DYjx93ƒtല!Ɍ*4KXЈZ==c:<8UtSk#^ES Y}EiIٖy YUd7ڑof~pvXZw<x7m/I$tN+ᮦ2KkGy#4'?72 h>nm֦vy~P 7ںIµLDܽn>E_XHyᆳ>o#?_2 ،goW,tJc1󺠷uto?;xVQ+'bu5f2Z1B:^Ά{j`ܭyC.(䕓4U+Y*A-WS^'>5_4ohZߏgMFl^+8B6w?0j(cBc[PpSE"<(),/;䑹2-xNC c6R-M?0>櫻~VH[qquq׶^bm;DͿ=8VC֯\FO]x$1"64LҠ=ia+ӴLЇ켲V 2\犟:e*S*y8zHv\C^lU5$x ڙJ_Vtoڭ6xKV?!G"gО! <8{42v>8^As챚PNNFK% $ۍGAM:qJuQ"Y] @+md0i$qZM 1WEV:UFs6oxì+5ɢ2%%G"IHvlBGdP &4$Ne%Jl ;cbc*6oE&kJZI5ZRmz'ieÜJ< kd :stqH? fVBgKdBbgBIT>YDrdL5^X,D1%JC6AGN5F@ǿ4_NivSr c2Nc)>=ԁISs ,ull̩=U,d&y|DcbCLV)g1D #r`JcSJD_[a ,?`3E،FvXVkDLCesfu]B)(L*[[FdQ@gkN&XJ3Mq;]? Y,M)+,8PYnbL|%+FYY<+9IaEɷ45ۦEy%Q][/s~ "oYT|)p^ck,ye82,G7SA4!8Êq޲="rbS|GIAwϺ~{;yW {v2yu v ,@zV!bSfSyw0A.E`; y"VzC +VTC "Vp d!v4c7N"5nش:Vqێs&-ǐsvѨh!R_k>P1u9#Z>Z}|ѐR>"iTM#I:_Rc(Qp|ΧvA!Dhs]w:w3wk`鴾c w;S`ܭyC.(䕓4AJx6cfOҢr+Gn|!˻~>j`ebtEDv5$jBeJKɨ( AR:K2SND((ݡ?@ fkNSnyx{܀Ee7v~GyL8 J8qnXSsCFؔ87|9 [6z UN0>2P|ƮvM^PϫR6(+$X?+5f le.O|CsltC 刨@Fm3LKqa1)5IlaUSH`LaY ˖-o.*Ey[*;lA@=N ] 8xUӂ45t,(*L >ȔrE#t7'"Xsf:y\Y mz]{<$WdSUG=+C^1s^ta\ R C^Ɂ#ghy@SG>]BX=NE7Syc6!Ja&X`[ eYȫ lPf>Řj@0U39Ö5<À5=RJW(C% a|QAWQaFOŝl}7>A4s@|E~I%+qIE,<"(/ E73mxv:ZnHJbIQ)%1_DFDFsX#Eȣ.]!y-0L~7\e6pympcFΰl˞ A҅XAE#4kXCđ 㒋2AY\ҍ4`d$SG \%v t腩/hl˞ E;xAyE פ,kԛIYֵ{Ze/͏#N/ɾUЩ8)A8ZŻCxk]n5h Н"'vr~/$f.,gEQq;Ѿq oD:H+ޔۡ^ z&Vj)c2Kc4#I$lrpT_n8,GwYa:9+?9 m碽'V o+)G)즎ܢ "%{A @fYLE&$3.]巠* y6ڭvbncfOYRЋ?k9;X{<`<Ӌ#{'Ŝ+">=gF7?WoՁ@173T$茡cvl=^h,`=(M/O΋ˎ e|a]kx΋9^F 缸^kq^fw}2u8Mqn:9xA>L4$Lj8~֓r;eIy)~fO=ddܨ LDP^tposKV`o骫OG@oG@z}wꊌdL8[띇[5ͳf`̫5?f<@H o-v;lq)\$y5ǮJmٵ={y5'ւjGNd!N8&YuN5ݝԆWubJfuOl |>Q~٫&A3@8ҙ #2%gRو/K1JLq mp:_?ڴt8k ղp+: p*IcKU/w2XzCVG= ZްKo̷qy#ȃȗL*MO&tpח xjXfZRv9 Zk,rQO?(9+MjMv` ^W_Y=Mh:]+Fsv? v,XL!{>Ѳr_?L~8wގ>sTBdf_䞯'%O?. WՋ5UUh=7 {*wM6y/l谕y9EهXn߮#Lۤ'wޓ( &$Cn"#2y&ҿO[y[xJ]ɺsG.(R"0هtA2 |t7?n)9f8E|-?zt.w. ᮑl;]@q;Ӝ\ _֓̌;m_eRדݚxb{n[K.&q[<6>Og EʲZϫlف.yzl>agfSOš"\tDEvBB$F?jɊt.[KmJ{ '8@7z7ެ>CRHap/)qeIdkLXĴZX%7 eR?>c՗ {lSTQ:Z.nRfىnAYcg /l?yIm)~Sy3 xpxk߾eeJwuY>=V mތ“#5r1;4ځ*S qțc8Q0&wp0>6IBcɃZǪfmY Rǁ; 5VIjrג^jb9 5 f * |̞=RZO3L爍~tl1n CI0D,u;F=86'NT̏p/;R{N:)#[J}<<Ηn3Z4ImǜEkRErnΗD?ٮ/d'֢٫ ^פt(ܐ K%niûf$bAe0A"PQRW,S>+,Z wdYqΏk Xdfx!=oP==O̫O\.|/D-l6P9g3՝1!vB@=O$cNn̓&!WWH^ZP}iW"YnqJ5_J(}X 6NJW?pzMy{'\\=wzv`CWJy}K/l.ns\F矺 ]8{*84[W?w<(h#]gDMIԍ+q%Q7$XUkLHbi"" $^)aHX!$ ӡQeʀQך.h5]h!&U1g@WU?r?=}zM򶤎oאѩDvJamptqgdJe|zJ܏Y5vQ:AhqD UP{t`K2߅~šM(Gy{J2ƳV:.֣.X<{!'onq&"eӔGa}( H嫌XH^d[;)mi *d,/c*yBWc%!DԜ9 n޲& ~QLⷀ~ZGd3|[V ]hcxI(V',t4O^˃6apHUYƃ0b6RZ+b 56CcXow&8X񴺅 6qH )!yD+;8Hg{8Dt=kuZ92 OBƪmX3HE: )¾'7No4at(3EL%$ S8QimX4A.tř;tF-F䲽AQފ{&q“XL_1aV+J%q&(E24 ID,1:c6L B9ž=M%>xM$VIZ0JJIH(L qqf&JxAy|\ Of3^oy7a:YXm}5^?zx;hsc?zUi8X~r~3%DN_74(Ky EN97\PVd ANB Aa:7z" 0S!CeܣIRrO'eBɭ&m8õ-`l˞ 䳡F^LLN#y3hdR1^tu%ƒg(w$GyJ 0Yǯt$/ Qg1M)b+%Ź5G)ZU_Gbp7c}F8$\]5ڕ̈]J 'Up^/Lqu3=}heF]@< < 1JmWھ wյ} xnw Z\Ԟ)ȤncmPئK{A˚J ֶxºdsƩ:<a&QdVaW0>+%آOY:K/3ś5ΓAYm+ReY.bU~*\WExλyh.%l<[yosIL|'CLk Fw 5ݧQdkdG*ĻIA[˧Ye|x+AN}fW!Ir8'Qv} Z}9RE)Bf!v~=& ^mK1%ŹK!AmI4Q2C2R$!X6&]qa kl4SeY 짊5Me ^ eo8+;I A>B%Am䘌VGC^h.cGuD]g N)EC 1S f*So'5IcR378 M"t:d,L^:g=~y( GiWKxJu~$gZ㸑"b72n @Q"[><߯83zfz$*,[.zX5g}~ Ͼ0ϳg>ۧfa GO~2=*M9JXN6U"Yqӥ _Wׁ*`4HSHL_W@.ԡSVΧx=TzE>a&*s~@"SVF]QjEuѽH:hl-Cc, w֡֔# -<4,p ~~/X",VdTB(rNMH7H_BHdjE!$T8|FŸTI$69v9^͔@`Q^{P!Rd!$:o"S(P2@7(`P ŤA049t1]0A$ 8h )Alm"5$+Nƞd{&f*M8ٹXc VXYcOف!`rWH4=NXmp&G.&Rԃ9ϫȴW$ EAlQOo \Q@yOKZ>;5ūg7|r -5Pdf)kRtڃ5mSp (I$A:0@%@#QFpbF.&.tG8R.0?dO$b6`6r1C:IS(].zE. +)HڔbXߣL DJ 9Afs䆝3FAiD?/>jE;hOJ> 9Bސ֐Ok> X}MHq) HIhx7"im$0 8n1%-O$.΀J{ Z)Ӝ2$*I)8(ZTT1Ɣ%rq>qz"$xJ=d!L')N<,ѤQV\g5nwTdb;lb'n9+ r+}&qxw HfdoNufA!1%(2کPRl R@]. 1:@7D!)111pLl ZcQ#AʼnN7b[!%H}u-KmdwYoM4Rd%l ǔk`fYn"e'Ӱd"qĠG0{qNڃTr%-ɑɕ̀-;'K:Ɂ>EsIit Gh8>t=s(l>|v@ .¯m紹5x e,Tu|F`;|ɾ`=93hrLr$s`)$*3_H\>`Srś)x>V}zxLj8 MѴLQipceyu`#uuSjqZ4}Bosm-& \L_vRdYn^'ƹS縃W#]dblC4`y1]9g+NB,YwaayPc$\#D|cڸۘrY5SAəOA}m|kmɕN?Gh7; 4FOn 1=GFs5s>SN#S;SIhgP ;py .rг\ ~|ȳvv;GwxmO^jFf ,OA}1 .mX5x,$y\I<`w0`rTHt{ymLp"lTp1J >re4.$'CsFQ Y wCVѓ[d#d8eqtɝ~2Ԕs\OmKLD U m[»4-Y;IN2=0`tv@(\gbԵX͚agRmB BsؒB+ռcTfs\G]Чz#P#5< yh$< |VD*&1Mɱ= 㖎+%- 4sq*Bye쓍O~ UcJ1HH :ɗfNթ6>|e8-wן7*q^oZӮvUo2Sͥw/jăiL_ %œpNl~`).oߧ7 #V UXztviZɮN]*, <6dQ'Ϯn#*,|=;f6ӫcT']b4u WeG_QvEuB=Pm#(Q4IՒEUGde\q0L.(g=\z`js4ƺe 16/JN8~nR nnE$kȱhbh7Xwxuchc:4|&ū61_*g_4%cNXOܞśwg6e-(~;W7+hcv bo~eU8hC*#0{psK1+ʗ[~ `C eY0fVj -R߫ؼShvn`nus 6cnHQLJF F;AQ/sz5/0݆dӇ!;nչ܉a,g1`1t ,m]7[ֿޤiCyVezi҅.lXކ{O;|TERI%_tU *`1pPs" Cʷ %=bGQRQ)idK WWgonnWo^}ښb~ǟ@vγ/,ʠ˂X{˕Ӿ &ㇳ?3Yŋ˫KrIf^<7v&0ȓdn>qc#6vn>{v݇ߙ9_y뫕,]'V6bS5^Q{Aeb?|? /`A2$ћ$yI$k2Z )XNX`9ǓlVđm3Mδy O13mL\|rvݽxz>I%7[XT U]G6FGո&5GG :٣V-c1LLGl8 aDeˈ8"!h&㑓"QN oeP5T9[<Y;rTsf RgA%&*6NLWÊ*xnkd Z j|Bqq D̈́AːT^AqJ `E %aǐ\,sMC(O }WWڮTK$ɼJp9 \jqȻO5rԡWU03SY ~jfyZ vg􈉲+zs~[3n-Ti_5G"rR7h>[UP[O) NޕX 1֣tsMBv}kDFf}!{C'g_ PkfMmA}>9^|Io5RfuXzpty4U$A>ʃQ -գb>#Ϙ<8F`A) ]0B;`22^/,6,PyboyG`k=cVpR_3jVCR8 Xɐ~ZD:KZs HlTQ2ߋcڛzv5#bn 6<|~)o]p4ǂOli1IE|Ա}-[bEyg9|".Dp݁BMd!k!Pc:~ |=zrf켄{B֦dMǪM3N*DsRJG0`0;a c M\QV*A$W4x`KιPAJ#uڴL3{tfz,Y]o!Ta"Ҕ O!xHF{`& ptS H#m*0ܓो c4E)FzG%-;)9]&\-!X`)ijb@ W$d63\Hi0UP3]Oۼ>H%l՗˼ܰ㯿N?'t)ǤG"Ad=lr9qvs~(p3(׻Gɣ>fֆ8+˟ͷEB{Cϝ)Lշ62$ sxc#~Ph\6P>a Ԣdqb0֚֌UbLRQ*!6ca VJWn2!P1. }{-ʆ}īgJHǰ)+ܽ0R#Y([ƠYОF L~(Uy@&dơ\[hλ @a}%ư*8ަĹ4¡ɌN8Ge.&U 3my3 {cwG.eQVK$ܔ /3m9x1n#.4uqYM8w}Eb/kW;&s`^@P9xZ1O/{`7_3;87OFzr!|l}5_cY$_>= 7 W0}N(u;'J1sr^vz/u~<'q3>^)c73_~|s2Jgri 0R1ݧGdRu2:wiNO$ HEZiNɉRo'%o./Zk- !%jkVLP,.u/mMm|ua(ٱr1덵`p$}ㆡ ٔ$pdIIä*[,fݣEt^'HkL_)e14AuQaU*m֘$*sZ%pbf70MdZ|Y@&d`T6vYELn &&d`.Ѥ7X%B40njS1͹#GA@vl? L䱒fL*ӬN;͊ﴮ &WAhW"9ǥn D2y+=' >LUGlsƎ2&rK htW]rR! |Y|D[Ipx<&!ݡXɛT=AX0H.2Y8I3rr+AJsP}m Զ؏$0DPE*h]HhM nМI{B2=J9Q*2]K Ґkחc7b4`=O<Hϋ~ o/!Ȼ_G_t[^xh~A;qa%iKk Gn& ԪEg֯h}|{&(iѝ%XϢ P\lx1Mݳ"萒~e40#-1h$\1NbG:,;QED5nDvDȭ_3Aի\UAPB iyL Ag['hc_, Xe%PvO|Xv}V>M3D%@o` !:-adL <1۫.7Glu.zs in?MF\aчOglyoG6,ÝГ8=TP_K\yW˾SqV=@G3QෝDk1P==O 橵G  cx;0&ۼ`41!.ΠȄ_].|i_|@ѷ~n N @-oK_)[q>a>PO)CjtYno?[Y9<CSFloכZoֽ/428i(9Ꝯ ' \BIiޒ*wC%`pY(0L@c.ᱢ+$ۦ*zxRcHr0< 6W^t%3roΣپ\M?2w;Gl:Syɫ/nzUEƫe߫S 9|AޓX8kfz"ލ+Ґ[[.][]/K/^.Kye颇KhYіt+; 0eQIvjm+;ưo fEN5Am wԑ"`l0j(^#ˆBp85>0%d[DЕքWK)(!*M \9m ܇^j6pczݸlM/Jpm4l;uvXPq8nΚ]Snx=k<ӳw^>x7{ lwI΃^y=_l" ƈiܪ< ?Ƌ/kwwM&f@*7#[|2Xިbی?i7~!Λ>]^Y8׉RtID-ū<ȨmBFƁ@Z~:'B S2A jȭ_䰬ؼOϔGЍԼ5-,0k%p\jObhK`lbpd诨ܭ`WQ(t_IQ 8 ''>=*29HeE=&+t=hhK'5:h[^GS V*1TpJdH%jEa B<(`C)xQC,y\C b9 KG0<ڑ'mPhpdCjyg@Ino1jS?4 a,waᐌuʪݖ_ߘZF;VGmUէ'a}WeQ)m1LAnN_L?i^"_~u"Y8KkYn P f%(ת$U%_/ר:] qv.M;= [XaYqS+G3 r4VK|UW zhdvw~'74 5`e/_g- SA[]P%MdY--^vj;rGV o}dߵ K(;ӟZIKGNq`yV-5*yAg^HwيCƤ( (oU`#w 䡬)'E;1 #d=%$dw!]w("( ?{S[ϫʀ(@nc -tǷt:)+7÷v=i,74wbv$$:IkBґ5 p7hlQR{ޟjE;PK@Ϳ\ C ;'1fR$wf l*ג ъ-t֌|vң8wJk\a7Dڗpu&ԷȄ_":̭ ?ȥ,҈J !5CKq)e\J=Yv$-u&"eW'h77p2psa{rr9#X'l~; [^qqގV}_4P?L(K+̢0+ϰVo\,l*rD>΢^i[%*|(q:'a޺u%הcr2l6 )@qҶgI;;hrNy/Nſr|qৃ .UIk]$IW  QaFҾ tt<@gϬm'W7*))c<2;w/Bie)RFjp_8T t uhS˂+ / R*g-ehq?~P\W lU] sKث_(7ag*-z?Jh#]K̨u{|\ 5IRBcC4Բ . Ƥ_(3vZ) 9+}u70Wfd FYod{d#ONvGuIwϩgdirJS[JsQĹR\SkhN#c e 'UzlMW+tRȄQdr܁j *|pS^F3lIUBQA/K!_hGjxSd~r\2h|_+Dcʋ`Ɯ΀Ɛ.%5).H N'~|VqQ=U\T*.*>GWZM833vo"3Gli˒xx, 2$ĹeĹ/Bk}ZNV(M6{!(WDU ͚T⿮6壜\ޗ.Lp7)ndڲpܘs c,ªXRU}zEoޔڗU)}MBCҡrwָ5Aeq^LJ鲜wt o #K'Ee]Y&a"iBqJERbq7 -(#R?,:+d5./%JܬH88U2XpQ-XO W^q½JJAK(ĝEVr##1v1/[ˊוdUbzWJ[ _?=~]DTgE>񣟮?ݻ3\ PDb6|ŃgJSZ,vo;7lJ5~v)Fbո'I\5S.P|'.*'tU6ShZ b$2}=6zHRJ(m-GO0}PB G8Ew)G6)@$wBy qː{a7'A =K8/Ba/ϸ^ޙ/\:9&a5ZMTM~N{[5rR| zkN!eOpǻڮW',Nk_M:8|py=}n,{{y-(T9Josq4׎āiټMjC0I՟I;x#6 δ<1i̛yӃ0O 4ae̦1};Lob \}9?9}ŲPWpv; Nk|>`[8[!L2EU'aIa!a $u=Ρ6 Y[GY1)р낓Aٍi5r]:Hm@`ȍ@+8)i)e R͜TDxM9jШg gBR%cIQM5t#*IҬz#II뒔ٍ$iR灴-pT8DZ>L[lُκ%tW|k“.0KxGV=E}YiN#"&%F T)B8~(!ڣCO?Ϸ_]/,XOƔO ]WIbN H>/u0L DpC+eG}^XZRʃq58}R\8N0/ҩ樼^|ܠ\\*A=,Jа(wz?ΑjE"*aû;Pf&&s;EfGKOkl ~>{/JL 9ƒjBom-b6QVO}W*hyPAߟs#Qg,XdDKO9U.üe$,JU'nPOov@햋AFpE2v˞hvBBq_AyH-w Bԃ X'hb`fNHy[4G-7V1X&6 4cZS.m/6Ғ ђ7mRU@Z: afHNrNFWՄ(mÆr!$TR]׶G&#ϙZc̝yVFMϞzZ}gf3g$°YKivOSkcsKqRn1-ؔ>jlZl+ȅ#tnvF *_kW!ye]&5UhB\Nl=83ѕQ45 {kެ %;wk[Q`@i/-JYA&Յj[5wgoono礗*oVc?yf!iْxrQC]',l[=hn^6&g'g9FkAMϞ]4yGy(X2='U%<);*"ٳ˘mOEB=''6Y&{.aOwBEk\amqF1B@˻CKh:9uZ jC kC _%5Z 6BSWNz @+6~>5]$0ƨ!7re{ڀL/Ŝ?)*amI#CWd;5Ӛ-+n|l'0Ƙ!gl LA ibVIRhB9`H^;jk%[0Y9{mYk¤ff9n0$)[ƣf!T iRrײ*u;]E{`qt1փkd~;G?Iv2Qik?H[v2稔ߞ zW28yUvHHWp:}+$AO{GBCUAOہI8sÃ̆ywfsKZ6ïW;a:I)evJ)2% qUV j 4!zk9 ZܒRje`CAjXiD-f$b;ܸ# kE;j!a-'S{36s1J2T .nb~KIb^'6["O??S f.W`GU5zݶ$^ur{{zY &7l*+ݙsNRlv'QsL!*+lxNjZ['xUB"EM~-SrؾS uLӛFEٻq$Wr/8eZ,Ip.0@nn?t;If~q˶dS)Kmaf[*=Y,Vѷ+ }H@{<8xA<|þgf:h(JGct{hDذG$R(i>, ^~g"t<|FeF8zǎC +$Tq1sC,XirI UњU[+2e`ETjYjg4xqMG.m* VN(*3öKf}tc1*(Y2CYa_z<j׸Pysc3Ӕ0BՖondZwr.~r~|@>h.AdHZ2n{tbHTjPPC JJΔC[R̟\7n*:9ѧPT91o5ݎ愀TCWc@1+Q'*M(]:ЅA<3}(:uO٥HNϯwUh88Zz4 FONl>[u)8l—y,]ϥ/RRj0 V3j0j`)dm` =IOd`5VvG3X9yOLKorN3jzڮ=q nJ-%״ʽ8YuOJ87}ٔ*-+unΉ88+ΐoۏo9y//SrE7vy<wU@Kݖx[\&r.;7J7*bg~x;;EopJ\JzH}"dTŬY/nݯzT֥THCI"D,1ĭފ\S|a$/uϣԑ%z$~=޷X$40!91@YĔ/ϊU… NN̝y̹3Q{;6|tcZ/˕ףz$4 22 bvVR1t0YPvq\ p.É2#`" Ϯoot †ǛJ9(j}-ad+X~p|\8)[}U.ÒSV " :; `8j J%OrE"Z}^Utvw 4S2⌚ڔXjn+w<MK`%yMm ~Pt]^L57wI6[¯qp”\;ڔё %M=IM 4uyPj},zqfxY$]Q1H9{˥-xv{}^=}HMI@Oz]93F[|N4+MqAصz)^n[0pOofJXȬ|s"̜vK5pȃ=żG LP(NTdǴo8S}D0ߍ+Pc/;?Lc7̀z:wB%ii\&9UV #rSh'Wfjg_ōCZ3uO/C_#ojϵ];SkdQM* zMَmHxnY2p:_.Bظ66/۳W%v(elO[D("^L RՂ!%ڹ3R2kdeÈ]նˁ +SH+f]@LUponP/Z [;dAEΣ$K"P.RЃb,s>$a.4)4)k'JS& ˌnm >O!#%}Vc:ZRJiRP9F["XP_:*c2IE`J@1""]*hЂ3LwJ}HIR2"`jּve ]vov'@Bl5śPyEG m$@ pߙ\K5DA,s0KFGPxQH܂:ZGIfw8m}UoҴaAne bvd -^I[\«個{Bę!K=jcEO񤪯.c]$>k2:^~uXVb TU +UNxs>4cA&2g_']Gw_UatA$rwۯ^8i >I b]FӉĔS0Ba4I^EPBWTci*q Ϋݷ6s'Y^o׫$ጛpNHA+} w]R!"lh .0h[`T ɭK Oؼ73PڮDq09!3:IQ8ąTgek$84YU))JCgS.EHrrF IM0 +[5Uq.+Qv'; ml*ٲ[z=vI|{ہ>*ZHȮ?z|yc_gZIT:Ǣ?NpƎ&A8Q$2  2sXm rQ l q^N*-#IBٜ [NT~/f11I.AK"GШ\q#_/׋wTgeR6!_;Y?L幣Dg[^_79%9_j5aNc\0`CmXP)WviC `6nڭ 1u{v ձ];}5~,Z@"5t>n B\/Jg86Xh0J/&~_U!G5:YNJ$elq\!u^T%@4F~=/g٫/tPo9sNK9{b. @=bAg+Y%}w|J.7Hp͛ Qh=~MYe@WT{rHFVk;]Ba9Mi62q"70xAIK zK x/)pv8XO3#4 v^)Tf:?CI,͟#C-i|^Hd\[pE"n$IYЌa&p\CUS<^3CeCHP^zr("VyRRЧל8SӬD6M,0yB 5LգvmwL{* |-{FEo!ub*MO 9)=Uݢ,pkOY9Z9!d >Tӊ,R \-:- ONqP]@‚UOWֳO1ysz1 n= PZƙ4D8lpDZ>̛Er2㞐֍rASZQZIC1~ͧlb*=o aȐJ- ė_]CߑKM| 7V%qU|>4?1C 80\Tx"#*RCt+CR T6i, _(# v/?]|X,X/fl6pۮq]|\շ …_ܺ]os_VX]啗~.fY=~h]")Ɓ2%™QnY XH䚺o E}7ORVoU~{䔞*|o޽1$&ФI&_beF7s?Tx7iHQTDJ\g3x`(lVfeņg4G\ ۳NJ~;)vR>SLV*jQkkAmNYas9hMVVqn ! [ l [Dͦ|b[L!R$./$Ye9֡xFnQ7umfYa9¸\KMbCJ: 6 ZuD+'Bͩy&n $JaQ&AL 2D;f"5K%ZY*$%~H¡մn0YJ $B=bjn xG %G{ɨ`0Fsxd)cqۅ"u|(S^dn^\y˭AJEErܚsM[ѩM H!tp*ʕn4$ђ2sM " N 6n.7jeP'P0~&ML'A‰AN(rST(Lnn=~% I&aڪbX(glrd Na7f`kIb dv .Rԟɋ5C/jGR,KCaB5*L W~X 1 Z`SHg>QU(@qUj>]b6L<}\֣졼pU_}(}yw|\i0/:5ۻ1ADD.!}&bb 3BMxrql /)f`<{t{7E Ih\>w g*0=K"(jSg|ujC 0׫MAVIeXP4cDjc3Sj#`Cm{tȱۮOAOUw{iNrv&T* i2muTy3E ka/ۙs]izz;\Hіw\Q=pk rByk7Mz!Ji NPp2HXI{á铜ݖ>i kޮ%]eI"`v6uz>ljB L~5_zv  xgҒqǃoߝ~N&dܦi?Nh(B=h_ǝs|.:$?6cjյ휾:!z|mqY BR yӞɛC$ q}LwϜy?%8O‘'qx_bm/̔;6gt3E8MXƼ o!CsHTO:0m[C)o8z=niw1tkgT )Unv@N0 'sIȾwx*,*ɥ,yM{!wWH.m0/vh>.͋ BO1q1L'Z?wbn5UK: Ϸ>}4QITc)XhQ x%;m#>6k:1\g& %| /ᗸp\ǒQ F'9'&MX 6$VB8c 7)SCs n&׎ɂD5BNVB)POoύ.2d nv-qC7 25HrAß$P "铦F,YnaɚF)DzJNz1(( x0D2$I& >-a RAkm!3COp$\R<&LaoBf0ọ ঵}7^C+uef Ck!J Soj9DHaX&Cv(q.5X&9Op+W6N! Sw:wS\雿%y}KXa}Z$PDS 㕭`ב̹O[WW[T䅻/~~[O*"fl͆,( Wq]b3FW̋va}ޘ>ol8PSA[_b|ߟLq]>_S=8IǺEX +;쐻C:uD#3acP,X&p3-<9O0R $NUD3/UI3єhHP{X*3N9EE"'Hc)JP i ym3*B Vɭz"LLqǣ85#xͤ9I uw\ޜ{}Pt"eGtQut zk{Bum-n̩d3򫏋8 LʼnVR1+RbtOLJfXQ#"RJ=!] r'4O>8 rrHN8TkɎF.qF$Ic"e"I:R 0Ű/2&69XY$T'ϔi,LNPLXT26T#?H~k\a YO*̀m"iX>TLI)zB\^|7[Ps o^wc* DJ cZ_39i=b:-~V_ߎs8BhZHk gb$mwV'ΝJNC*Ъݷ@R`5dkF7VJϔPd+7+rT uFgK=`H.]N .]RpkCG`F!C+*)s&ɢ47 ວi^mGg&nxJ Sp\~m ղ5l˵w7K˿,a "wLhq3Rsvo>{( YْK|q'qB{; 9FhwoRCZ[Ӣ>X&FuO[&Fuj-1wF}CjFlƖ7U|9G[{G1dEaf]%;Zk+q\l#Ty& -Zocr).r vVVwŻx*;!FWCbhiGV#Cr(\iѰs6$< O'f{Jd Q-qTߒAb+Pw׷ [1h5S/R(UhHNN8ZpKN~0kHGMv7Sn!Vy0O:9tGOO40U9H=xuGTi6%JQ˱)c/65mkRѺsJ5IL R`a2<3[ ,41i X {b*J5Î#.xf1}/" wXCj-)``j+@(2 EFQ-I+\)4 IŕЈQgZg̤Dٺv B\iɱ fjEH8bvjd6[?\>}Cߥ 8@76= b+qt,C̛ʪyBLA 1oÕ+Gk{Q|O;ci30=bW/uNhV4 Ħ DM752t3%:#nB3SMʣ#2\?'y{8+?=Anb"q8B A:] RghTju/Y=S+ҷ夒N̓{AA3o JJFwMR3AkM8,HB./;݆լU,I>.(͋! Yխ-sj[|Gzz v?אQq@9/2*Wѥdp42l*8꓊fQ6/*1m]Z`t_wS{:{"_~o+tYv+~Rrۡ7 k2_ rZgk(=IvSJG=qVe>-/o^}Kɟz_dLeY,r)1 ZIE拁&pZHڞgUK*Y$V:t$(y|7]J⵪5&x;l$x/,"U4sCj[9lƞ|y([. ^mH-+'=UNhbT8m*ḀАW:K(R>I拁BbL;rBh+WQ/%u w,&$ ӚpZMW[__n+?,Ie,by%L(Lg1I]^3./;w䦥)uGɁUhuDcYh)i ;b=2|-DΖw| TiUIyYሉwID+% 펄N#sB蒓f8RװҎ 3DܛZ~n̷a4C?Y!M%*^x'‹}8Pj,&3F^38wj-rr59hOlܙb4L+ ?vdۄJXntWcN =bxrwh,tq/oCkd`O''(HKZpIUtp8p~#q2s;۪6 6=b]릑4Tr:-EZ"E'ؑ?؄vITm d3Zγ`sޓKqʸd :YG )*Haldrf@ A2To0vGEDnyEq⤾,3$)7ϔEx9bb㏉P'gX8q%s?wt"g;*ذJhEЪooLl;F۞/`n O]Ir|5՜JO/yy:S4|]ю|Ha G W;^^q 8 8zw}M*'=}?~: uC0/t Z!u!*EMvNτ;~Q\kP(Sٺ7imaA[ ?]鷉TcʧIR!r埓>8촣[YG:jtn"#GLpm/4$(IZHmY_?]5r>Nm ^$(Y$#,D$9&"q[P%TX&O2)2 D5é$OQY*`^eyA T~biy4jDl<,,͈}K)|;a*7S* ҆8GLSi 2\(ϒG QXO) S>䇔rh" 8Q̒,Ҥ 4U4M_;j4?G<=Ζw-ADAzMo;ASsfttdt۵A3T'>j~(]ko8+F>X1ld I("5Ķ\N ߗd[r,F:|<\yLM'l RٖOAvgm)Q8ƽTj澭`@X~z!Ɣ*!c r"Sr 6Ù#l 9JE{3z1L/*=Y!)xgȤt͇exsdLj,92&ZiyXmm nID79Z$(؏R fw%]R} 66'XuCjdj;niwʠSN]}pKrErF߷ _(QX T]fKGeb"^r^ Ii\}i-܂])dAF)Tv )]¥Ǚ0r]aZK=fZ0vq4IA8_qKfj~GlDʥ[p Uy]8ڔ@^5T IW\Jvtp9Zue0noP{ݾz(Kֳ5"ܤxʽ2iOH@R8-KD c_y&Dd:d*,( 2NʧGCnmvۤץܫT'c jF !k!L`D%kܛ)N.>]J=EiZdǔAS25q+U 3mon yA8Do(2e\I6q /wCSnԡ:60_I:afgVS@;goBD7dT3|:R4&Qn@!A0 e`"3G0J)`\Y<\]L5~3l#-_u6wV1+4;sIVp˴:s&XLL8"s C+Tj}hlbeڳ=йaxtvm}帵#evleL:5Ǎluoݧ dGq$})O{L0lel 0-v[@kLDYS p.,HlEg{ )9XڮP!x`MV-OD\>uL9S \ hoΛS9grcjǫ~Z*JkC ˌ׏ۨհk~6ZmŚ7q^Nf"{߇O;/3|Nxq~vv|z|j*_K7Dե1{wg~IhIۺgDUus:.M ԵozJ'SO k}ԯV0MtXNߞ*. #~,AfJ(Yd$g1L-mcXK"3Z$oHpb}"M6(n(47RQzswteV F6QOFoUk(=uۍx CdX7`ҀI<(ZQcF;AEWL=[Xt85zt{dk7lKd;Hw܇FoNIz ݣ_-~j~>JT7[Fbf5-|dߪR7Ui.BA[I08=;]-O9>)$DJ| :ʴi2I?z~D%Z+q#އA*옚q(+*]N7$n8laٯUO1\/Ju8qw\*J`,C.ҁOUf+H3!wfǝP₤~ĈKkS&κ7CdA z2k 5ޅٌ3뷦hխY+8T5XOG#)1V0 uq(6037j-JSE~Rs-ů(>fgH2M}o s(̊`<l_wU7wl/~x0Vٚ"=ِ@l6W5ܨ,s)fT}ҾV7!hs1`wx5R^jُ.K7 ĬF D>#fy&1A,2f&3lau 1%ҧRvmeznRDtjre┱hҽK*T^ŗ]U5tїFyn?LWh!*Uc c%!)A4#Nw%ƙvC-5|cе#K>kY%%ixc`Q M̉1IVz;BO: {ɕ8oDB",(3Yl]=Èyِ ˩fn`TL:;ifGւ{_r k!.]wwt0 *|ZrpHc[rX&ӅȲq!08 cTP @L 4hLɓ̦S%5ig{"9wK ln g;G9Rz]ĵRFUYUQme*-CBrBo&V`5 IC]Zнb]1'f".$MIm١s1ޙ>/. o3FJ) quI&Q9S{EY6:sJLY X̼8.)sϰ2*)_/6!'GNO=W$,ta}۫$Yc%bϣ8D3\!A<!] mnsi}R;OQyz.IJ1a..[}a⟪RZt[ˉӶx)=\3Q9u"w>+;saΕ?sem;Qc6)Yu&vI6oc"a!& rOK;]sOUX˳~,9e?{Fp+U& S6LdgyHM.% 8R% XXd;^0TU:;  Tҿ((/SY>?e( Њ,9PG~VpKOԁS|[=OL f)cMǍluoݧ չ7# ܟ= C h3-3|CLzBPj!d @ܶ¥ˆcJLHe!-ګ T4'To ćO;S+|\89rs|Tn*j.}?S;#-NFOZ<<$9V:A9{fpMyI9ﴞ؞fٍËCC~շT k=H\Ϯҵa?/%MI7mۼ,\oUk(q?uۍr ~S` &{v-T{Om1ijSeQgaGI~nSt{dk7lKdPXx}hF ԇg_j} nh^<ìjx|bvMo=_57ٻcGͯwUP]ʭ$'JJG "'> 5 SEOQMq<rfky0w+혝gQVUbUɻ*wa;~կzg"7>W}^z}%R*V9WWnPjai #iMY\YnD?2:7`2+"_0Hi!r_Op#| s'C<9f;Q} { }& ot3wwcW489DQ~JA'E7Gpk2{yovph)^Ыf Ѫ#b'}B[\>}dBZ{~p1-~9 H8i6N^y$l=m:ƷщWgdxmp1`&CL p@O A30{} ){Z5Ͽݻb廴:E3g?~Ktj"uIUuSG?v\@BMQ'A*9:X2A'cM2{'n j<td$ l Zҧ/iѪ3:-vVD7&*9 D/z@mp7oonx'?a=e<{iEXu.Ѭ7D$_M'+3_Aq`o>)^fz>U{Igr) KdUv~'*Q}Ǘ+_R1]nUֱ֫zލ B_U2:檠%VWy ~%+A`xуA#qm"0#l4p\I/ drHl$hlDhAφy֊o%tMNSv7պŽ6Dvjn#7AU7Whl+_qEMY/ п n$嬂\aDUr6g5;! S1Ì&ƒ@ gŜiOQ26-7ANZ3rO+:|HB#hmѻ 9HL+16/)p:Eтaqΐ$h1.b Ʊ$dɎvOK t[h$f+R.Obʇ$:HI;-2IG\R xF<$= : K`#Q"̘JX-! a6Fk|ֺI"Tn*QVdX)f<82&ȕT0J`| 0r1N;eRd@ –>0irpC`Cʫ _MedU̐I"+ŁG,_*F9 ɧe OLDYN,;04H99b$1Bm"G&j^/pbivߐanjzy`uMBSf{ƙKLr>!zV+O%et> ![ɲ}[M 'ۊ9 G[dz9I*Z5ŒFq9< i냓cB5{y9`]3ۣL2Yv׆.FӇ3;# %*u~؆#='E<EۆzL119$tcgLcqig6sx]W8X'V.B_bA2"z'(&-(mpiHD|c=P'=X.=]IՃ-%aVsE.V鸖UѻYM {gG/:C}:>L݌n윃>gŅR\܎D{r4yx=FJTP`W}ç.orOi4cGbdc# KiM=2||;ApJZSJXpFDF266b5&e7"^6*E5b]NSoo)]x \9%P+`4n\bP()%yR1D I:LBNL)MQ<)8b(4ѦyصNVֵ0r.,N]]~wa.rK⛛B<܍nf/s>Gm rWgS S4*}h!?w衵<^ Z5TI("IdLWLM''4$C(r3ٺ<d}<{Χ-$'*.xTBrId()㺈[͵m[f{J*ٞip%d4唣t}El-e23ςRr7 _RH&|}d^3.u\ >.N|%e Ys{C.m,a"+6/FNݖ)窤rJʹ7  mNʂJ$\^hJuFd  Wƕ^OaֳnM1 {n=kg)ACN lzΉX[ svW;Y/k *;CE)&Q0@*>TrJ'.=HT)"ef&2膜dea͚Bҥ6J(X&yΈ}pĒ?LuEd+@x($irֈ)y"Dɕ>D+|BNi}e'4|mJo (HJARU RT %L:IL$ #3 )%IdȁKa(b4:N!CN9YČ M*BR ŭ%.)\H*P+\A,CG"׵lr5lIm d\j& .G̑씛B4)ed9?/lZuZRvh%i{lL$Aac[@%iYK`Xl'/XDCԑ\1'|41\4^ֵNKҺ6\?b촋%d"*-XVS9 e]a+;K$NRYj$ LB#473 g?`dB+qcSDj:{}aL7!CA1nζMsi:ծ/Ϲm9n:Ҳ^}»zcZ{ ӹ=ljŇKLU3?nHIq̐c-7<;OwӯUoG 耇␊"Q=8>OnjtVWF qG-r19/TM;8ZF}Z}+>l;µjt+;.Ǧ_AZ#+w\DW+Q}+lSBZl.֫Vvp+9K~ ǧ$S~CݭFmǾ{փKrGlblp;W;n=w"J?XF=EGjKHuO nNmc_5gvhf;3`vmW~<ؠ A1טZ(RL2q\*Dx+:H|NZnWW߶QSn(oqr`}klڜ|!ráA~5kY4[W߹h߾f P _ m+EI_{e;_Z;0愥~\0v}Zg#;|^#"0e\P UYP U}B άQ$dNiX]e4TX8\^|tvI&N8DyBXa<bL0wuQIGlM;twAY! m ЕCvCACeI&䙣,GͲ::> C> $[fw U6eʕS &کMI29notF4e4NXNK^zѭG &کM)tnG_m7;'}}0HEL8VK0|o_44Ա]O?Wbe=%^,>Oa8 fx'DAo@ڙvDTtUDܩY>"VvB*ۯ`K|6[:dxR59aqTc"H 4[U, IӾmho@sc=p)SdMM8%Ez\;y=zĤ w= k Ho#wVͳbG[sܾqQՋ/B6W[ ko6QT\& oNwJMr0<~YP'\ A P"-쪞Kx,N)JeFN x+a$Ə|vGož2u[Q*!#%(bpA`!̣ /\^p!R"(֊^]\MɵTR|.pV/) L4Y{xϓµx0E=~ݎ^Q3mcs$X-:7vݓhL; @>@EustiΩ$DK=m71}qbD" FbU QYL(QIs 4F%<` &T4c:ӒGA %~M )lXڲdn$/Wsgl^ZNC<⢳@a-B ɩ 讇y$@88+>Իfk#0g]6;S$I ;qry8??xy]X?8qwonir.U+mE?`.Ro_\A\Kd_ק]t=_NH[nvy\%y$^{(K yB1C.ru|5hw_cI d^!Xty ƈ(͎p;! ĶmGb}bodۢe#mR rI $Zp/i0DQOnRPK=ADONa31AgF9l_6#X^aRN4eqh+>38w\HcbTIރg1JȺZyH{oφ)NY]cK@m\ɓ/b;.[M1?T+?~Af(( {Su{]^m`Š c 0h9\ݛI#=]r|nm]WduG^ ^Q ̳ Ttd]˛ns<=ᏽ4_H$ cQ<>t28{D]+]0۽s+}bgq!8LD C-u%$vhApÍTek _h;$2Ero5vxX,tשvÄ蘈|"|ϏN]N$~6|%H-pD1u*%h[8 ,$әm8N e*2fL-fF p Nm@fRS# 68\B!LMBL!HXJUd+v u45ՠ (sZCT!u+AaGP!Ň)M\q {U, 7T7!7p*B|_DH*-v*99w!& ®W%km=NI0EU&10k>XAnf+ې*r&A?q|Bi(Zjz];y?]` !LFޕ.a8[}B&]!?@c!@fTHtB*Z :' ;mIґY5H}_Kb icOM&d8fZ,@20N>u!Q %u]^2㔪$/%p>R}QK [yRCwLƜlȋչs6t;8*{VDSeIP.}ʧ_7 ^ݣTy=oa3bHbƚ΋:ӳႹs @q@"1Da孼>Kؽ,g)L#] SLd_ R?in u0U%7D6tHץ?cB&En v1H@7i~YW0Y2l2Qk%F@‰QgyݞObh絯}瓅.o#lAģI^1`HK. EͤO˩n{5J iw 2 bd$^F#J9!'=IP}Y|f-zLs-zpe^.dXVZznp E:'WoӃy،t1 bT#gQkj 1Fhهcf y N$g1|KO!6iQP:0>\PNÎYPLXcнhpԻ$PaZ̽~Qs8Q伥PS,iFR%UNRWm#R"۶g%Z q"Icadbdi N+R017rBA+[7\Q>>eU!Z:׮}+J!~,|5?mmI+ Bog {|S2A B6 QJ$j& (0Q ѡ(\NW3"o.Fil=m$Ck9'.bC9/wݛ/qܮuO^nevIim9FUNq (wC }*mc5S !f[ s wle;w6ٺp:Z+un13̶E.%X6}~lo8LUE {c=+Zq)Ll˚&VrZ?n9s2y&6OX[@17WJi8m;24Q"ٻ]xs)gkؐDIga%5ьSbC3"U1T%)U>5{Qk%$2› e&"<7W>dPwV؄mDɍt\cl 0!bB(U)K )D$Hj%]!tfSCYJrI,4@c" KNg&}@ h[L-g'Y B.ԯ3~GMti8 DŽw=4D15@sti<^5DN(p;'ؤIOV~rUyy't|2eba8Y F76tzkg^'|cawea#A8#A7Tb',DFa _sKg w˺UT}&:f8(I4 wYg}ӳ/Y`YdDkCS!6*#b+LJ"ƙB !SO}QE|1LׂN' Ξ]V]lff,2'ˇ5MMv'a$fֶNfPM&'^ n6Öb5>iOIF$hT2 9c2 S :nrG"'޹1 "-a`ǵR-$p`^1ME#%Cv6iҾ1ԧji*WONa3z2npYOY-ho?ƌ^~k~P( ?[GLQ1P-/(&D::-(8;8WT!.`J|_MlYoD-F,Ԧ\9~96o}]|{"*xH/ oGgt/1Ȕ=1LJǟK 2 r)`4#r>%>͖yɌ{%n. r9 KEp+zz"]Y^nŭ~+]SqAw_5{w ]ww_8T*AP2 )g,ɤ)N5)C4F(AH, }kAt-}fh]κcͦ4I^%Y'T)5dZ)xw_'[-NvWi;\t'R ,hO`[v։\ ~<6XP8zy P0!Heåp`O8 %.m[]S+eB,S mp[pE@XfcHL(ƍVB֐ &mn~X- GrKL–yYUK9eJ1EE.:-riQ1:'8IPə&¬b%Xh$%)wmHrmbఘ̧-&/0Ҵq$'9mlu^le3)dU=AKd42H*Å\gZ%&+ڪ6QRd }rȜ/V2RyFCci I@~E&Cv%xJAXw55DfL'm))ǡWdlExL9!2odD":m1g}'bPo5PQBvm1{iFJ7鼛1!D]ߗc9r.)_{aWc49gKJ}z5's3FZ+\!_CH2 99wˣ> Cb.c󀁇)@ \5K%O EXqUKLazAjռh|)%=JR :8j=yD :(<IJvA"'kr>EZ;LU]כNYȆRTنW>헿u :j\\s;\p8R/Rį83d:pKh:_dUV5o%ؗM:a 2!ۉ }BѲClûbr/AǼS_Md$5<3R&GFFF07xFtR?G?)I.u7}W3,9zyC."XІ=5ۥS'=yO/8Qr_q-n-hEHf:} QPVzzd4vG5; "EŬ'/XP&M@(zt<Q{ҭm?{;d*e3eb$%w"@ =B,p+ٙOElxCd_̚ulpln{|XimK;u+`aY>-h{0yS8o| \ E4}ڶB 8yڣ=XOmoSF nN{RLcloPyߤԧxL;s4SoWsbK۷OG -۱8X7Jґ([I+q6uS΋Fm5eڃ6 ,Sbɛ\JMA?4@ ]{Y,D`g7"^p}Fq a9[ޞ)aehq͘JM_!pHW 4]Ͽ>"/>]HfAY#du 5hB.Gc6.%i$hr3]ԩjׇom`x X?xwsUտ^r8vqʞ[lXp3Ɂ }8B噠P ssuܥUvj?|qI*[qο؟?<<ǢNTL.azEÊ?+4xZ߹>t < Oގ(:sTo%{â*Q 4$`茵!h@0ϱ5gI gɦC8^))XK>$>(ˤ5OO BUwTU]x@H9TO,Lc6k)GgQtǰG!|L '3Ͷw2C((kAΔ֖KlhquLO&8_i F1# 6&8oϣ4?x?cd$W7 x!ҕz/PJVi(sTNXѪg!`nT20S:e&5s7rG?j62}~ٙ挝QMsw7:ߜM}?޽— xods5ݔt54&0<)tRrWF8w<*9BG bBfH hd.{l77,6>q-Rs. sU0(SYQR1 ^'BvKނQ`-54x PC)GPP֐<rY+Ӓ"GmaʣXW[bdO:_Я"Ǟ?3.R Թ,oDv]*6wjbGݮӒ}rZ%sin2n:Ȯ䮏a{]&9y MP\riCpBVd|8Z)煕2Vk_pd-*'k;][~o>S 8dCeq[&)DtK6S*T`%_HFQEc!aN0;s}&@4`^gXgWs ˤB)rn 9Rtv'K|:Vd<;$S _eJn{=ߵˆq{ɴ>d9BaչGv9.NڇF-[ki~c̀-&P%cyy$PJ]K x6[_A+*5Ʋ#K2 Ƹ?؇\ 3"XO\H23p=2+jV~gד  x pIeɫbɅmUj5哟*=p0V'֪+O~&H{rg"tKun'ycV,ix@mDW٥L'ULe]!Y~4!iiH4]f7tHW5} J*آkiBq@^bKpPnBO q[fM@H7*n8?}nl{Ӊg;Lh׺8|nYWAc_SrDŽRkJNEϜCa.7c C3ՠ- zs6ԩi awtxڪ"Ab7ZoCU;L׼w)d;DV 8Hf><<鿍McM_&o?!STAj p̞hh0@0pV`D2JM(tɈ pAl9zMe~LiA1R҆g?PZA‰Tpd PWea4"46Q 1ًR &;c.:7f0̮:HBrT?Qo4#0`|Z:+a~NDc+H@J[]k8hUs a (oӫ'60:/Q:*V;g5Z@t2q1 x$k %B$Bx7g?^C-u%w'nou:x?Z'8G+a4H(R0k\ᴯ#cVj9Βd8 g\FRsm9zd"{77 WezF5z 8O#c6hvU7m<com֠հ?.SlY򏢸\w.rQE(i d?-_UrAj*Ql]ɒ>>X9)%{+RNIiIW"6  jy)N,؁hOc k (&TYqUI18K?5H1(Fi958I{FXޟOWTWxt"hȰH18W[Z! jK 0i"WNl T}̒e->ZQ&*iˢ4G=EeyIYYcTN"jN\rjAM?GmHNڣ:%)o,CwHp 5n|$IQ1#u  }/4K3*OͨFo8ڠYn\KK"Gq1(&njH(iI|j=hM{\1DTȫ*?շvy4"#v>U yO纛.Zf.j:IkhXhأ%9?WJrٰTYd|I,/ݹf6-f{@F1ZdM#V̪R|EUEI.^FВ|ڟumycW}_.S W1 ΢B~Hg*>YA8a"uj9Z,dޕ6n$""إab2dV"KO&o5%ٔ4I5;،-]|dUFx~q:@d`@qnM^'_t9{/C6eVenKc<Cډ6wRJ䫶+aOlu٪AcXMڞblQO1@%$0o(]*%i\rP#5bL$S$LSb{j Q ޯYDK553o |1?tTjkVHq9(>PaMǣawW4K"Gz^Az64u9wO~T>o'OxU!(}*.Y\Їݑ szDơ7Mg+)h9uq`ze͠7u+r~XT8$O6W@.(Yżm%:[.}#dXsR/'#gLJr=l;/)LfĂ3g`͋y ,157K5d6H>ܩy.}ԥ*'m}L2WV"QQJ]~N`ݴyc݌AA`;m5Trd&b'mTg{yUkïX`ߤj0(R58xxK(Jbc`ˉĎ2G FNjjQY1s,4`O^5)&ovJ`J Svs/pSpuS0?X t8& 0 Wg8F ~0]L\^\`tN}^H9#T.Q/ߑiDdw̆4D`t'"D"HCGPl^;'!41!BDDj*HՁYiZ42|4[v}#kUWhdnj7%2 KJm!!0сF)%8]\&F#R*&QUp4GRh1*J@WcvS {8J(j1cXk c<2NI5(t%w=?̪;\Ï;\*?>~|nYyOͷLG?U a 矿=a<.YS{]ʼne3Z`l8h>%Wo$>bN0?G?  ͊X⃞X8d9y/1CjRgr]/4AEf8bS`K *]Eq.2BX9g+a%X ?NW3Člu d)t$3ד뫢Og)?tY1?]F4ft눦"TFM! $a'&ll4d/ӻ5q 2cԓJuX&cQQ  xt8eD  ꘥2(|:0jZNYy ru55UÚLѮB(*Ҳٞa!A`4-m&XsК AA4(<ŕc++3^E#܋cT5gn'g2jqNa4$d(%L5,9$ G3 EAHLde I0BBGbM"FؘrAԂA'^jB@va )_[z!B8< |zg鰔vPcp` -{O.q u #HHUAFo@}u0,ѭ{BNTI*rh.紥dq*`8pGO]"(+YL9""*C-%mwbYf .GMٽOIR[t=IrT4]jf/nLompoO|kmfGwr].]NM5wܥp%%&>LDZcX*ՠש:U^jb5h i"DeEp-` 2o`A[Q0F%GDuUATwKj~fQ{q;dQ@.25/MVk G?a+$j//!~agH 롟 kw_#hc_][CTzmX˅rӫoo= ޲F!yѥ9"p q52lS F$7"2l>."`:ѹ*ɻƍ$WzTweulq8ó;OpْG-i)u7$ A[$$:^ 뻿߶!*~rCT5e]:qq׷/״r/g}{jo>9ou09Y !ۮX=E:eVt_/A`ѻ pтŽjc;(;M.;MT+MidOl&ˆf,Ŝ!@FDHK,40t")ޙdK7:UaLʲ&p"R58E S괢s%iupy؟\MX#`Z#o^)=a`/q]hNe{IrTI o!lԗ猵Sj3tqbX/t~ zmm4hĨGA-.tZ!0F~:#?}TljVBI|VÂ>'  rU0(5h_Lu?)if|( >90\8 2Gcハ\2ֱ@ @*A}Fj7 h۲VAe@U-_5Pk 3>j#8X8Y#rE@3rn$ɽ,[R ُ )V^B: #5 -D˒Ȋ͍sH8EpR8j 6Z!A3 9eEUsg\ {DTQZ(PZ#Iң{S%h`P}=z*Wn(Y IW"Lv=*Yـ[(;\A_ڝXs66 *רVQ6g6 J\!T1εm'rw|eN fcKS)R 'lirC+V+هm'h".V*iBHRcEx}$Ҁ$kntM)c 2Q+CDSSDQN2&模WfjNT *(Ǵtywb\?_u1dpo{|p[Jt~ m]~Z"XS_]y?=O&9ϔJ^GW+`Tn=-UQ\u[R\1>6X΍HB"Z%Sޥn][-%S.xaFر[j!$.e*Or f9|O&HsI=Hr=nyy^$\n>ɽ{:; 4?퀴Q엿.7 k96OfE{mZ=o$!}۹VK_ BdoGE+hMH&TdF !EP-=Z{j 䴛yB^n[.b*୑(H 329BgVYSuFfo)A'u5H)$m ,XR(0EZNIfMx3ZRZ*py i=u7N񒜏zT֡e<^e *%Wyժo``P V hQ\/)˭ź|J\[r6Y}S}xUb9G0oAv[l`l#Bvp࣪M 5j%%Zh S%#@a Ė[J L9A$<& >A9nR;'RzX W7:j,ۛ H{x]^\ͣ/s4(rݗ—aeFc.s4yu6l:%l`u@BkM~.)+pyFZ t_w ^ӳ9mNPI=l7ǛOڟf¨ NAJTܬ@"W4K~6n4K`(뜰iy)8tv)/Cey'HGB\2lT҆bdj[P58goO+] F#^ k[m:[Ň:m=6l~[76Ӎ鯽d 4 Gт r?n<&,vPp ԫ5ĢKGV`uВ+\Igq-9F8,@`JK+nR. pY$EjxZ5kkuBr2!pIE Y@Ejj1.Ojɫ$q4dJAXSL`wR$xf̵t}Ff Og5c~fCH\Dd=,[-%S.NjdP !!p/SAMYd6ۛQ`!4j փ  5p:1H][+edIO9%ը):?tմ,y.v*sFz̢sϘ50^a:3o*E^!mNzXoy~3ty<*_} e oE ZHKi׳_:9Zh`CjJ̄Փ󲝌k[G_7Lν_{30L̍g_ WOO*il&Ť28>F717:2e$M` j F> yGFȴ>00-ȽhX(6Y nx Yܣ(&d!D$Vvm9s"L9+ez~%^=(Akb~H kHRW4H[sdN|4q{x*5ma 1]o^x!zSgrmyV;MְG?-A˺Cn"E٣LǞb%]Ğ&z]<;pE&<(>\_=j'c3N(Щ`{z=[O< d,`tkq,4j˹X|9q o䔏%Z0 <^zLǯd5Fγ.!rEH|6"ӣ:+%=uׁLNpilSx}2ݐvw:ڳK}gUҟY aH,&$,LRq4U.r#h4o}sqwf./*򈬵yqD"Ez3]r1zPi=_{I BP{IuAG%>JԳzwÀFipSIS7q8d [cVآbEl??`_}#K\c"qax?6>na!ڔ\0/Xcty㕿og7Bgv25|`[+F>]:|? tE~-~[[Yv%9Mz}9l əQ# Hk9ahz.8nplcq I$~=}oDIBJQn 4@SGbR9f!Sz°N S<#XUa՚̼*X(s$qܩHI"%m&!FV"jѼz!z23:Ji܎dHyEAuk;E-W/DO]xa*U*ŋCNfx`"0n JnӋ@7xݪr\gL!a1 uu'N( 3fw K)*~nF&zblpČ zPWX cuej }ENSleKvI`Tu" AupF6>TNhVmhwHk( C2A{\>a`Yc6ގ]l#)%o^ܚ{=wom.&=6>m\7ӫ}deͼ]>0]}{AxV`XLOI<ԀYlږn%uԦm*mrZIm'QZ )Q$:SN0\/S]08S"Z#0*hAoYŰd1(Sc962C"Bu0Z)_W .*F c# 1(~oE{,1cq|a + ׌Jx)sS /`p+1v Srf l#2cqu."xE 4AޠbºسNPuHepՃg^ijvX@1d; kuuX]\!E8hgGTKA=++eNQEGps1[x%X̗-o ԂWy՛͌cr?-ruɌclK',7ap+[?LKonF~cl\鳒jd6j5‚ GߡC^Vfn~X^6U͑}쭮pO]=SqvѸ7ˋAuUP] a83+FnO+?7Zz֢Tnu&|"$S]S pR\{M)؞k2rl_)(~*1 mWii㖲UDchg=;^imؿn߼7WVL⒏IQf<;v?TRyBaG9xxxIBldO/|> ‹kQlvn"WlRQd&HI&Eǃ&J`/ыl|V"\ lu.:YdBrb T߂L !z"S(Yndzh$CAT%0 @GXπG2)* ԅ׏l]jց:Um'Pةw1#;x?yX,E-KW_I.;ZUZ*nk0J0FŎvSm}: k~˻R&"Xt}̀Y*[Hz5k˞_ܬ;hO(窹|- j{o{ﮟ93_'0C8pgȥ?ʱæa!HwHfZ#M'9y˜/eJHA=Nu0scz}s{E9|z7qW–^1]!G5V>b63G}c\ZPLЉX LSWG5ҢvnEth24M ΂$>LF5bdL E*;]WKf1WmFf1uU#nDF"aP~]GG gw(??$ynlk!36~ vQ2Ib oL_A30 F Xb4%@NiYiƔ<57AN1Vd!N x}bE@ǭaZcШzC2 ^J`㉣P`t%P- whAbH,Y^7ڄzJ "T:YY.!CVVid\hu c\,Y,SRGPGQ-qA*1ywI-Uk6N-@hLIFSi7ŀ-I}G6UW8ъzXOh=QՊy2mt=ojau4:tOKukBE e]~w^&#Wp rȨ9={gӋ{^s|R:A6tlBF9dõKY5&,`Xs4\Ʉ8VQLDnCʁS"%Fj?\eD 7w/W=?%XLnfwރM[(p@Xz2c' joF36w'yLKzvЕW #lu0  a'&u 7lLBûƅ>>~lXB@؟(q\i(pp1wN|, F=~*[VyK `c@!Kpd ",.Ht(%5[Y{ 1"y$NbJݢ{$+3b,f@aH03 憱sqŝ^¿i +G/.Юz.kӘа: }~F eh3ШHmGVp:G-U;՞p%bx/HDO%SƋQm$1v7ulARc}HŵK+P0߽K+ 2+QCr$vޙUUR{åACݧը% ڂJmi0z//lYz$YE k!SɳԦTe!d І8bI3Q<)a,4NUd8Y0(F"SɵE崼 %2h42 Ew>.._Y1JrT6Lw=S48RMݜ"\˖  (X8  *o9g~}6z %RRꊓP" cݡ΀ >8gƑwSf1(>0uALxPɋ}"yqyMwH?J5oosy!/{q$^kۙwjk'l6.y8tXN`M{E<2_.15B1Y_)%f<;%rpIf/c h>]y\ k!x qK@ư7G5P gqD+F6a{% 2j⥤J m4&C Q+b*LBgyrm=pB;]wNͷo§\AO;SXd6Y~OYx&m5Io&RSo~K{C0I‚s]1 z?dD@ay>Kn!W稞$2ȤEz6w'*/J`֎t0FHB%1 jx*y*@#T8'w3s6Qpl"JF{<2_;džτ ᘰٷr8 ~03TS(b/ V\z+ lE+"懐&_[*,M:©>đk(Rxxn՟ie*l>y;-hdف 'iN:'&ۍ#m5ϗ`Md JAOM"U\ԇG34B1|B 3ּgaYDž4I8z- 8BRb+~CNZHl"19"?OCť#8{F>uڪ}m|[s?C8!rfȩ~i:(#(8w}aG1ϟ7qVd{i{ fd69C$lrVQgODjA/@O?{.HwXMB޵q$2b3T/:=0l'0ٗ݅зP‹A&h8PpVT$-$.7 R&lsυ0 }' iXn]|t8:Ї"y`e0vXr17ߗ;/j|W0hX% r Ki6.XULR""cxIf "K"ȋzH:7d suγSoM<_n:٠,|`*7SOx(4O &e8cgBiJ'{&>,̞AGQ&9+ʵG)5iy|:ً'ͧpϏWWo_/__a |/^{ś_yutw}k)]y~ׯifNG"?MϳSITo+_:iIi8:Z E9NKtfGً7ɟγXwy F=lX%gi绎Oǰ{Xg2g99g+l hvA^#A,qpA]71w2}q94ZӏS{|k\C; 6U>zϋ~Tis?ϡ~#;هLd4 u6[d컒7avicdvS*~7RShϹ.dT[߯댵:i:i:iUkXe%S/P9bwT2)G-ON}F!#h-t?G%O7JmHX<;Sh2`@Oc\!dyN;Qx$N!%w <;ElUyN[: BcnG5dg&Bu-8/q;P/GQ^j s:&Aa}hf悐CEP %$@cT|plsЀ;rW!cy9w#uctDUL4jɓg5!.F1D$T0uz5(HHNl#4g~0 m6^G 3ohDX&♿,6i`-G"`}D*0 *!5rY8 t#شq.#9Ձ"(Ы!0bО`(<+5jQK{NݐkIa-vPA(GiΕ3b7X55ǒ?Nls:2 B⌬+,6 <wP>E)o\+RzTKDIfTXMVDtIQ90]~ۏejK׈zrŭUNŵ_/gY GXf}ep #甠 !Llggu ]L+XgތЌNOowbf7뱘Gz?kqZŒtq)a2S1čbJy&$HM[zRg)$܂FDg N({w,$TMVW15d,dCs[:ә&иO=\*,W;312*DcL6+/;Y'sF\N}P _O_JL}f8t0mEF`ƙRrtvsWq7N'FA0q\٧D*h)Hfg 73K7\ӟ\oj>0c1oh`H0W S3EX͒cY$,byfRϕC+ōW?9%{#q08dNб1S4+'7IيAagx(wv^Q WF]UNR^~rt /Fz՗t|v[`F'MzauT`r9/}w缨Gػ =@.m%=PLIn3'mB+ZuB> ye1'P [>HatN8DةN.Z&Q)6Kmi^,%m mк<Udט.vBb\]AB`Nz ,lmPPHAY'`wu L~u"%Aо1Mf?G7% $ e`bj`T,1k[ C6AQbT-&Km+5JDg( )~t˹ja%IwW wnvuc %bJAImUv[3#&J|A>"(G,Pp UAY`l]4EW k(=L.t" Z}"iٯ;bW7lV:A5@y$H%ku H@1ꮺ9=LuCKpu/zel0㥈i,VH*Kp]qYJԀrvD_%΁WC⬸ZFoAYZм2ٍq׀O,ER\V]VZ.v}Aj%v7r\u6Ct5.wA3׋F kR?\zX V8d,Ht cԆǗ:3'0/BmE&20AaIs?1Heܨa"O(+H@aAJT,5`hZXM<̆uYꂢ& 33ei03-͑R0ν`)4!#-n0+c$3cT U_ZXh9k~~D QOF_Ҙ= 'N1x|>x1Y|eS#;$ 23Nw۹3X]gm ɖG˵R ?on ,|c֥v9)kټAδAAOZGɧ+UL4o.8~""jd/cBKAJc2 ]ںg;dis&5N8gKā\N.4:8V~%1mp8\mӺ|%0gz`a=Ւz$سCC1Xɿvg:眠y^+>\iKgJy*,1!NCLPVqjU:[x}~wT6~E|V:ۙmsc![k,cR]8nF+}PJ)(0ꋢ0L-wk\鮭h爑9ҊPX76 $4p,f1bFPqA,5aq5S-b^rN[{ka7M$“BPYD Lr&a*1BLsx5X6n5daoOyIDD|vAR-y&QH|pYs~K׳7kDJM=FHu4b]HQ3\B][s7+,UzHYVl6xMJErM`"83,%);׍FїbL1l%;{UqjJ%&AӁ*cbGu"cio2tFo9 I/Jy7@C˰D౾_,g7Mz`<Ҝ"~/~qW #c rAڋTK:9 aXC9w_>h`e9 H hCE%=pli3 BR0ĜHm4hk# =O0.TEUVQ!IViwNJZGdv*{"&a=cz3Gg%2jY3c[,8@vXtht`SB5qu3nA4abP--vbsIA,TdNBp,jď`\9G9o?$!C`Xj¾E9DXpRoJ0`fXc/ ̸BcTfrLY(X,ctg8p2%Y ՜ˁZ:}L< mחm=R]Ӥnԭ.bzkTҳp~պmѦbM4r&:'v8Gw}[onͦe`7~|o#XLoobvE&U-n5MtvƟ|ǟNgp!Q{iK'**ҩ;LɾNI!O~uK5 R&CY!4p'DeHc_-mbI T|0fDFI *yc`JpdÐ|o0,SDfaGkGD $jnn}%0+`N)OFg4vBՐ8ώƠRlsSW~^/7!vhIz_Cs泛, ug[HS(mJpէ~,Ow_ww(n+]GPuX8:G4C%47Rc(0-NyE`ap_̙$= ֢Xu~HHGt6Q7QVYY f xyɒ#p\#V]~kRm9DFQbQF"kqk5e~&DaHޑ 4j$H3TR-ITC 8?yBb`Ug=/PDJ~N糸x,N"u0fm+iXgxFP{dMٕ 6zq 5]{ g Fn鏇0}\֡Bqڟ`6\׌Yδ[Dj&$䙋h}S]@=O beqh1!5gc)˜Sh eyoߎgV[N) E, 5|KtNdӬCG>#*t}`Cф[MCѮٿ4d%ߊ9>R\_ᅸ-xyЕ%t_l㨣Fa]ԟ`m 2'C3j I 55 6dp}.; bXT yaT_"2#eۂX@ik NׯAJhð>C(xD6X?+K} fZ*YǍ 4E=-=ҧ N|9^ywM'z(PAJB,tr̭&p9L h+uKN1ԘX zR5nQ❇wfS Ie8pd3BJѓ`g;Wy&4笗ؽkdx/FTI,v,PJ*Ґ!W$OqL;T[+衷(x0c_x{x~}=i+d쿅3רCrS/Gmix9&x8t΋⒢=OYE #KAA9F}Yp@m_ilv `& trj K$=s -GxVhcD8PRxz^lMۿ~H,(|V& 쫇iq9y2>joK"n2+ƍ&<\*C(qdD+"Bq1W9 hWߦd\5p&VCd]ϕ#?dvy7ϭ~5pm4`$nbR{E8קan'8%oqս>?ǀg& c0$ƎXgrͰfIG8&l F|O]EʒdW{=DgnʋDݙ0!dY;vף0Rv8u~㟌 nB*4ܬt6v@rl7^yHj a @e0#ri@˅#au ۋ4ͫ `w5$^[$T"m.w۝7aʨ1L2Lr{& S7 k] d~.-ۇ 3>U''ˉ]4 pC)m ў'VYQ>~3c3.QjRKjBi;a2f9Lvv}s.ÊGEWG |cbfjU0"F݇XG"G5?N2?/)'_oc/F&/kzk>:oHI,1GC6Px e|'۴dLWq}8,qX.\r*SEB؞nlW)x"T5͖w;nvܞ>B ǿMi_JtQE{0h2Ծ W L$3oq28wTdzCv}C;]m/vyQw]S0%^&ՖҗIӝ30_&Qo}Y:GyyyǸX<, )PngS?e6~o{=:_vBB:er eatVw; trܕbPK#b@qW\"rpNu|58E4y6%&XF2£I˿kf?Mo<2 p#2r$KA#y9$3 B>Z$d"A B`XGP*3$By4Џ!N|yGr7vMu(QfʆܝHHӆR'ĐHB̓@h?px\JQ wNn;O;Þ&l@e $WQ,TR Pi*.3(͈%, 6&xmeoCUF ̈3M0qʕB 84M@Ⱅ KCK6Zn}\ƱP\2/Ga\L c8a5).8 0 ,?/d #} 92 p d^Y)o5iμ0G-Π8_$V6:on$6#8F?vC&gk&m@1U/w Ăp%"i1 (ȴSKgB6視W i(M9+3HvT[6Tz9`aNpCB` x-WB~̉A.՘ȍb:(FZ7ɻZB)yN8<;] cGΖ}dB~jFUQid BNo/@D:FAyr3Wy=O)٬d7墮LXc_7SAf{N~כ3˕Zxw˹8qwjfk[;6?h jcE/n.&ťp$R+a;CS( H %Xw//OŎ>\$L6.d>Ǝ2)(sjUKȓ8<ᙊ*a*2ue]+4g6E-͝wǨWxLF/W[lņ-8I^1p+v4ZnU^nD9%TCsbO-D9ּ's?uBj!ƒ9BpEELJZc%`{b_mlUsle-*'1SR^/7n)ObÇ.I[Q"1_GNS&1m4"Ѐdǀǁ=>s} J Z qmoq9#Br%c$IEKXeN  y&MIo@$љX'<Ϫ "N6>.8Hu}# 12x% +5"ԕ2RUIWԲ A襡Ɣ7DAErFym[-F@xp!6h#W~^!L~U¤yx< ڈ|$ ӿFڃWf M2qq"oLrpO9W^HyieqQV70~6<>߿{?N^5b/Ѣp fa%Gi{xG4rOO&BD._CU[@GQ4v!g ?MfI\fsj{K4.S2( ʆ$ԠTӭx*>>Z:6w\#13)I%T[m)Vcy-T 6["r&6RIYuhr!Ґ&ZY&v*>rkRF,AAUڼ[Wqk󪹵4$R b\"ak, RǾ4ja9ÔC"wF,#}FNO {oƵ,"Ŕ3sGISmD%ґ}~Z\:4;\Fq^D[eťh7vA-$lツ%jGzD 1sD`]i reRs&tt9 pZȬN?KN?R+I6vQL9֤hϨK+\k!1)Ȗ&@wbKNuї#8? BBy<(-SK xK 7.bM^bn(M tAjQݧU,M]U,M]5KS@W{'ÞD[@ZCC5!sAqcj&e/?ت9LFsZ,}o~~cn]|ͤ a57n?ȩVXn=xM] ũwh3z+5ݲ?IMTw@HG=8bT7‰Z&!]5V/_[r :J/Z)twAC>XmLMa!Lth(@ZSDoJd *OE`eC[nі!\ SW }1aaӽk7^V}7s`@GSR¡siMT1━Uah]K @Tnqkd҃Ʌדo|xd{t# tu3$%F'Zf4.(;jE4!BH! b`U jxjΩg(LJIc>gf&3s6;DqJ% nr8&X 8IՐ:]ffRJ5:y"fCc;R2b;s!RIZƊ&s+ȫ  q"6X2sƵ%ʐ KSͤ+BrkF,ps":c˖ DXup}sْۡbQ;^ JKJ+#@5Rx~ہhkccV eY%=x:b(B!D(=VHAmEVg " HUkxD`'Л %8Xŏ&%ھ;'b/)(yS)}4L2c7 vѤgÚ ?'V}-o ݧw٫7SӚ8'_g(=ޮAۛ׷?o?F/hneMuX)ÌFsvt n {U".~ʂD>OMK UuIY…l}znc}FUɞ =Dh8K3j 5Dsر2为ŭhoKxIӘ@5`x;<@!8:NE龜^sc/8ޕY1Ш@ \g SlHZʊk 8BCL4J9@7Pazm7 r{aCr32733nv<39h9n݇I83*C3)xxu=DYZD).`;/9Bs!.'\Aqkכv|z륄X=ho´s&==7ѿF%"zx:Y{6q2џF'kttbAIY~׍DGm^:c8tSH1Ϡ\:Z JEn8JjAթgh\ܟ^S >ZNo4^]^R KN9x7vIԭI.gpELg[!)u.l<{twW4F*Mn8H5(>Nld HB┳S3]Tr䯪xׁ}%]ᨺ ҳG38NɏDh%erL|K~+PtpkmoI8W>Iݞs7[;d *Xk7f{X/J֢@! \X͸V ȘټC> IȒKbMWH‡}b'Z1Cqjj i k-4$8HXq= @-*-؁>DN QU{Fd PࡆNMM$~-Bۖ͢fGpm`LTg8̫q9|1Q"U۷?Ռo?N3#edQj<e/UK6 P&nַ{wӷ2tdd[ճ? (n]v}֏1_eܥ Gp?":ʩP(1bv*#&4OMvlMEȹ ܰB]Ll!$Hu['qՐSZQ3R m'^RVTGYR՚Ej0A93hI /Vt!Y!B$EpB2͉*ɒ;FK*q2a2\n1>\NBHx<9)Ql-V &0`BY^+%)~AVSS.0xdBP-.DX 7+B2Yrʣ|4ZhsԧQ2:?KLQғFߧl 2L`8M"U@ V1ٻ޶r$W,Y3@zҋ^z=~k"c{%9`}bb\t 1>:,~U,XWya&jsb7XߚĻ ~榆ؒQBJSCX) 2 p%@ 2t̒q VsƱKюp$c]ojz[Ht .2kH>:YWӟN:ڄ=y%("W5_]F%dG?4Tuh*DE @NY;8 'XOXC>ZJQ׹"X-0*$gd ,G%õP eol!Ť$t ,GEnуĔ:$~FT²+?z\ <~|soQThy@\Ta1-f^>]hT #NS{[bJT苜|N$aGe>xÔs_6$kxEg 2ߔŢ aDaZL6ҖH+lnkr5]nTZlueԀ,8+Qӆލ2l a0BK h G?mbѿO'q۱QHr$Z͞-Կ.W?/M`dEFT+P1 1_ƣЉSiBH+ ĥ '*Wk&(0mVVh[AZQdoJͰ@fdVbxD(8}MȉCUvƌޞَš  YR5cǔJ AmIJ4cK[eF?BV7Z Yd? t-ĕ" SRc]K#@iNc:jWʲ;&+FWJ]S>gC!$?˾z\f1rܴIxp(a1VS pPvc؛,ojyAC%; QŎD&+ =h΢ID .RL#ʧ!sDZiyHO;Tuh5;4I!(h/Nr'Fv l#s:槻hsjq0}Q&Yf-_%}uNo&fM苊w߬ny`R;_VKB azƛ!⚹lW%#D@W Twyn*O?ܜCdޥC"m*3!}T^jAT|^V13(;>ڹ8F m\'P\9H&QkKA]VȴKˎ i&YBAc|=|FhTVjҞQXaPteb|ωSsn/.9Y. 1|beoƫRμVOՍsj{R] "w1`5t@UJP޺"RB;"z_ Zu˛/2Fp>ww n{~ `{i~*kKBg3ߩ'zKЖ"TWbqӽkqmoQZMKu:N:N54_ǩW /vj!BSqEZm5IW I3WtjΈK%6jyv)Z̀\wS JikB_WjIJTxP5>r5rmi" G?d3Qx aBI&\.mw[٢ug4EvW4^Q9NS\,JDd._{ ,  +Q:ȼ'M ЕUyG_e_ӌAf1h4zwePER Yyelk{˵^[)JPUnb,7cF4gEH*(*v@J9i"!c|<:kȉVs:sRbު腇rB L!(.Sb3 ,z42QSk_J,YIFs'$M&mΰH ūF̢;\@ Dj n^cݶE89b99[b=2i*OkUҒUsMu[!IkOo0~tT a>Iw(bH1C1muh I]3*ԡ}٢$o[ikEs֖֘yXMJ8}ka#5[<EyǗ W-bB !j? E)Z4u$UBrQ\*2!r´` b"6@< H-䈣jHۚ9k3q߯ekYZmu[Ur658 PIi\߃,"4qMjT#i[5*y-ni@j! w*(8z:N"R [SAPvk^mlg£LZ|zkT-QŽ6IuP}yːE!yp CpW=Z'Nrl+|6ú:2ltD1c>6 nδr`jJ>?MQ30XE//cwпyi92[iKs嗇8[(~9uZa@k}6ݤ,{~ZO~P'տ,84KI{㼢ڱ37&6En!xN6x:e7ָޭ 9smdS[WC݌nbiF֣N7-QwԻa!gnlJ'e6axN6x#&P ]{` [r&zMilm=/#8}uq?_S6`g&?ApV2VŸ_C]3_PuWň.lEu~puJhi \Ŵ4` jq ^F0 =i2zjx bZFEMǕ|.w9aB\ 62?*WpÎx E̩fLOwje[ GSpJG1XXΤ); Wۨw-h&0AHAS-Ả1f4SBKJJM!x&5W4)œ$>hf* dq˼h[ |Z 6H3DN" <řJg.(fAѼO$w!|8| #6o*pOv;".K$_n2yX== O˩pƧx|{K*?+~ů:Ϗ(rLRDOW4+MnrƋ!_gcgV`=>"[QI?jy[W>NFjɛ-r]6[hsZail9G GM>G;拝Mͪ雼(Y}x9?pHVu5`^r4Ig;8>Lcda:O?}B)6gm.ЎS2`!X"0gqAHH\k&X0P  <%UH  oyTG-u6E,1%>EV1EOYT"tf1LʚS$NAڦUSԧGPX '< BkH0:KT1 .% F0FyɝTdb+Aː%ˑ%b(!9<x^818H<613 \oIMl>6^nn D/ "qQXJ&"#ɘ.zE'5y!DWn!4!ZVHsvׁUُ&c^"r7r}تB5V;b>p\  uW:2OL܊`Ɛ!\wWN>9mzÅʍ΋s ;+<7T6y =K--!* ΋Y.]hl q^&Ւۯ4!8y nwez%0!Z(AB@5hj>xUY)QF6LQ"\.lG-x4DmzƫuHx\IyZyg)-^2F$s" 6Z!qNc[\hȴҔF* k1tŎ}#US>n<׌n~VɘD8XqeQ$J0&rz 1#a,/$%'$%/b.n!%~ 'vDj!R@XLh-S^zoP~Թ2 *yCEv8D̐¶4YNޞ&XH/4zVl~F_sd&gT;Rh+ym4!8(9i8Z$Z@N\S*Y [ɀ=5VGXE*)k mCIFYGuK$4$#fZqO. ʸ' (o __Qlyӷɾ#:x9Avu[5z8VfQi0̷Qh!͉ ;.[>X+9ǒ7:XyNO dDi.@tLINr0‹S,S›V}#Eȗh7RɌc_o$:'kq`q"렅zf}T%AkKHZ3i@z)\єԏ 4W"E<6F3bm6/vmr~?M4ƒ{lܥ+٦]+rW+[ ໳Vop^치kvC/P\EO;3 CEr+u1:Q.%PAe(@RD;s;(0\&qۅ,+s .oּV븐[KMn>Z̞S˛mV qhi~jr[ucGƻ_.:<2|/'3WŅUIt&|:Yu\miOJu|;S٬r2^^~կXB*g4.5B5ϧ%YFT 8BLfWよƑW~_j^V.951గkhčJRlxJ j;Za\7%[KZ'GZ\"Ph^Z0J֐*Pb-Ug''rW~|4U ;m<;DLȾ3F%!c-#7/%! ݇pH>d.x~`LIGIu_NXs c9R|1'Z J|n07.9 'l(Ie"g sC&T98@xv5Kd1,5Jk # p!FZ7V&J(CҨ,7K`F!!$o mmvg7rs2V|PLJ6 k{'A10cvMUۉutZlB1zUMcC뒢*Uk}Yڂl-ndk =Ja:'V+;0k q[7__vŏݠ)U;ʫ,6Ta;NAPX: ֗QddUw)w(۽_[{Dt"UC{|$@7lӯ,%g3 KP%/kwF> 8&/> zfWh_ ~*kq2DS E5-%ܚ_Sܒq2?O~v3iw`ʞ3|QV;O2Gyrvە/o,cI R60vͽMh3$W@R@F(ֳ7c'z7YPDxwjz6+XsN͔"z .H yT!:'^ڻC)sr\pt%0z@ tZAQߺTG[&>_jNp 5=PV }|NW h;׌ bխƊ!ApJ_rARv 'خo$舲B2LO!,Ƿ;:a /5)(%?. n wrn^Q,(@/{AuJ xpwy2paUa;Awp=ԅu78\=*TMQUN+R<gVXL^ n469ؐª`"t7H]`u:tTSPځ,fmdb߾pdY5ghݳ?ՂJ&41H["K @ H~+:8+EWñ;,Lłֶg1뾧g^z"7.zʐ`3+I$䍋hLyXMxUb19h9OAjE4J;ٱYPb":stnq<%j0WRH(tmFtCnmst#a vj%7.{#\ >\=u+M*Vg9Ɛ٣>. kQJSTInpx-Ø+UXf hj{LUbHP!GnR -F 8 >|: e3%ߤ0m^hs"6 hl^‰B/߮i`%'X߼el;'CvG' cjL72)w9J.Ҡz\hjY=) qvO `ciZOL67v:P<` LIUHik,1= XO~)&FVz_II`~P)D q,H2AeNGo$9cw6q\Ws|UG! 9p˟•:-sf7#(* Ya2j lu1R"kc+@{}2T g˼XMnub-9++rʝ<?{Ƒ lvt/ 8ٓ %WYI9kzHICr(pfxi%VTnBpIk$ZkTt_T>ic.7Nˤo}`,VsOռ~2A~~\\$y97U`-cUf2b99w9U.)Q™QGGSHF!N9# bbZFlG _59Es^$:F!.5H`6]ɕRp P%}3(G2R4T2hPXQom'W4Ph ⨢W1HBw_/W2U0F(i4 1 yN鞿.8^q <~x_z`Z$8\ߝ "bȠ?~|x2-d+ A_wtf<onD`m E1 uO]C3>ہ8F?*}G߉S- d{χ# GBYll) 1.j^1*FR)gHo vy/`kxcJbPfy!/dGtGвO#"y>meP-9b',i}BLD>4L}tUNzp}dJSpW'H.P)đ Q Q3G,Q kZ ZDD#Apg$FhD!")S#8pf# ZNw\:ΜR"{MЙ15_()QR_Z ʼn2$3FX 1!,f1 ,‡2gjl8Pp𭅙IS 17xJMƃ۳7oc=T)j<'^V+yD#.3e `¬Ö@r#JhXl$QzlaO39#ffRXЌ urR*rƠ6CCJ]aMgWvua"n j|A 87^,EJED;=F(.0O:Fc Lb~;417YG"0;oBqP4*t`%q(X=U+9ӝ|0+ChuRrqZU^/UeRh!DC-bIN QJtv*qL fmz((Z+l; LϔB]ו-huH$\E7lպ.J .]Q9RHDɑV¥$'$q.ع 08kDJ#U@ StQd l_rכEK{ LI]Z"v l1 .(b}y 13QxTn9ٶ\-h!SQ$a 0_ض 1`&u R0ީ($|OgO- GI*F$ЈvǕWVh.8OQYAeΆHr{r*řdO2kD؇<_Fzx6/E;OϠއjÜ/z^#hK@Z|)es`0!C^G!0#K frqMu5mK H|w? 3i,Ts\rfuW7oq=<ل[t|u!rֲZ{/ ZȭNM杫ooz7@&M]A' d ڵfhCnO\cTxDU)C_ yιJ?ւbX- N]898;7uZJh$fyٸVfcIr=(Wy+i>hEj8W.Z![:^h6#R~14xWD& tx\1* )JI^GbkbX01u0y=Cn{!5R # ,y15q,쒋l0CAŔ{圥"*G7 qﲋ2"~x~?$ R^#e#<(Qx '\Y$3F2%[u&KAQ5 GtQG;n'ڐ\Dcqi78i:F]3xڭ EtExL!ӵxh~<3-H*$=we:Zz 7K_LZL~rJ2̥t]\luAzWly($~Tax ~ff\&7K% Scah|pI: p=o#O|? uQ6)$j%5kO-f)Nmǟ}ΧYD}F* $*آٲ_SJ WvȔ6P1OeP*Ĕb҇4R9NH/峬HJd;5AsUū~-lM)qwwz<֦]j)(s$^h;Sd?%g}r{_lm("4j.e@9"G^}LTPΫ%s"wʅ-cp|T&We5謷 [bj_,z.jo<,G{X c)N#J%Ӂ@zkcLPZdWDrzz)i6wH dHfem0{~[=C[-)QRcQJEXTӕ<&ނI.zm8snK&H|7eR@tvՠ$:} tIRdR1v:" 5(,35i{{$S'Zh`4`Uy7gZ0LfBh"לuF7T^ [#I67l7zvR ~* ! =Qhw!eP&c Մ@7JԻ9\R%cԮ{ºյ^+`r o3-) N(`2X3^bB#|·E*D8J J!1Bƭ;z8 k⒔;.!J JJX **B6(p {8c48M5Ny 0eTNv]#YO ɿz!yT#|=͹CH4#^ps:1[32һ7_4Ȓ4Wk{0da1ƒ!B)z5]nbBZړ~StEêiU#`)|yýj~́|vs uqvZ<{hT(I-ݝܹlvWy6t`Wfc(5b%zk=ug^p%\űIHK^ #jH)+ 2jnxON`WR)n$W 8GN:8,Zy@)4B9r;tAV]LfCN^meY3b{::'@- AS nG 06[m+˘BZDC꜕ڮ}jVj ?ЁOԻGc̑㲧 07H+# ì}vL7X?&:N{KLr#$@Y)N}6#Έ\ ~j'KZ}5Q2Ak%"PHq@k$Fu*)]`}S$%͖ISq.7)nMU#!g]5nyW~JKA|Ay?77w>:}Ef`hXw'MKoxGQ_=^=><1ûp`y07 %;O>K@bs>N>2H%ǃnA7GȊWG#B"MPo4>{5x ü1!u%K |/'/1@m ,nB )#e>OHЁؔҜ. 0P `J\БK';YD񡼸h ]SC 5:iǬIRը2C:Ӥp|ʔG@je.G-tӧ[E_1HA>LHc<~nnU\s(t*2ʄuwbBgguT9c׵'f P庸c$ y&ZdSRލ6XJ111x#(e[z1,䕛MLRg騤I*;j8?/3.)k3q_ճH/wW^wݧ8\]oߟW|3yzf 'nPӘP/`+<ޜ`􋼽97hB]_ ilh_j»c%C.9YUrUP%/[HW3@K{WopN6G8YFvt;"bD,bSi#3kK<"%>=Gn*}0% GdmLdAtf (3cκ~me+fb`:d1Y k]d@d2`q`L"HeX;/w#66:՞)#~bQɢ {Dz"Dq4zŬq #4cNbî_Q\)9c9;$)i,F>+_Deōa!D۔ 䌼=DF1ES%~J>ɥ=~lta=X_=Y>fvD{<&Y?^bs矢gѫufE=UH$˛Ra1tS}qf󺀵.ikWmSIKzrU(g-zzVS˘5dsV$G}OSQonG:pfEXES_u_>㳟Xzw}kRz.#bug Lrr(xkZ_eW#xXǙ]|&?$e>ZU1hF {l>~Hn3]"9SE&l&GTH0`ZwSփ3A=`hß}%l0QYAgM0QX 5w`o`l$+j++IA0ug`*KP3m<߻lCWcɺ>6!cl,rPǤi5\3s8T9{It^:PQdX:[[oi OP ?9011 &(׉wV p`o'X &oZjƼ0҆9IBYjՆuУZh"'LJ"R(o{eYN8jQ%)7ޜ;"O1A|F['QXaڤ$<+x-uOq9CtcǨBΩyqsj?H L[vWC뻇/Uyo`@bp]]3矅pK6vU07eyeѬFQ_}z%s}gALJz?|tA  WnbA8S)T>p)7p`>ENA8W\ ̩ isG .>vji%a"* l|&EAŧ'GaJ[+BomQ o%@5k0G5Q{!Ib !RR GjͩCY@с-L\g| T7åƌv.]#p@%Bq_{RKnR rKCp<ǘt#?5!@ I0r-)Xgx ޭ)w;$vڻ fz1,䕛pDv#轚\ >u7)20QFj^ ʎR)Q$_ʣL":GaBi&!*t5=d`i҉`ڣ3+5UAjWcyU 5KV5@$uvR"!昜@}](\QX?fic*A 2^;Pj&V7^#mG${Z(PоeH$5( j|]Y gqa Hw H3Lbdެ=1/(Q[ 1_DιCW%r}4r._(0`4Cc΋þ ViUZK( AKad{1e=`N+%h[A0,4hF8 2W\HIi"yT!2gDHROS8^(q{N))2ĭgB$+ǧbؠAj8@\y4.$XiKa`btC3K$G!0SF! i(p&T=Ѥ\_^^g:A7,'>}~#.GQ(&8 '4yAL[FziW~ M{N[@\h7gIٿ\\"?-^:L+-TsljFɘvNoJҾ9?gJS 8 ODYv)ur` (ACx~' LLΰkRzъ`ӡ3Y,E;z1nsؐ]br b6WMr+Vκ8+`-H\=4J@u֗_jp@P`՞IH TJ x<ՁVXR1M|ù!1d1GlCshOd_BYJJ^;}dc(]sL$0>LCIq##zX"6XZI^ "AzYU"Y%oEf3齲p-HRN^)*Qަ`̤E(L\6 tON;J@,N(6tKԧ wjB+1wػT^q+W_U> h2U#oH5h&APSH+R|3Y`p20_OwYkppUgg>XTД5 Rzu^YQƌ "N^EʑԼWabRڲh^*ôK;R1m1Q[} ^ eND1H[t"*3yX2#4wؖ/2ER 9rlީKղ9w訖Dz3fMcYbT:s8HAh0{ytDu{APZ c{rY5T U+V<%J_]2It0;`$odj%,:Q̋a`^0P]jw,F Y=˪P\)Y:+{[Y) I\8 WLZYG[AGDgR mSbWSP,Q.+C21%A! w)5_e/R%ж3`p$w,3,Cb6򥞂 vJ9X%|2B&+'3; |ϻ|M_efetEcJtU_ը|87* 2Cwz"Vb2o4Y7dhO瀯n/&qECK#Qv`ccѨ\f$ybC"ܐd]E"67D#4f37*5#*" _p#(󩵅\p,z%TaIvyGa/~&ُꕆ'9Ii1q(0R¸㎾H=åm.,zBNf:K9 . kiCJOMKK,\B޸ٔ6&J11>D#4IzBu2[ y&ZbSɛM!bc:} nG\]/JdFMȦ8zz3u/3w+ kb,cPv=\M*X7уmD܆Ta=nZb ;זO7^^~l::ַ_s7gxp``|#cELs lznh5Vu?$/sw|C%pI.)w@|dᏇ:_^=^>^/e85ߧQ_f'_a jXIKZQe%C}uQU9i,uØ5%zhydM hS]ݝctdb>bޣn]"58j;w$Ȏ<^lWjBxJ|OipD߬G,1oEHJ7O螳>q4ܷK^TA-~wL;KiV`Xs.06y֣LZsR`mTCJ-+y^#w*}N@NyLvހxq 9݆&3MY 2EKBLX,uݧ֝6#\Ij8JWCѨ>m+ >z/q2ZAQR#J6Dx(>-xRh!׎MkD  <ʤ5âL3A(A(<@Kg/IEKfKeYGivYwu5 t-D /B ؖ!4MBH8RJ-RCdؖA*4r)6CKL+g<-RFSZIR _M}9Fh)AK9բݑ C=&kK)UUO+b+jݯ:7Ki'9|V\&8\߄L)/\ *0U %"{CX ~M`m_8$Y70i`,8UX$"ʆ'mi T҂I+ȅk2 pTFB[#<b.%Z&:⩷R؈a9@м^\[N>eFiN1r ;) ҷXA,V)6)ˉ hKmlʋMB,DeJҤ[0ܿx`pwDL\#9FR$R6niIZߤdJ%O&d6u2%)Q E R{S-zdkڤ޸ջuk_#gj5&DY5E$IӢq*&Ky$є޻ /X)ݏ$Diѭ'PLcD>8\sddF"] ?Pv~I {>qҝW9 t6>wu>@NFլ|4G:}^8?vi#"=53:1,(F'cLʫ,[ǜY)1#sΌ=~7笸G!ߖG);bf1|8x}-!Aڭ{4:!;w  t=kDAx[@d,. Lz@rdLEUK-uez @B^]] %pXg TX] ovh6~W%kYJ@s1(cK6xgVDiO{&DZ4L-O) H49FSUq< ݚJ]5l).^Vd Ƌ"n'$,d<:`$2s~H"ju֍/H wFЎ+?~ Wh"l "%)tbh7oJEJaq@u_Gh2obj<3 5hx@ɚ]<;\+9ک7K'}tϐy#KWl獻\F&D] B!-K[8uH"&%EpJWs"P:L ul#VA=WJRx}fFN۵Q3ͧm5 98KÔyFf?^ YJljk92#(1lG''#AR̈'=HGp8d6T;_:΍p^G+91yѼfQקy2"ӕ='AGrtڍ/3`&$j*QI8kQhW-;y\7+oj7{=Xp6{Zox&l TS^; =,)Q@>8g>AGRx#rEF_^;uUZ̟r) MbrZ~^L@i %h W$Q_KJ\,kYzBLzy/MU1tn ]CWݍⶱ $|esC.+St0! x(& dJ_܎%ȟ.]WO/J;zwӲƅ "-%>ݵa[2ӿ<,xx:~^"Į細m*sPU!h}$-veKS+JK9pbL` O9Q<$x4Q" rXOpaztRqץ?L`{b ^MaH^_r!!j:c#3f|!pJM? QA5 S% DNEΖ~.\i]<+m]n_gz L&x^)`k (Qaf:MTX t@n-}l$Vc5L&ѰfK![3m" *ƸV0פ7 ,b$_7vcl ;U)ᢷy8œ39e}ku|$b&M߬1cRӬ7tr UNWXPRbxOY#, mэVϋܧ&̯#pJR8~@w:N~wcL+ USvzvȣ{qWp?/=AZ߷灝 %ro<-FiDr:AFtޣأMc={`M0#gO؍cq7ڐRB NH j:ҟb_RRCm 5 mvr8u󕳤ɤ}7^yi:2*Rk"D|^%#)P@<Ն_r\rR3[cў3̫K>-*@8 ZHL/D>p{C#aI4E>,-W$;B1r0^ڰExFk0Btuj I&b|XڸۻUVДJfF‘zQ)lƩ!2Cl2 @4OhRj @NYeխy3K":@ݵ?R]&; ۛr|A@ɱ ({e51 L%dŬugA//At*2%PVgPG2 YUq**ˁN$݁a\& 5Zl1Eh/_,q 9Q0%഍̊!FZIdU)HI&*Ur)֢v}mH9JttGl7g^b =1Z͏!dVR+-Dd =xy)@"M3BS|7Rn\Pu4\ P1:#/(j/a[E(KXՒkE { {TlIAvOgBP+J8%Zga{Y!bh51Q Q_v3_\IgTbu‹6?]is#/nDՉ@>Xʝv -u߬G,1kyZqb͝vg -Feyϼp]3&4>k7 .j7#=$|%ba&W^g0w~H[)iHHcp++fpˋYŏן~]ܬ~ˈU ?,O"m1*Ƌnn=@<9/_L/]cji;֏ê;77~9ܵZ /\4$Hđ!FǧHt<XAe_G躽Kvy vCu1kU(ss(Iy/ڟdbѠqyuCB+RyOև{b+}3RK"ՠsGg. = P:,}Cx Eor. )Jߵ7 п6+&2W$^"N ph&O 2+Í+riI*RPuI d$l[vl970EM}ukSgPNՅ`_,MfLǐϚQсh  2g3cbA;~Qbx`Er<ՌI߅,]7h_>?$%8GDP9tÍj|Fk<x(n`o6W7B}uv/yazpqןuB oa'gQ|iI%\Zy*&*DDS;-,hH:i)ŘF,鮄;59S ǞVC0عQ~Eif04YKJHiX*^ƙ~>9 .i'?NQ5ݕaD|5aSs ο8zC4.b(Uets;v %esOIO2< =>#GZߒ8X~D-̃\=R%MR~S0rn0ѧl-mǢ_<ĝI_Ij9:/a:[3~9:G<+1뱽:Մş)H֒2#Ĩ*TnM3QI P(~}=mB,1VYlkW R08DZ5J89 ֟נ&K+)ԓx%QW^D]yjJT'18%ZR"WUa-#5U)KF2HB=c_Փ8[OF&Dj'.-m|^*A{w KWPG^pۋ˯溼\ү廷5 < Ug%*Z`N2TJ!eE)aͬRQJ&ơrL۠čգɭcYo% -w,1O;uMA.5^ 2k1y/ŐCdy>Pf[CĖ?OW8XU:^W +%5"T IҾtF8]N{ib.xwm'#Q.eeaWݥaK]V}.KM,T W8mH}j$07;@Q NS:\u" `|R7r19l5(F UhUw)C\:ws!U!,C0tP#գ{7Wu}nQU߱_{V JoG{ߠt{`H-/lOьtf[2E; ;76aHYjMٕB8M:zQ[l9{`z|ZBzug`墩x" y"%S~n6vjX BD'],)è7_:P.DӜ.hʒCʆ빀ꌂWr} /wQP̈_7nj+~x*J e*Kƙ#O4{ - 暺?;ѩVV&6mTѩH^:~@)A$,TƕED3Zpbs4i/YAX2H4Nx'Te!㴦$YDp#Uչ7XֻcwM/-CA3Aiuen?=H5z̢zF6r%{hk%I;v9;2¾ܱS14{ $e8[Lث%D\ ]44Я6&l8UKVu^-hWb ƃ[(kVHxЩ*v.Fҙ(iN9m$ ;&>rSBd$$u3R:JSʔȗ)֥0^H!eqT\I]膕g+ĥ%i\WRrd kSEa lh`fEI5QX;peǥR ^\!Nue`W*'h]ZDUi"Т_{6ZKQzXPx~q×o~Ud]Stjzws},:GP0G'zW8fH3`툆݄DX3ea`?Ae+X_[J[Kau¨n7ձ VKzЏk$ P rʖ%"VT!)@K` ;4`q@K1TSa0,˞". ˛#0|TOz-i?}Ob?o Dqr#FۻQu=?,!z|#6sSL~)< l U%ۣ۷ vol(/d?pwDf"*FE|FazIe֩&tNT m'TPBA KJ0d=+-ߙ]Ad)!JNF@$!b&҇HKvO$Q0'BhMY]%H29]Wۖ1R@*eE|?ǷfA~sB^#hlQ ©;̨eەb9'v`)="lƂzcYTh S-rpWfEAF˰"z{j!wq<㙟.FC7yٳJoTq0 `⇕[?:Wab13K|S=n=%!a8#~`X|=o./1WN_jKO(?xKݻk7}x3_g?FuzNun6}1Vpg )N ES׽\拵EPJ":ER4ީ/W?Ii;Y*tȽ2,s|!`PCE FE`QO'DR^y!BxfJHʨ4%׌ .#12:.V kILH,,n4F/?^9) a } NK'ĉ":@si( ;ǒZWI>SSy<T x7TBF[XvQdF z_c}u@1<+k[AHn@aٌqI1ÊKYce%tzPj]L%F>YVVU\IWbl*+* 6[YVhYƅ],}=xن DZNFB^m ֠.XfyAA_x{nn ?W7(d/&1Q k DNٝe"?_W9|rAkMD| CqGqb^ :\P e-`( e/ʠC[.HGJG`g]r+EL%W|,SN($- Nι |3aCU΄ `)%tC {]4Q1:_2 (N&ԉ`)eE`9;9mpU'iJ #Ds]O([H?in{Hy}8t:śdh`NN$!3+!9-Bjf_@ >ˇ%S'DS7 sj2žF} M7ijElс<{YA ye*mFe6 ;Af ϧ~1D 0@lE]pUApukn z84|AFpn8;G>2py,[M! RC(G{[ #Ɇ§#ϜiO/[ ]yO|WfO5J NT?wzVLJmWwD(A;R]c{{P3x>xt=_rIb&' R 6c+ND)D;[m{nv@uR;ѝR7JP,܀%趁eT1/Ho:a?Ɛ&nIÚKj4pұ*u1>}UUpU?G(ut*Z(>KkU J̱BJSxG(O'>Z M "%y g VڇBPP\!OAd@@AZm,rqA})2߫EH"+r ƲBJe`CÖ;e Hj˒"+ H4´+k #{_OGy1a^dt|@=qzb!J޺1 f4?7fp3C,CF`Ά0,n6R>Pe[;KmSK`yۦS i0vg6˵IrIxwm'#U$IJN|^rw].c]KRDgU'D gBSm޻ɚu~_c5)-MχzАכKˌɨ<̮}g{7~6w݌v0:a~Q\/zw>r0}^m/f-[|z!~{u}gݞ22e oBk$rzϗQzjmmk@MH'abHpW<ػܡ <6>A1Ďt7䰟QfMc"w*-QTcv. .qib(0UȂZV,P*km#G/w{~ȇA=,wf?Nt'?V3=~Ŗ,Qbj5*zcRiE]l>@Y/ySajbzCunnJ/8|UW*30wbP^x_n_ɺ.oFNɏ99 "Nɩ%S} ]"TnM6s^#oo >[/!CJNP<PzX^4S1)U:Fj=nK:nzFVlBBsM)ij$!G햊A褎QGuӿ4ٶvK?/Fj&$;X>>7$As=uOH^ `@EuomB{6\ FBZr6솘l 1 {<[MXS3ꜳ"1.C"!1 hPG^&vZxPr\_9.SFxlNplˊr\jpTQ#owQݩtǺs\pK5,n5%qg!l7_SULiUL!!sU(h_{ 5LM:)cX(V͵`j(CQd޿o)s?Lṋ bmK3-:lVSh;uwSܸ_rdwCff5ڛLm,#<^tHV fJ\9p+X!GX!_2jҽB(ۛѢ3&SOs9¤{w/Ękx*aߟUps~lQx0S rS2%Fmn*=[ s(Z- ' ~|O1++~ i{.|ݙy[W|y]:#V!w卷]\]xoԸ$ql'ѩwz@ƙH) t>pD~_>#Sn?zOo>l~7\yݚ`$dy ̔qV t9ⲠX\uǼncL>SKĞcgVB0qg;{X0!>ys wVeyf˝۬ :+Y$I c|PMZ*['V%_1R d-R/'!Z;23j!Nmm{);٨^6nPTEHl38 \~56o3ྚ! aݸzlJ  FXnkcԶy ᅭޮo=Tw:@ʌdT~Ocg9yF-a6lT[UX0R`Ti\ IL"|uM,ޘ,gSM pjFh8_#O}' jRSx%u:7!dkoT\Dϗ)Tph&LO &s ;ЈA;mi|"akP(R fjDK ,`9 84?4-8Ǽhވ@P-F|V{IQ1Ƴab+8X乂Ug&B 9jh/|JiIzy2C@,T|w;ݤF=A$"1_c>,/{7hTr8U_uN_uN«ΝFZZSLFeI! <+pF*b& ɭ*/Y^`o~>%ooVKN(u7{q~]"v-}&~fHKĵCo \i@8_i*qUīĉWP%tp3k*!)IqXN:,rB$ŋVYnkYRJZi-0]4 &9r }:-l/% g!`9H1TP.I^88XSNK5τb祉< a嗸QҌ1i4*iQQS;ƊPE"VT>#R$vAb(F #rDI*i  "&F!SD4eƹy+iiK)pzҺ'66Ym|OCMXGauj{rGߒ??Xx7wGހ anDbH/}w;|\a6|Bo?wff:/;wkd~~=h+a QE@0B0[%+/]x,фkv:r/J~9Mg/)GvZJڽY~y{q̯fj=7#*^N 铧TJbGSD"x`VdO3Ej &U[ 0A|Q`4M\[ͼqKˇuIDX1ZPS/ ab͘, K` u`jCF9$سih1_ \q:G 2nZOC sNՀE|hJY +?oo,cАBGH :FR36661o4zM9S-l6SbS=D1Ekv6nzs'F`58F;_]N#Xx#2<,n5K޵ NjJ>V=93阣`o֩]YU-pNFlW*&%cҢ+ͩK>~.I=qLR*VEF)C3[ .Tyf *8TSrImQ$urXQp\)?\zyGqdT^n>]2} M'eWLu[X|pfB(&!N4H^C*Ϧb4oCOSF{18Y*e\HMT$n֣?0p~pB5;NT(htK^adeJ8EpZJay=^kŗlEz[Sߵjm_aqf}5:N|b0e_F8 Nԕ}v0\nTS+#`K$15rkN<'IFl(yFPN5G`$/DƜUb3-Bp#3V-:PR/$ŶS6]4M()Ï[<̖9w/,7WWhù FIɮ|- ,wfST.F];E~̴%ѱnCa!3^ ڴ,ІQf^N%^:DUorua^<ry)|$kx髵sȉL];wJ y۫4X?Eu;sLSFP/wwIC}{ I?2 )ՇB!djE{"SD _8B]ZυV&ѶjZe$qLDbG+R.dsu&˦˛KAGC3rcvD36NfHN)+o&̭j~P3<70#7 n;<#P"9H1. ) 6,3S;e 6ǹ. Bx1ۤ!L!,!\# ajXc91*_u '_'ErRPjl4c:Ǖ3\LX[,6S6f>Va2k˄$Ynˊ܈B eʄġLϜ{V筶Lf>ةhsTNi._n~: qJbjM"ג4O/r!e7E|ݔTUj([%+U5kC 80v.tb2r0yfx<9l7 !]daÚuF1B \М(,v qf3ĹxMekHfs ՚gΔ,xZ UF 90ol@DF2$%'QqS-~aj,ZJ*1(RǨS-UypgRwj)MHw.)2ŐN>n*Qb#:cn}`δ[Dc[2Ř& mqw*v[ssbo߼ic{mb+꬧UFk+2>O=*y!ëi{-cg6}2|fn?w3uUrl3hcg7_n)㳻ƫoQgȸD5dy%4rhxs}n{t>쩹c#ͨkF1W5jYmŎέ9 dkE2>^dl[h!m lj*moњ c[*c-Pk#*rxͶ 0 oB2u(NHA/,T&7 ЛPkp](4Pl,G\{r}VFšpM3sʤd@ǑbX8f&W|`.ܰCig:Y8x[YpM\ɳy=$7x\jΩ@! K3gl@SNud;itH jP2 a Qt0*T&WMtR@xjEKU#1014p̟ †(O{8VX߽C Nؓգ<;ec+^PϟTnZSZn|БrjnJVTmC쮺|>!m?_;M#*Rm?*aAaWam2ӮfZ:_~S%-۷UcܸmU:˵[b1|{nM%,K6Ex}ʻ@[>lPv\ eaZ1eD4z7Pp;V?~ŗCߏ~zyabҵDz]"贮0a%djŞBĬUL׼wI^|o$mB>ouܚߋy )q\߱F NOƏ:U~ui'|fK鹨@d(/Fj HzNAo+mnE/[xɌ?$;{4/xHLtcK$M; PDI"6Ėiy<ڊMbR}n&;^D'!#pw#=%vP8 a%}5:P-%^mNQpw$)~9`A"R'R(RԀu\hFlT:CnnwԄ@03$IȅJ5@ƞ֒z" 눶C[r!WZ+:BWG+?𿊠Y9wuGJ+8KW9dy+tb@ ,=IjyG2j_<5tD'5ӯPwT͓ŽC'qD3FПt=m?7Yd2, - a `9J¥BIqPkMSG, @sQ pXQ#f~cq(Q!K!#c]z=?Q>Rp(_ݟrC_\7[pV[yU*ƐpM-)QJyogXŨ0Ari 5ߎ'ΰX܍%Ycڿ6sQba1F9Ѿӿ|;z/3|Ƨ #2?]UF@LSq9`ⱪ; ư$CĉZ["dB9Qh&ӑ5L7=N/M!^uHA5:bz=>qW]ȈC(<Oqx;?Jh < ̂8Y፟gMUfFç[*uRauV ƹϞ0̀ }f6?& sЪlhARyNs|.=t7Z⢒5ଡ଼! ^,735}꿥+N2z pQQɂtA|d/o1-%2ӝL# kY+Ev'k=[ vQ2QyƉD=]|K\7bkSk %8'}A )B7ѓW*QAbF % pC-.H:~8rAc˓Q} 2ǥJ (K3'SSOƬX,e2Nruβ3}''˺ IHIq+<!c})qMypKæ;+>x=+Pr햇4 \޺J׺namn-8vkY^(!a熋!i-s/NW8 ?;wpEXpTf+Q h9O$h}(JK}NrWB{ixU0U*ZW@(}vsuı*=UxP 4$=aV,esv3>;KʆڑgM/Z,kd UZ ٗ~éy٧hJn.ٔdPÀ% <`#P#sn|.n&ѷn7ϮQq 1_f@Ac85m#k2#•нP[om "\ĴƋ׹^z{u^gsq(' $PBHA$G!h-L|i}ЫчUzZY.ia۹NfgVt桿ML/ĵ^S$@<یU]ƟN.vGBeR򑧄#`6FA P$ $c0"–S0B #)5_ŀ 39"kDjA Q#B@\+`4!$q$V: FȌD0V5$WvIb]*d'?ٳ"޴Hl-9hq:ZTTtzc3uK% UbbX&H&@F<2J0mVnNTƥKv=޾P2vRl=4SjtK Y-'{Bd}mkU"%5Q? o?},G3o1 IX<ݭ.h0 cl% $)KW2Yq124B!͓6ǓmliL?=/޼NC(YGH#4hlj_ Su6CzX'*xZ\ x;a1#X*8lϪTױSZTʫ3Ձ\B;$%:Q!fP]|S"ɐm:h9 D J,G̶Eqŕ;qP~"4^ނT?FQeoHHU&s3 g6Xaܦ644ė0g8 ކӠK_>; ni8M<9!9Y΀y䅯Y{y Xv] 7]zhGܗCU|ǝQɻ"i%ciq󻷏3? C*h ?lGÓa$qĚV9B'&jn i2LjKYgX@Bv<_R9nf8sʵV %fWe9B0m<-;Dwoܽmp p$ : HP$D0D(ĉagٜ%S/dRO6Qiöl5Lv,?鵐x$st`UuuFf*XJ "4E1' D 4a ) U4'׳X߁:|\jW $d8B1'j@b '@$L t6Di3B(m NzK5TM$+X[V嬕T7VV)e&w4 {\7&pN˼<2+_ޭùwCg"/ZO) ̘Cq,Lg3c `J¬bAI05"3 p"" +#%h=7QتY.ǩn\K8+pa%} $ @+78$rӜ0<6.3A"!`&4J:d2J^t{9S~mFMvG:QS8>;*!?޾y8s2k~ wc3 +o7`Y憈Ofg;0n9埣 L31hx 2 $mAp_=2 mGo5B$08:3# G"U[,b^ފ(x (G*fY xCf`bU&B>};*#mniEm,_@ao!\ˑ"/e"ny/*FNU%N S٣>.J(grZTMwבU:%Ni]0A[,MZ8&]_)M6ݔ&uص4ڌJ`y"nQJYRJ=Hi"Ϭ\Y+%<Y #IVZG6%y%jڣa2 E j&0\j/+7NW-F)½NW _u#].tZ+*C$V"i ` BǚSUSȖ c ]&,TuXA?ΞꦻMBoۓ"צ`H!P8d3s/IDA+{E" H ۃρ&4f$zYM.\0PĜ͘I6G`Fve+$i^f܍sN#W܍wvP3Y*==fұIٍ?7x`NAg-/UZȕ7:mSȻɬ@ޭ..M3xg-Vhjwk\y=hSXu"u\p"_,^Dr߼)̌Y4lVa,)寿Fp}GLZ͟3f@N?ؿVti[AdXǠ:oYM APA}[n+JU1sZ=D[ԗv RnegPp`*wi.mƀb CADcbFg~ȑb"i7r{prk'3sߏd%K!NjHW2Ze}޽gCq)hÛD+[946*4NTPIo^qu VPUh|TU"8IgJ1ΔUsl;3H5矉C#1  P|>u`gh1r7ʢr"H(ב"͍$th@[E3oqlfkj˒߹qW `V[z1nRXC`+zJ z5R+Smk !g(ca 1+At*dxufz^f^7dsffsrg٢gMcR1+A侗Zcc]q>㖲A,))Qb^XD2RJZz.%RUb+/Dd4ձb4j N:ƎԘxaE"^ſ UqI&]M-"gdsxjyx`jN-*@vȹ8+Z]q I*fs0spiU4;Yɶ+W^Jږ#'w$VeZ᯷7֚˻cS7UdcՄu$Dz6%#**hʞR|_?~!N~w.!wb|˔b䯪&4?DLiܥIs-ȶS ur=d 7?|@7K<_5UvB[+ X/?Q+d捱?/dwSPs41RYd۱T-T)-l9nZ7F(RZ$DEQ40(5 ʏ@Nz<}Wm+09b_ӽD#KAޗ?͋6ZY}ٜTy7_-X!uJ^7io u%͝ a3/>I}3S1SW(4'v\>=19sO_=1Ov_zDyDc$]cQO,;*RWt`4p\ ܰw, !pu7ߓjM:q{tvszԙ K:x31Yg}w_oߓ~춝jmHZ4OlEզ[N&kFdأ9g#ThtNVhbOVf/1BEn#C7E>3+UТP ^]Œ}knT2;g9_y -o4(SܹH<N^Z9oYT)H/gleΝ0Fch&Bff?E=Ιh٩LH\ges10k j:H~5]#mF4/X(wru^0@\ŪFLj ]>{?$C{7oh>_. =L8ݒMt7_P(wROwMj|WkeT8%qݮaʐ.]xNZ? 6YK*]QCYPӷgIӥJ՚eN.5sKĚ ӥð 2&pDzˤӨA.MJ)膰m3-sG|ޭ8,"f/iIB%Էw,Kf(3dOnCH.WX޹?3nΆH?97صV=ZdSn\Q`^Tf:qRV3c=rH~_D93E j!eᩆF4,|w6Bs,{il6`:ZwFF/7ư!o^{הfֹӷ0+_<ԞN3:/ս;< ZF7e Cg3v$::qNy{ d`a44#gh'Q#a䭿IU9ͫyz1gk>HnlMC:(8$f @ƅ\TÚ,0K[ViLS5~Ql;_%V+:Wzrƹ>գ2AUKPV@"AkZRᲪBEViPηrj(z6ǘ&((U tn y!q,Cy2""D`@;=KfEr"AZ'Pbcӕy+zUeܳ%tF)v鵎b74!G Twu{dZk@PMc)szURt:Wͮ[ts[U`K%PiSWsb/و;;L3+]?0)}YoԣIj/71JŸJHXVvcTxzq*MS^8VZ]8mZ[!jۦ+PS6-Lk8l?Lq|}h^u"FUd.~+pj{7Qb+?d!i<j` EU1ҩGnU%8)@-sm d!Bڸ|>ߓ1rNPUn$h H]5q5P7,z*O]aX-?> mhX๮]^9K 36u zWj\Ei}\ƁfwwV}e `A ۊcSKQB5h7H-lP:jZ;I(:liǣBD(` Mw^F L.gՂ)x aLä FqUb]jq"7-93MPJ`ϫUX)Ny6*)Džq u԰{* 0TpCke"JKI`2IwErPc w/I>سdZp'%4GȍOp^~\FT35.D]׋h{Edmi{hM2cNl )>9ZN?97KOd5N1eqVlOWӮ" bA/|S MmEӶXXQs2{c#;{u,^9ci~_g|;M&v" 3Q31<ϥٯΥS4휤X&m}g@k4T€VZy[UΦ: ]}vi25KFoYWRUj4V}-"7yeG-xԹkr|Tf+hqiFi%IGktjƑ8Opɬ]'O l:h<$D\O+Fs<'cRUzd_ALHnPZ%ZfޢT}:xki[ ][[<E4+iY 8}ED/zO)iA݋*\݋՘U{Rw% PS3^Fh5Dz>Cgb{W1P,~ ݋GY9edzn dU>'|4_/ a~me6jv wqV^ `"=Os{_&a9`q>>GǥCo]c8'Xˇ*c󱞀TcLR1:IF${cF~Lx4Pn} dPDc c6E'Ï obl#^9υ>"VѨICu~x9/Uj_Q]8+Ҩk,*6FQٻ޸n$W=,dU, l0 )Q,$;Yb촤nュ @"Ů/HCL=}1\oMMotZ[ꕩ*b8i/(Du3w0ٙ'/:AH*=Tt|=#̡<63+iN":Ɣal=#̡|[M;!J E?|;γ R2<͹|ڌ? :%,IdMՌ hP~)aUwOSdRK2K|I);m? ]qj`сѹH=w=w l3Tbic?$9N犵.Ũʎv_8oi讋w&ߙ=8j}Jy%N\[ݫ%7d 0N?k7뭥٢_Ӈșbrx4lbَҮ|l?-zUm_apGU^E(KMy(q9-pn-zFsT~Z-p\-u:CY7; A˰1,O=`ZZ#ҞYJf|m豖Vf MQ@i~N9h&bu4F?>|Jj u I=rp-QYEHr u4Hr#FcE@=RV uH=ts_ m0AÁ\ C\tmo':"٘~Anzw|Nf ˫[N=O2H8_6mhم_I%8jyzٽ0knwv5Rni\';Ɵ).͗ٯX?[w[=cQJȟu')ܜHZӷݧ;WItYdn?Vہno11@<߱[?xc*-29kHU#T"wg gxcFqE $;-LU6VB*S.!2@Ыg itU[Ud77M~xw[`mHW)$ !=:y'ʛ<ÃO+DBR'+yVnpqẜ\}}[oow&1}c 7%\ޞx8'\ܦ'?^ȬmǓnnlj:īS6ü?kzjߞ~2f{dOfqpUpCƬe!qmc1Wj1I1yA (^3 M*rg .ԮNwD[qʑ uCQ&>D4ΟuIuKLљ2Gq56Y{txhg?0AAa\3#čOPdCś,T JԞTGd5).[.*<՘:T-z:{ʴSeb?CkVۆ0)G9X o'HeVdP "cE~KHqLZ l ]dpuI'uu9g CҩJ0\C^xtA&$bJ0S&HmX#'W@Bjkd9PP )T"K2(&/ n foT$r HѳZ@L'*u6r 2K-Eq8[sh=Y5c.V&&mU  d,U, R2:=YH FLB 90$qTgPJntJ&5AjqAR"$C`Iڢ6;/t*j*"9 sTrԺFVZdx2^]H~vcD$[9DlZ>`݌sRtx70 5̲$JcϏr h`gR7ːʪ$DygRIO/a2l,.:gZu{ إ4 el%;# f7?yG! [s (Gi9|8Z(:Z58H:&FeyPgოsWla<8#;_)|@46,F=_Lrd4F<)w=exl'^j,ҾA:i" QE=HjstSg]_yo۸`>-k~3ykw'[80a>j}s,?МIl"5Bp6B1jt4,ގ?dWG̼2Ŗc~B^8x8=A_)pGO^31J1OK !sۼ !sO 1ΗK3i_wg:9 E<;CAKŌv8KvgLnrf9f*3en&V?a]5m/ZzC~N^|Mq., 䉌 Ǯd.#WVƛsNNP A񀌓~a6^d  > h/r vU@Ю)bc(%ް;>mg㍍!bxex"y\|ubxb도uRvk?^g>.՛wyyϘx͡ϻjrGG;p%;L!9sA4el:ox@rOItE&*:i*-A' S}Do|2Przt*ɝv35}ve zMS0np*-Q3sq10ŮHf2{֐}-7D@Ҏv>pxf"{4oHUjk߇ϘQK7ZDBTǣ3dJ2 iI8hKFki!w2Abs;w+n,n+n^8h_vt!ӝoO/?{.O?[kY4A%F{_ki $l^ Vh ZLtU׋v*Oknd]-F5Rl!&EYZ{Um)Jm2X /iOo,o7nQmOSxry/䷋Vӷl(@mgqחZ_el`?IPNk.\Ǖ%s|098U˂^^)k|'vk1 Sz旋')Rox*g>Gtm>+ou9oߤeQ\4kw=r1Y*\>ѷ $MIJݝ+ivvffm@y]UOX|`;8/wWNS#fTБDe/] 켘^s Zy[qGvCbC;@Z/dĴ,V>x=]q"RkL/돤xC+Z Q"hC g5Zf4i5tڰ8L'C.tR 8{$M$r:kZX6zS刍5| Jtс)`̱ 6NaBhn>F|tp}cFizkJm26 UIDNF=`"Rfe3E1GcaQ1q StSᾪH5j;aYק#Laey#,b(e~<ƌDZq|`ظ]^:pW\[1pl!*lsl[MB,"/ 86rZ"t%6&j%$ V"3;I=sl̚2Ccu߳U`(z԰ j;b˫ tcDž}@XrTV\Q谲#5z0Zcj)fD1.BّI-KTlkN˥Q&/)՗j"?fCbth>4b]N5ژ{Lvڙ [ǝ9[5qeFD^M p~u>rFް/ #kb5qrV/b ~{79c{-½ʴ+%@k(2Zi#֜h|neo]?mn8;S wF'Ht4|hHԔU:9lIoocDjNauu kzÒç\9CL#];ݒR~yvcV5x8<>eH=iӚ@uC/buTI1 f&&{° .%,g*;ѵ-_woDO(1B S 5L }D7b)f&+3B q>1,uڷQtjCڸ` @U`W>g 2MУUpʆA@lC`8釔8IYk`aĢq"MI?ilZ bLf9-dlDEkxRպ%e/YicE2z'aq(`77!w&!n"zhHLƒxG:z$<(3U3_Q;A&i`Yz)EK97}} Wy7tzclܒy v/8M .gr!'48ވٔ[ꩳ~Kip+F%JDcgaR4AmF{ZL. CsjbtE;T^ YYez^CiR+mUP"75)4N&&FI#olrnr[UkBHuy.a~1fkbPxXw[:cpa8҂ũDlq,MbbGxuȐnŞ̿L^!\B(8WOY_"7)kUoGfptDW=Z4+ƀ vQ5Xg\]Ʒ3Rngf|~o=|a6jd4B#N@jP=4/5mrɗ #~i $T#YvMENQЀڄuʃ821+I`@c䋎T|Z,5^<zKު>c:Y8{! 6Lf3S5LiZH99q УYŕKZ'.ZIQrYVc(;eK $[_qNK{ ;28j+ .po(*Xlھmp{Ybx̯Naz(XD7I’0# w ; *t3mK]tLe@cлCױ.~R3ﯻ2:c̒CWI-Wi01IovKLVLnUO=C`S6*݉9xPĕQ'>UVu373\zvK:*U\mrvCC{q_dDDr:nOjp.]%b@ԌA] gߧ,«ç/W7Nj?F 7hIzm\?v축 OIW&ʱ_e˂7=PzN?^}.+ϟqHf_?cc5 "TPDi뻧{QJ?euJ3k@&ڢl=dXF;_ׇg#l@j4EXځ\pxnq0e 6I;[W%g*) zBohV~PбVCo~` ڠՏU9hzឪ mc@k3 b HNyѤڐWmAڨ2H+D96LJھ,~6!%8[-vwzѐ(M`zauγ! 8E0=&] 1LfϚɬK9@V[fBx+nHIA|EgG 77sb[sLaѽ 1Oƕjj]"ӄcl(uSl7W|#<5K +UC2+AGhH?N6&fQt =+Ӄ ;ϸ1A5)V @&Y 8w] E` @ Dik^ۦ+BMiP y.F?9-K,]{W叠G^ld'xDZ[_<"- -wlB_q%wyY?%e lLcBв] hr!Wв''5(#*,4ΓS5/ړZ@(=kdYj͚^nrUZEMi9ZQ9*Z~L5pq'\]\쾤xLw筭[jӽ;NFp$.l'*T;#p^Ew4\]VQ5N8.>7Q6aVcF4eZG "H-l>yU7)M*GUMcAZ$ a-[@E2C¸#W>}ӵKِa~LO1 8(8#?TU򘗛6$/̧ago{Պ Uk0m؁qbZiR:§-!Op[wn}{&XpޯMr}ٸ4uֺF&# 7?[G1r>wf:L2_~O{:= tҪN\M,MϾ%Z63wgj۸їIxь?ddN4i$p'Ȑ'#%(yG϶:mQ `+#QEşTńRTWJjtXgM=qKUkXM7Ҹxsb|w q>GgD{)p"hHp!0x,?9x0^wQ Sjg$!cʅLa!2Lq+:oHg EoŎm#Q3@\s#1Aު]ܬ۸Xol"UaF7#f<4? |M//. (0 bk7AZX{+e.\ŹcFJbn`s0l7dK亦9uԏHW 52z%S⢎e/^ۥzDL\oE.wȢHw&ϏjbA!:~VܭZjC+t(&cORiQp=Ͽ-g\sM@3#ezf}ћ `,#NBӵ3!(셡"]}tһYۆ)[:bqZ,> $@ qwZ{yv\>NI_G6JS [n)YKv\@Ub DSn kyhjȑ8hCLE$|n͝ {йe6O"#8~Q穢8x́bcz=h͉ѓ{7,=KY:Xb5^Z0QT|0}o7w?V;])XonǓ1FOvz] (#(dz^gIt$rV- ~:TFr (Ӄ#0ds8h *=/}+:>PXSsZ]&˧.d?2\3v8Z݅J`Dqx+ z `TCf΅JfkwAX"$\슢f "Zτsͳ?&r\)(?Nf6pm$19~S-+dJe{}\=|kfz# Q׋WZ鰇;;k #E{jڃMjv W<QvqW6 ^:.jZKͥf綉m&?}:֮jmӝhT̑Ad|1;cf_df_G;}5x.KH&'QzMZ{{kjxî͘OR.1X@(9sX%$('iq|HcBI%"-^H0qϯG"#E/G/ꁦ%5eKΕJR#v\(}y~Ъ"Gx8E}rԱĵD7ًyP* XWeZr1/TJ5e$WZE`*[}1I#nׂfG5ABV_bEkE!@-x4ACuZ۔͐!뛣p=t^-T=ɗv[:zf(`.(?4Ay.wfQ=Kxig C*h}H[,_|:ARc!1w]gك9E u}.Auz'J,@l}'?P~)rD2_AտvAӽ#6 6E@]#zoS|}l3˙G97(X+;/L!)OFiTH.%|T Ap7uy f#%D{.=EHZMnj* #4QAک`%~2+DDJ6Tl(Ϊ+Y[^uK(K$pBp 9j-12pdC5BӢ/Z;,?DXz"|:(%"_jZ^cJЕ*ڔZlb˒%s5f[ۼ&{\#%;-scb,9ZbI6> \SL{T#($?,m jPEiINﶩfkLK5R/"Ҋ!ksr=J4Y!290ei A"V_$8ㄹ tq֬_9wu,zŖV[Or/b? 9g CNfOG޽}+gZIR+pWIv;{sF"Y9/7W0ǰ^貺;ȁKۣ7\ 9LbE:Mbڗy8 M`ƨ8Mײ+2}M> BP V~xw8ٚ¸ApBtm+l La9&b°X{n)ĒBZͅH3,# (u[+7 D #[+ 6ɵajJf<ҩ[`ggrK2nNysSٴT$݄%S9_CNɛTR!fw*;sET2v FRS촅1:NPJXt)rm6 ٠Mt(@V.񛏵OWOOc9ѪScF[6iHzHЙa-@ @dMsso׌7B2Y͜Kem-DPȂrfuN,^!>L` IUz v`)$Tҋ^Vom&>Dmw%i&B`ښ1Ž캐,4v{&"d5SHmu@~ih)s-UuT'u&턋ѵuKfukCCr=Z0e,ȳ7oYEgܿΉi~M4' r75wfzadMK*?y֍&R VH]O9]L"XonCҽSAO'IEHJ}O V( cRmҰC㊣ii| +-ckV%A̅2Fͬ6LSc7L<(l> vM6Sy`j@t ~.X ?KiW>^:CMfQ@~oj'"/_  *&-<^>Pdd66&e.blb|fc;]|)u$ЮVɭ}5 ̬ mDƲ 9R|~sn2~SRprY. #2,69vj$R͑:5GNIyHn'k~Bx-,+ G(+UBV/S}^A6_  d3k1G\=wkf}Ax05VEJbEē7`~6gGɬŻY 3/&c{"|n\>y/\NMѯ`c|PHW:n/{WFn K_6_U.)ֻ)c3KъHT00Qkqihܶ`ϲ Զ%*pw_ns ]$cH #67ꃢeXw U2%up ˋBt7࢏ s0D)9J57oWQA6"uf\d}eR##,akPD&%UZ<K&Rf@ikED"8(e'< e"0.S S`S*SlCHg8 Lo"h ~zdWSA]oyMɔ|??>|~%~ы7w/0Ac&e=?_~~s5ʧwwtqw@EJۣspb*$͒k# RCxۓAq4\>| ^@1BN{4a F_em|km97[]8+|"(t͛Fb8`JDyēALrœGt%LBK[AmJLa||Xr.)?Gߦ_W~* ~֯W7I!̶Zys,>'?yXal}aѝ{~٣_m{?>->)6Z- ?Ir,I_xj^)[xRsZ1bҋRR3!pQʱS+ff7O,"us9AN&QOؙO?KټZ *t>VH׍εJbZ8&ôpB+5~ gX fs^d$Y2:KΎgĴp ۧBԩ<J iKykÕ{L&侺=Zi2m.Hww(KQ TU#J 8Ey~%;r :}X_5Ӄȵ 05dBE{ VϊdFʐO-Npr1wFCCrE''366s*ʥ}6}2@n_3=}A~.TT/իUG|-N:]U=QKL9o.c=t(ZGAk*젗%՚jtg|KC5~V|~wՓѻ|=hLߒٽZ.E }Ow~1!fHlmXn?9t?"}sٛ&g#V9g}8wBw~׵ܪBVYa JP~yP ąm5}TUZOPkhKTxkRiR>>]j?yKb\F^k}qE휣ӥޏ% mwT)|FJPos5Pd &,7:6E {?Sz7S];wө*팫/BDgޭ~dwkBqcSyW;-A}Gv(ez}ޭHք|&eS֗n|3/q[X_bZ`Q JG1n9v+υn],k*HFRNHdפԮ#t{F]UcaN^GƝA/5D?E>*%, u!rbؚT򦇺'V"l=t@\=][sע@\HbUD9?z@c(3oFfc{iTNiz;c{L·(y OOtӫjS;*U5 H>gުniB7猲[US;[Uγ+kE-Qh$}f+Ve9fѴ`qybvjutC #x;"A޻;-1_m&5v(enyqice2 , ٠x:x"īR JK!}q7+J2`GoyZ,7=˞umB(p$~|Y *5 { )3@)3rJ'B)L<@Qg)MvAʧ 2*~'PN_#Kx{yn4Ӽ5VC5:f$/F3V@~&Dw /}6]y'8m9 Q/W]y*¯%hL ,{3WDw/"\oՑ]+lEkQr?м4x`\@5퇀A6(h?,8oؠ)b)kr#yw0yr?[Ky"Xݠp2ɓCi+!a)KQ{8RHXD6JVGzTv7VOzoI?oo6' A.u65?~D):ߴ5.XaxM.Ĩ|ߧ:OQ2-!:8C/n [`"Mikb6I mgvDZ3%.>"\峓fֿ"\y^H7;J ! L="ÞT(b";&2&@lTNl^~@Bc^a0 c-r*NԦ(pgrعG#u'9Wu+ɘ {nnw~(-k5}'. 7(=y)8Zr)Lݞ%LG#)53dpO{'Wk~}[c idb=9AIZ\&F(t`"-ąA!R_H-A/8-J9hƉhVt~Pz;ϭ[qVQ\imA~SmU$XJ$2"oipαuLal]؂H P 8B}Bj8my-וCGlyݐ w5zA&sHQHŊѨJW$hl!{0a):Ss6ҫ}D1Ua㽗*|NG, 7Cs֦󿠬b#OwьĴq̚6XF 1WB9Ρ v*0+O`fY;ro 9<01y\a@N8b3ilN9FέAi+khCwh]M+j .#kU}Эj@կ$ЧVuexU:mS9Rva[ ^jө[mdݻZ+]}i 4Hl>l~-iw ~#8Qe*&(OZT =2K./9|=?AQ?+-8VWz1lf/y5 {%Q~fzJ!nubߒ] cwd};9Y3ilcco(nJ1ZqZ'[N+kt#[N Y%1fGAZvTW7"':7MRVY73RdsPց=FR3vj"=o CD,(VZ*I1A842f"$0U VTj PTQbǫA핡ѭYjNI|}=}Pd-xI6a-`c(u&x $%eɃr|'/q-^]^=Q`d@aɣs0@f|c,i&((K~\- ၞgd1n[iEC?E{ZZ7”V:7hTUhS;linnfNm怛/]T-ԛj;E D}S+Wr)xfNm.>Ĝ{1mOQ%7DKOkؘ&`rsw6p`:xWXjizxnjОGr낣/@x<ܜ'9\pZ9oo aBY_wY5U YliCB΅d:<Pk[BBȆQeZo/5ge Vkr˧~啀h  \gK͑ӾPQ[w&"?0mmn/0[ENV5G;uD֓TU+˔N k~ `{N9; MnәZ_ih`0%.5cXP^BUZvj<zϝ>{*AYl9!+Sw~~a ֭lNw4n4tO>+Һ!+SK%ukM%WH}[],λvBJ$Jfw22ݭ>6E>j0^.yѶ}7fD 2޶WF@A-'l~LL`,`55+U~\紐i6JJ'hZU:r\P Df '[A% 9 5\]uVSÃlsLTFSR2]R:lMTJ Z4HYd*9 >5waKs _yAM@|Q2J.3' Z4/6[Kh1V*>:z+0q@Qox C )wZ*~ChLScn>:zt}!taz}oyD ItE0#9l/L w ^fgk}-"%,8 F" $ JBElj$'<3p`SD!/ QK}ґ oCCV^KcC` D(\j-C)XJe0bXQ4m(R.9n>Mݙ2R 0R~,A.&ַ ϼE#L!R'﯐n3#>,T~\_4,3? / 4דxz_wEDwBퟡp8$_ &r ԑdB ы!=R(j0hz:2 `H($y]"hoxཷE FZN4Q=GUM)4ݱ>|jO m@ԭTGLI ;smfbP*J9ïEvK=1z䍔 ɢd5TӧV2D:ÞTvWٍkT{w v Xa tHn]ñB{Bgw^wZ{ V9茉M4|_HKrat_4㎐Q+@,.cLeWX/cگY.'/Pa荽_\ VvM#ӯB.'s$XpvL&# މP Qš>Nbb%vx(g͉ov" ^zF9s6S(˲4J##ESi S+c$0C"LrBKG0{B,jX !-*5Uf+ E $`pĵ(RF; E3F<վV^?FɃP;M=ΦÌҙSz,Qn0Vý+ h|?^\zyEzJ;WVcxX7֘~Pd G*W eSl?kKCU,Ƴm % ݟvJު89oiWNf~2Y}LeǠٝ0i aS8$T W`'jZ' 2W!)co:.)ޮ7TX;k;|3% )ߍ/!TY㦑W-x6Mz}X>i D|+P0ˊIiդ;TbB:@"UKԻ)/Fe_=X ӮRfQ6JQi8`)Dܖ q+=.W}c \}5+2x6P`R:guWȤCnU+ԍ23'~ZY )Z,KJjWw\Yk|Z!FTN0 d\ '-yԫ Ռ3Bj-9Zpq(ko""E S!QX܌ 1DTcWFwdjhd@bS,)PyWGrl8-Wg}kFB)CO&yCft,^_6L ;, !$ dЎcVyA/6%0Qҥ=jdꞞ`p,]U]ѽbXPR<צtIDGomtHטݣ92}!kl0K,;>E _cSh" n\"[p@8d} 3H&Q h=0M<K1v=d:_/*[C~lh\u3d{&cQ^$unOLh)&$DLlujV6#D@`Tbb2"1)#)Ba5i7 s=ϋ%KS%-ϋǺьƉw~IJ4Iy6#"OUWWwWzekulfNl-/@U&Hs0'P+ 53w+/㤟-pE..^NnAp/^n濿Mf2^fËJNsg 7`(;.%hcrףb|bcotk^\/^]r'XV.U^ne{yteN/Epi^zw~Uvl2nsx7/k sp;[6\*>{_Jڇ< vڀLw ]_GDn1߲u'|Ż]y֊E!vH^:7/6E_QaO]Uj3/Ox(wUG(kڹ^qhaLVŵ$G_^,$M[F^(O.ySV6->>KGvUGO5GF"Ɉ~Wқzh?Y4{Ι/^uObT/l(3 o >/z~7 _ a-5X†'91Mrny箋"Д6olYG6e\ЯzP=eN$ݩ_GѤ{GY>!毧ev_"JЕ4L Zz;[xfLVثcf"t0!m2ژ5͓E%yGo.?.z?F)|f>z LtݬYgtc &u7QX4:b:`UuS|u  c?>.| yV#anDI GYʭ0FOt!}!>ZFÄi)M#Ie,dkF[6Lo˵],+y3Zdwt5'8{fl@HV_WqdfsfEW!"vއ/5fEirӃcJ5]wi9IaP'?t,B0)Hp M .|w,Ӱ]3Ò fy| #SO)) 7| Cv4}9-;44Y; 4RI)oI0:|y/BXgoR0F4ϟW,]ڊ/(ȥIZGIru|U9P'qbi=ʼn2 f$t/]$ լsU{7 6>yYj@XΞfҵ?{ZuklgF4urE9g Wi-@66UXdfkŽ@QG&RjM#,b*%C1? ToŪ.6}{l̯2L)B*N)QI(98b) BXSEcGDD1QEQ!T[}Xay> HؓZƾh5~H@ P~_thE0To/JO`)$nOń"SY p8|x7 "e,GWo} \w@H"B E?; VP) ÖYiS. 9pI(ƅt$ DVIez&()$:b<5Oo4O_H·WBG/D\X0T4\c'ͬߍ`޷8 8*?v?r߈mo10WXo!BE0}i, ?=XNm˜gxϣs#INaJQ{WN]EvzJ9;NY-k;Б-M@l=>AVV<һF"֕'o=I$5țkgXKKD:4IoQĈ?$UQj49gt| +f")^J: o[u4ޏ+ojbTm<\$x8Ѵb0@Ǿ?/ ҁ^zmqy= &}R}OP(3}.D")PƁď"=~3I0eL6jc {ap-a S0nkO9B2A,r1i34v)Pr }ԗ QQP(!ةȘ*1&L=Me 0UUߡ 4V6=3W a;I &]1fU;fK5yGQdHjG;r&0jji’?X&p= ܚ7 HyWPC$H#P10 xRPon\#cjI}HjSYOлarqb(B5y(41`R;^la+k*C"]go{" }!N1 V)s(6ճw 2l^b^UCwJ9B;@+:3f"ryfz?MfOߦٍ=Ur\zOϟ]~fҸu'+ ف[kZAU>IlVioOd)+$1xΣ6<ǭuz J!x RX60'}\m89€{1F"g0 3 cXiyCJ Pq ꣦kc P~u0P{2nтP  FBjƾ=Kh/(%g0E2Ɛ[bD#7O)brM9tBU|3MID|FF!"Scp- %Q,d˲k-)IleG#Ei-\ekás?rtanrx#o˺(QDcR_:W L}zѥB&72)BNEp̴B,&L(IHt0"$0^DŌn2%)nn6*-33gvS 3.m|I*t7ķ, Ђr蒑B*@tHšhaom$]\qkK "8&-j0 \2٣p͓YV%Z{$\neJ3͗X_ժF]Y E9 AU{E*#92X6lF%@Д{k6My5aKh01'FלFWF0SvFr8W?"U_hPHne'};! uvY%  \j.QC7aJc(YEB,WWZ)ZеPe"Vȋ́pG^NsT)p*X]jӝWEtX}9%]k^tm9rB:ٔ`R+q|G)^ npӹwBFwwst\Qƻ9LƼ{vKV톩,Ʋ>iKFgyW^2kp>3qS-F]Y Ngklҿ42YS^@٩X7T9%oJMfHpEIgKUMƕ9Ajlu^Xkh!t]UAjD"+6(*ް`t^X a!OD]l 1zۻ\+62ޭ to̻Zz:,䉛MaX~ۻIF:9b`50:T4W豛3mVq'楅Ԁ0⟶c,K%H腗c,r ^z^KY1XKyi!5Ux9]'k:#R_mJmi/=i/u[է8KV+:xI{)%n^JIKꅗ>'զb㟶y)^ 楅 ^z^RV䥜@/9楅 0{fTyqXFtڀvcӿc~cs Pm>,oCɨ:\_O1Y#Y \kL왾շe]:0^}NGlv}$@R^:$[x& nUx ] AbK}V@-<JJls,XB/\»%d²?=%Ho;B? bO/1G0JzTˆ?ZG2؂#0=ʻͱ1i :JRLA"N1 ihX*L䉼^8k;)LӛO'+?KvJH)ySG%̏IDe8K4nɝWvxCqعK NDwm22kPx)p1\Tvh )â0"b_/r+,?ƛF5CF+M^0Gl>6ۙ68u(}07ۃ ral? [01FLC&:0*$SǞKGy!Aj]jc=g>@`7f0űB$L&H1L b!IX8SvmlQmc(RAmwނdžȓỪT͌}<3QGDwՙ *ټB(f;ffBon/#T)M3!<M*PjB9I{6QUX 3p+c3mYŶg+ &l.g&g/<ƹѥy6*'s Gﳓ EAMDS˙: ꬁt@1SwW ̀Ȱl qVL^Iù祾z"5υcv>ڼ?`Ҥwm#3I쾒kTnKN;n+=LT)fK3+謘Shr^[hTb^m#:=Q1ۆhmPޤb3#ʺ+f{xw.f[>ݹ+o#H1f P߂gPl^k#$Uk=4 j-D(DWg4aih|ǘ29vsvZCԪwvZ}6TFH^ĉGPdo'>\jɤ8V{'Cjȱȫ1Q;%bېP>uN=jCj3d|mR|M/L9X-|.S5*O ԋݾ6pqw4n˾,LvL:F򍻨O)it \\MgCq=ܧI*iڶnKնuiakmFA2gofS{f~>~se scxRTnfEqάx.;fYu<1JSdw~jtwѺ/:wtܨ^]t,ǣX$SEqy>VYBIԋn%`Ua T>탳n*,^VJ KX?96>NKݮCYb(/u+\IMʜ2e"P b&/yj ydIAk"{kpDAɃA!Kܭ O!ac8ǂ9^Ɛv HWa $賝ջ1( J%Dؒ?5C6O jsrr]Q:<}xNS{Ů}bGN=&\i!~8DWEsF2PĖTHݺI`6lp&Q#l}&-6]0bBS/<@Nl4?FAzK \J1NjJ\EwʍR, Lq-+ꆴ/\lѱ xк㻏}ɾ7DZyE诋:܅#7wsm*[}.2?R۟ŷz`+J.F?ݘ0g.moznۏ9&b_|!evk̺<}?Qu|#T狙=Z$&k륇͘Y3ʼ%B%Mo.ϖ8BJICJ_l1YЬcAóڪȒ 68|{`c(L:3汆(9X$R2\i"ajcB ' $2CI JZq qO*JT"LKS A*:<6Os}`C47EO\logznL8-N"޼~ƜK"JtvԠZ_D;ޱ*Fgr=[Ǩ[trƓdP|7A}^[rd+9sޝ 2ڳl1^۴=1A?|Էߓb0}A]e;P0,<(5✅ $m٪#ݹ5ᡣ8z[a mm뛁Z=6/fzkkj}ڊg[5n խi+^F[C̈dh Dr7;wHϡ\`DPƕƘ/ėok}KՑA&s@Օ iwU_sZ>9r_$Y ;t w"$Q}BZMP}9bwT 2Y 7w 0@g 0PIC]oS1t7۪q`Ϫ2R o[GW7?w|"K{j_+7:j$loŌ&kFS$nyMyw{iZ;2#V.5޾WUGݭ+j+^J"` 83NcEXL)(Yʨ&92KMƍ"B ˘"x56䭝r5c&91*dmGW#A 7d/olr7 |+vPxn m.T|`jUgh{$QË H/rxDPI7CTZe:y<'CT _::7@V+]LJM?/RťE)2%JZ$"vcy Yr] lKSo 2CC/ CroV,?ԋoyeqHSF-~d[IH說CpE:yQ%:(Q=5_C #Oceꩋ_t8wkj*V,q=+ot9ʿS8O2Q33aRˢ 9UBPNļIBS"WhUKe/ҪHaP22CҴE(NP.̵ 67m -S *];bwxIw4evyM2U)Q(G@v6fzӛe˗dBԲQ-?"WBNJ2WS]|ͭ}0 #D%3(>o %۾S"ۣhLl d{l4g+VgS18jlt_ݝi*=+H7X67΍.ͻQI9ZLc ,kaXXXifh] RlBij] y(Y) )m&U1D?n 6CrM7ѽeڟ8ĭt*Z~f2󍊘 1[[BS)]A>j+vN> L_?f-E9q8JMƊqB) f!HNo.K&*v/R kM^H9eP.?l9h]ڧvq^GJew +I`5=:JK)Kfaz!7Eid1A1SE,lMIRI9cBQsy.F%p8ӏjpX G ^v<Q dI!RP,S-VwϋQY֧Z_J}Ũ\Ʌzڒxq˚Xۛ &&Jo޸H\&&ܜDbfDt/i(<2HI42%c+calI$ˈ@#*_eʍރ#LĬfn&:=KlIZuD$(Y$RB(-U(DFJ"4>I~%G3ȭ,unvn!hЊ;Cbw}={H":gp>3ZZh{Hh<7arb1(zHhx # r#{<ǎA zMSO Q%M?ߨQ~䫖tظ>&TՊ/ J01 bdwC~u/k1Mvq57ΉA%W\ɉRw8Q7pÊt%1;:nDoI @ޙ4g6&@QQ)CS}%P'sQJ'@)h6"I t@8xk EAA{xFr&q J"\d91Z(hUJVeŹOɃ6>h& 4;z?i34=|작W_[kwdh2y? Ws`=*s\v?Bs w|'W1ʽ{8jGzc#qpQ >I)9}~t;ݱBQ>sL4ߛi |+|Ձۡf^l ³ẅ-Gg9d|=@smɐ`Yzn{9+MR܎5 ֻ4Pvo%zq=NlǑoL#̌+r/ϣ%'] h-%%7pwm%^[a-%.&S>g'wE4XȾKrj}}t;6?u 6Do6{=C-$* F > 2C{:Xƈ/c{@34i=NTB걃)!Ӝt{֏x'$ hKo,dc Ƅ( %TxZV 8OIhIhɐrbӮAzNf.gʁcdEj;9-·gM\kM8V9o#sKܖsKgo{z//h]+|4޿e [@5\x Z[j'ye2ys5}e4($/B(+fVkCPx6bxBcAj9wS{_AZ w~&{n^_v{A$@CKDR lhW [ONCs$ǽ4iS;@ks_Rmzxv쩍S'nJzG{B7n?}Hƌro>3iquE3w9J`ҙid; &S \gwPdF Z}"2Kpd{# ((<Y_r.ܴl?RͽX0%Ko*E E]ч|dv˘#Wr5 9%Bze$O-dd8zKn=eAvC8|ɷw7r1- ̓\2!E~M=D}ym\B<*1N.Woz4`à/q.y)Ih--A xHCCʱtoZ9rvU^,Xmuz(:u IEa Ҷ`)!ԋ ރcgFarfl;on4\cLZG` 3?JuFJ+#,>bAgQ-R1XS.l P`IG=QxFVi2n0뫸iE, ^ѕ1=s(>yUZorisA$YMTj)  @HYǰf^72Uv]BT!fI%-kК۹Lig/vlR:0FkDh䷾Io"-$ehI $3\z]ȿ"I(Nڮu̡@p x Pqr\)$(2]԰S`iJKtU*R8c= U7{3n+`ExiR@SlKrLDR v,bݧ6Gd(@A/ xxP\*%a*\D\Uթt\"Sxp%7cf9;4j자3Nr}Yf sgzRFCqp2&SRHJtO^֘qYӾ~+ї\X4.eLy[ !f,XTb`eN) KL&OV#chzZoB7͆m˛\Hq_)[Tj>E!!ӷzhDYgln2 E8wZ T-ġP*hӖDIǀ3 JօRS]Ca=S_LzRP8Nea{dMEu ACXwV׽X+R,rYĬ,7]ާ yf-<0>z0mBg+]g:Jg$7qczFhY`|P\R(1M^wY#+be㻚Mtdxh>Jǯ}WO3/s G5KaFV(HM~f۾S`=SMB`"H!DsztVF %Sȱ%S"2˅;5tō'@X >ޥ}i!Cex8e$FT#$5@ؚ5K؆as"L8-t% #QL-Ԧ!ΣRer CIJLp( I5PڈY2˲U3,sHx%!Ɲe<+mI6)յ:&p3)l.v;9 'Y(8/1gHO*[]4 oLC EO؍A6vbS,,^?|'46d,!- BY(ny Z;Brbsxbs55FqeB$06 'GDu?y|݈87_{$/˼(^bRiCvcG2=\>뵛ekmH~ M8D"P~4i e9 p_UuuUwu=|#fcL3N'mT[R`MRuWl,QqpV\?Ș=XG`Yuwdve/l0'`I{ʳ.z3%5 ` We? =21e0'IL1&SPz8cV,c/CF8Qt:کᮔ1JQm,kRg\Y}ys .EE!1I;Zէj1tւ;yÜ&:ͱZW!VdJbsZ;M!$U i])=åPmKyzFi'r=I1i-Ė{BjE{ϳv/:-7IMHE#LR0T"nR`-eivZ; ((xԄ]QDeLq& (=r_޸`6孱j,J)L<$/X fQJt0[AsG7i0ʩ4lBBb&Ұ&4ɐՕugJH*zDžV˽iɔ磳 5^4ʲ,Ɇt٥R S$){`qLh*Xfoqlzc9\7qXAN"`2zmc=LD_uF yf΃Gy|ˇ`# H$pZH%Cp>oƾe$no&&*J?sӄи";B&Gt2+LD2~[WIJ*a$4%F.izCkz#~Qrc"ߙ4Zd$dI$^UF/x_APIfإHU"Daf;'{8jFj!;uۄsCH!zgƤOL@cTق73_ɢQ;RWj3SigTB+%jX:s%C_| y33!0}&66$D/RP?k&GF[l*x=7)Z,ؼjgLNXze{S{Mڅb3œsU@5L}!*NTfM-h[_o$E [,qwCP11S7HO0e 99;Boakmx&h ^q ʾxF2#La[I bbx覹}) mw>%T  /ЁZx1g9nmyoyMBF?+5a!,p62AHVdЍ`%XGjkw6ҙ U˳xXxH£3a&0 F=xz3<>!ӂx1dfs`IrY7CnA FO <T>s"cBF9fu{XˑȿπYW޺!Q[,ҏcghb>(Mgхyc~5s!!U \*%eoy+ܻmN^*R0(>°D@)a'5լyo,NJ8 Cp4 (90h/LqxF9LyR30t,@DQ7[i+qH9Ք~ QJG޼ICY+^YB +kI ,-ŕK2'}mA" yg@],TG 'Ѥ#LӪlz1jq ;j74^^Ol`>]T1٨=?vq_uc*l]"᭾{>WŘk-+*4-\*SAW}*nowǔv bԔp:Z g/^q{Mȧ O6?kqǎ"u@ye:I w2*0!$~zp*``ɮc!)[G?ƶ=_8=uWQśpt8*M' zGF٭?E^Xܺ\^z}Th2ǘsV&?xu[5A7?M td, 3JŃ5u⩐OZVt|3ϭejcc:hWU0uМMXm"]4;ᆻ_ZT?S-FaPE(,dDqB,Rb(N)ެTW: j_FT#SE1!gKC["c[va،6K?./ ?~w|6 ,[3.nng[lrI(UT)K+ʘsyRQf&Ql Bǔhsye~:=8ގt~=p;H((_# _.=;*;Bhh4523mȞ UbxN?"ߒ+![J"Vy1 ^bAH)/*üһPw*uF3+6ŝHZ_XLR$$8Vy,zpEm dyqtYf> 9*W>9D% #@acP.fñw<^6 8>]fOQ{KG3 ?j_oZ:ᇚBS9G%5gUn}tc3\Iۺ'W?;RC^,c0QZ{ȵ@>#c?o, C't|8iE[L&Zb7S~{*`Єu7]ۨuЦM-gָ?ZT0d ;179pw>.џՀbr>~d@}-L/V}[ X;{tSXBu C-ߦ2+@\7!SbJ Wڪ!W&h)"@&U .մXp'cj 8INV-ůd"Gg<)Yd"MQ%H).}Ѝͧ_p>%7eqPb QA *.rZzK>0,.&5L[RQ3Q}>33l;&jGgIu=(V#2=>:lu.qh`&1BܬJ\m[ 7EvW,&Ohgzw:Q!F/̩>$98n KKſPdeO*(/g9.tVrZu;!?|vчh -8J-NmkcOC?ɹw?:wcGo/Gh,Ocݛ1>=:w&eB=dp)ePdC|9)q 7(>`LGF(mEo>[wSo s# IFn$h\,fnO=;0kڶyj5LQf.'-?+&I~Y2ʎ;&}~VǧHs׎ARWVUA5fZVi?nGqF3&םi.OKmXci%-Ix۞ CXC񷼆^nfn,g)J8%pO7N2eܓ>nʙݺ`]4f 4eMuR!!s$S>ycfݺ`]ۗ"m [u?DZvCB"zOԖ20XrU}9l=Jb6V=nюAQ`Q8Z=zM Gg/  8aڌ#<&7ԹгC2:#ɃbN5x F>>L eeHM,EE=['?ˆz}p^׳#!gQ(iǃ:$zY岘oL t}҅ȶM[.hL7ͷNnfq-6N4mn01#QJ$# QږW-z3{r3]C * ;&tzW$|:dl4Xm*iElhi.SnGM\76|tLY Tc„&]\!*!cJX$HjKCd)iWro&cQeUQkHG_ef +0]ۀQE7t׌>ҀGN7 <ɤ=6p-"CaV0bc,8/ V{,:"ݕ^jY|!"5WڭRWnw:ф+ B”#I_ Q[JyX<TYnJY)!|"v#˳c -=1閮ؒӞXW>k !T/YK-{ {b>2c(MqS2u6xx<>!晐T(P&~ r4N=^L#ƙw6ϲRScz$$Gy]`8 %LjLA 8'Q@8.L`UF 3yHK~ [WhlU >XqZ4m zTPRo8u(Zk_)L~65KSAz,aAY;r!ę 9lok8XeB "&'4li$I F2 )\ 8l OÈjgH'cDMAy)iȇLA0 HH 71~DF9{At14HUde$;!Ao T\,y0abH- J p 8a4SCr`rDZEknAn%LWíX H&W7gnϯ#ɸ:d`ȯ#0 L~놈-4L P[s9iWr%U aZ]q+x=#)ۜ?g@B2q44nc ,b&[}֛̅{H:) - q hÐr@` $fB*c%)@?}(Elr ` )"!PP1&=ĺU?~[t"myH5~$ 0_c)-:Kc(BPA !?^cl6,TwNaP!QZ1Q`!X>78g7u~+V ^7쮝3 Y s,RD1J٬y[>n̘}SG^+nVaB+* -~}}}vzyQ UB8۹JL]+PvYxV;Lk'T,q3,lX!H>(ڠ]^Z4jBdBd jB~kd۰&ZaޭV&d[:=-}੎W0N@>w>࿩dG78)SŁv,7PXQTD;Ӊv>JQND'-sд8b4LWcwB1ivD֧nFxUK"Em]|7,=ͷ(FJ*J|!r1&w2XJG@AkcʈJ(HA*X I|*!D&bL (\OAkT_ ?af"Nە-kj[D!ﶬk ̓ zo@DuT[Pu6H-cXbFXh#C*Hmکxt4ގ&ӷ2Til~0MjvTFRB뇻%C}]~BN.VO]H8u$@şE[2t=m6PI\ 8Jhpw'0V?/(J =zRfGvsy[+ؚh{C+\W(N$<;jćT4R(I6MUrͻௗ﮿z8Xv_σ u7"޶J-ٵuJ>#iqefٜk2y&6~//G_Z ABة}ߞp= ѾxݳEf+L껈Lx~歈U[ Eg36J׻)wv5Pk-gXާϼ{alN5U@7_ A0e&IJ8]a\N~ǎS?ٖͩ&wH5QQ'v1ӛA\] fn@~2y{4KdSv?WL:Fq6~W7.RZ, Ru⁴%P~剡7__z{irTl&IKCuM]y[UNuPR Nj#rSя^X2ȁNw4n;B-d֭LB^9D[a q>nQV9i&mj۶nS2[ySeΉw^'܌W:p:3dzWWſrp<{-֖728xk邲5k#`ss` sBՈ)6r*7\Uvnçki1Zf %#L R˹*/ ߓ;C7Xnf|yTKְ:WJg 5pp'Vby;+xD)q5'2$EVhf~8~:Z Y+(rT/ZjD_M?!Mo@wK^v/ٷˁFݣ'o98F+j=m4%_y$t+ym-gIDk=0"$W* K (5H2qm`̆`zxcgSH :;5>yD50z0{mP0r dj@J<3qRBD?l:O1Jifút]B)}>Y:>(kR|'U桴JT塔SJ RҒR~@^e8&FF0:q /U Km"νdJJPh)FB[g(x P -R#j".w\OS='ߏߐKg$V[Y1ʡ53um^eR͆ZG9>X҂ƩY$p?"%N|PI56>J3h@-[ޜq! j36Z{u6 ǂ" B84>xM9(4hkqF ٚZJ -ü `9F+zM| f&ghjs>.@+G䨳X@9ϨUʻ|2PCKWtY~I,y/{Aza]اzV# ~d][ا">4Jd6ji$ XfCK.}v/6bHپ͎䄽Rk}\yd~3F#qֶقSkUlN5S\A//P z/E)}꺋˶2jk&?I?[ekFs/@+h+LM~lTmqx֭ rM.,Un۶nq3[y]?ʳ&}$rOץSJdjaWzwYft]Bi@)Vy(-N)5Jru ^4/GoF5DP(e2LS^t'T3tQ3=~x?l)U[(oPVWt]I(]O" &#JS9>^XMoE6L A `Av\qȫ>A攋:[tY~,^.#.O:VJL`ksQTl3C֮;0X;7u=.墇=6(|_ ۴/sD%57ݛiU}0 Ο۱;%3Bh>G\8*~VGa-^8Ycrr!:;&O+iɷ#%XH?9^Jsnj$SGJ 19Seœ\IPw{M`6{xW/"pjWϷMۍ/ߒ3*JTGS(sȈ̶(oc|KؒsOs_h|{6[Ƕ1[DNۘcy8O ^۔)3Wk5`:\sFhǮusc9N(BPVB91>N/e}4_ʺyir)k/:iq.e}Q=SОF+s,^㮕k7nr%vfO(2Lmcc=QQ;ctN8Duޏ ;F

(;Q#[}ֲg˕$'Z1NEE 5.ɘJ!Y@ڂs"I.*<"I.T!Ԋ\X0+˾$Ҽ9)zXQ1D9h90mjC}rsUsU'(~9NkXr}6 >sӭ4kM7mw͖_$7݋=(^['ÈG>Б?wJgOs%W1V%D;G+!^TOwh$hEΊ{=B2kqr؈k&W =v&i}?ma_&4I0>>߶gxg+|{ GDG!RIz{7.QJ JНRڦUwI/B~\nt.Y?/9mksؘ@'nej%H~.Iw ?/y>D-tRկ IP@0??+q| #dIi\|ڐ]B2b#gBУ64L8nۅ?:;θ$,h2 5,?iҹ͒dq\sP }!KkAǏu(NhQˑq^^+Z9$W?j0f`W__ΦN81XV;Ϭ T@x'eTc!'GV1./_0>Mcq YO`xک _<oD3O闡苤SGn~:/$55Dͻn+AZ\~VWi0 ڃ%ryJ"C ț TO+:{('it"##&?Gq$?<{7t^=&eb~.zg=`b.8$dѮXr@H){9c8l#\v.2U58vgBIɐO"*KNJ%@Ԉb/J)Mt岉Hrsq5FifTie[4>fSԤCcPcH'͗~ѿ 5> sWu"NəQLrHD" }`"~EdNI#ty3]T,7I+jY#ZYQJ}0ΥG6H*AF=Y!#[3?Gnۓod*Ϋ6J`$O[(=[rnF&CG܍1ݬ ۜc|wY &܀a4ǩm]$"#&B1irM;oC*gsomCcG9 ppN*ˤ KFd.roA|w4i0үIiQreCmT[ ՚%d?~ T`o[T~ldRU)A &)V Sv:Pz8NkAh $hUR1UbBZ8y^5X#iM ؈詟EH!EMAJ&}e {n,H+dXZ4@9za;BYcfJvfqY+QCZJ)ȡ*|aD~/J!8i]oB5Y8rh5!s2%Hm,kC0*&S$oX@N^ˁnϩ 𾡡r$.Tt3Eҕ -p%h?%UI_A *}5 _ʌ#Go.r&ld׾¢ɛe Z'LbO)<BuO]Y\^w4JewN-zKp.[:G=зKx-dl zT#Z/hQ\K'yqMi 9VDv9_ų4Yhet(J伢k ~\h95Qc䥰XCшR`!S ;P[>'myjh)pF~.^f,fSQ{VqV:71 ՎPM^Er0C5Ux%>g;Q`|P\)õ!#[bu=g:JgjYwd0>xk7.݈K.f>bUqz]SOLur8aScL~7ӎO ADj4syp!2,@uVF,fˡu{y00TrZ{`x y^;YyiqdfvΎ}b|`ƶM[E?]^iARqEZR`V.C)c86eҙoMU ,HdXJH9'%FRIͣS(m˗Q " C $&IYkݴilۥ̉ZWy2LEœ4A4qg*[MQD#ry>/o뎗?p2Tr 8 )qu! [񡆢Wk,ojgSd5PlųŧoӤ YX [$X7X{uౘZ;BZwbXmq0^aͱٻ6$WR0],v1ݞyi(EF($ST,ʲ~,YEDƑbT@cLC(#UTijRF.E2Fi!#Zv'B9h c#V%~&=|;;᪳?}IJ)VOY Raa4qYnZĜ>GdISQ #P!ٳ:Kޝh=qXXz2 3M R,C?2DY~6wggt<]` 4rOh,zh֡  @`RMӌbzmL?_V6{%1̊%g$\+4~RF%.Y^6ɸ923-Bp؋&Bc&(7_j>}N%6L=fLIlk1Ix4J6%`C_I8 *G^'''M啗AIH $EVTRǟ3F:ɞtjk>*Y[? 3F:I`h$)P *W$S6=Qڶ2R {~x-QRRt9F?1eh,E5*% iSNUU&o( gx8I ~%9(16'ٕ x} Fȁ<5Vɀ ɡоEaS6}ҀunK7fJF9:5׎^6e3w EEQgV4JŒk 邥&K3xۻ@2͹춡7Y: ìĪ@t$2 .0&0Sˤ1L "owgl<8.*^S^[Lg@"aIy "c9y(m7᪐6*`SI#}|Ҟ3ǝnj +@LIB _2R/擸o]&S"6ulSgx0L[{aHYg ˎoE .J{*EAJLDR8Zi^FuKΨ9:uSY%Btf&E$(ACk u mwI9Xڊ=X~p?2h=93/!q'L˞%{V |VсU.n+|Ld-6rxSR[@- ^j3&I', koo|{ A,qGM-08QfQ塰+ζ`y-t+)JP"`MW =&bbUz3` -` 9րI uAb؄o\!+.c_ۏAf)q+ tAsY/[Yk+7G_=iG vbB>DtV)^LhC{F5n]߁PǢQ]gX8@o" L`+2Fput>n.mo3AhQGiUxH£3AxD|Vऄ`y_45]-mq;I}r^WK<|x#|DNVsb+=6|-GSwQ9*Ş 6|)Iv;uO\q_ooBN>šx.$ AW W2Cao TAE=%20`9hT3?bF&{9MUB.`]qd~xF!={7lzM ݟow:ztd~*Y+^YB X_kI mٞG{n}mA" yo@W. hƓhґivt<¹V$U[ΟM5^r2w41@ ȉ* "z"1i ]4Xw%xitC FqDH>UK&q|icWmR\۞TsA}{yX*K mUŕ .bkh@酨9dAgnޣ30d%bsQsYIn3ܓt B{v@=~bq >.@3ɸ@NX)[}1Po!`]MCmD-M?[7`ۈvmgmÞjsfPЫɜh*iE8y1$$h.`dk  ̓ [f/c?OhM]ybdq|,HjHT 8]Q lwg&_T6ZAo'2?N洢NK6VhSpg5X>mꥷXs.s|g]Ke766hrhgu=;Y8 V`e Y  X6BNęgU*Bj))/E&Q6V= \9)Y_Uښ˗U>MڕIt.-o>dzU\8aue_?M~1dX)SRb@mx%OpQ&JQI)ugN ;P\v#j22˳& p~~KnF~ \\;/]~%7-gEi{;[(O?.FK0a"QjG?f`Q@850i4Ô5,d]o~ïo|-ߙ/3ީ^ŠM]fP~h?,_r:|у.ڰпFz;[et=<,yv| iRm!&V0|+%Ae>gT|2[IGv 8SH'dI3ˁa\;[Z7mc^.xyq:~,>.f7oXVf_[vX|WtrѯP#;Z`!(oF$zj>N1a]?B Mzi'he*3 {n/*sȌ BAzjmTmB+ؒ]_pz;fTֻq{XmihF/Yb> V4@m^L]WZQEi6Q(4)P:dg>, >zm35߁Nۯ>]xi{g,:S E`51w!Vޜǀi8˽VFoRgnucTI^3`OrizOsm;,Lu{=PrA\tTmaƊȂݢ4E޹*g(n^N \'͇񝲟LVN5'pNec|w  NŅNn \}཰Hpw{1gSuEjv$ًE<`QT+ ۃqʎLNZ{]偖NuZ }?bL/O8|]/]OCjl3 cP\2e;_mIeLFh9EwZa :jŻMG|' X7XhV(fYuEs=@' Œ2XtI>dDњ,/Y[W-Bi΄蟈旖//Foznt[ɣ]bٵHR PF}43Gqڪ@(1̓ _rV@&n5ɢt1șϑӃǑ|#dUw!̗ʁHE)XqJF줞iTP᎛Ę 1x תvQhΪ:~\ B5Og\·Zןz9׌b_ȫ7?8dvWunu20:M(',P7zP`p y5<Ĵh/;yaR~\]>2;dp!qVBӃbM뿯s m0'Ғm-֜\z1 nG|`ʷDv9{n=l9{|$ a Yܵ59EH!Me&F}1 j]&48E;iS1z 0߀Lͻb{ҐbwWʿ *E`DI\/AdRec.LIKŠ+Q)کo~nr(ġ r7:/P>¡u(֪_ĸECrC)F6(;8溧?qGiZןg FNe㒶hzם=%sNm>!`~yL?/60y0@#'3zj~9[ a}@ N*n"/=9nb5of/Iˮg5ܡu.y*vp~NZxwW̏*/vٻmlW4rC'tvi~h@Ȗdӝ{@J6%K6(-4L8祒$T)\oNx{ӡ(ӸhCu%3Eur˹ (D śɛn-hۂʹ.4 1k^Ƕn{~l@^^_pO(C<[]V;sM?֝ͲE4z:}l[9[ṋwiDS.:.V& WWBCok؄y~:Tqmn Hخ`i -"]b=W,&?MBr| TP#0Rq Oy}*fULb}S*P]Zmw +A?ݯGKwDv{-0 954ҥ:zW܆i@Xܴ覷 0zY ܶ-Eߣ?&v4: ٧|:{~# eu(zz^F~Ft֝u8-e3 'w=0s:(2vq房*+x~+w׷}ѓU\YbV74B^5JVfWM|kb3tQEXq7V?Pvkb!ZkLBwvAXJA7[jgX!/|)qhJ[[iDNބ3FUptxn*L]狦rS1o8_.|'>0!Y\y*઴!o؁;Uq+hʻ33to"TK3d(ݥX28J(QX$ Y8 rDE@F VuQGVjk2fIYM?Ҧ܀O,aD ]?Վs;N·BӺrtFOWƼ3bmDYM*4;A(5!#!*HBJuɑ{dZqIp0>@.B%|}Yҡj뎤TBoVe<~$-o~wv_ֿW :M',<ܙ&㮊!Ft7/5enoѮ>[RݕKZ1yKp- bǸwuz u1qFíSM\Hnb!ZgLwu{Nh mTnfSBoQ?vkb!ZkL1yvuA fmv;.6doڭ~HyCt1quyfEPjW^FPT)ciSxp<_LHEjN.Q) C/+;ArqyBEUiMʘJ.TOaITE`NRGUR9&|3e+A_$_iazĪRf M.1*&Z>ou΄mM|㛛F3k$@10Mp/!b%#Cxh&3g +Epҙ[M $^>-$RxSyF˼)6$kgTJ̐TiP+vOYOr`>˱ـgS+}%LK40NR<mP9 4gD_hFd*"` CUgO8cؔ5&6ԈKex&tKbqÖtו.kK #G.Mx)2u0{9rx3!Tρ]lCϫX )5aAN)HH" &6Ϯɫݨ645Gݵ [XͭVZ<//ň Re*X.#C>4<ͨ+b2X( Uz`~s.썤2aN$%N$D2&R<Br  r/V jQ -g֌+F{Mj<ZRZC,t/ ̣rVX'j#s?Uxғ oRl#DK4`s4r-DD KfǛ֝i1'Msp qCPO3ܜ ɥ39\Rۥ)!8AT‰kI>C^8!y$u9SCXԩXNŦJ8;AIsU\jZ% 'Kԅ: I|-b&|6.@`)Ư1}WJju~pvk3츷Dn񻿀ΒӠ"eI:r.j,1--S8 Cl Z>!{ҍ3VfD&:0)U=g |D;hOY995k\H.H_`*4Kť按 qs1+;{&gVq@.K\.Bx{{CA|i9qVX!0Vpx7;hۂ@8r Qwޏ|~8-~E{[y~59wVF9l'.Vu 9_\pPzXhdޛ|Z-ʭ"Pm$jtWFC 8mĪ'zU%`ibzkY%ųo×tX5];N- E`UQ>j(O]@!4X4WA!xy3e2Q&ԒiH-"1G 5/e2 p%}I9R=bQ %M((Ҟ)RTLk$Z5O~}&pp< _]qʱ管'<X|'⏷W(7^U>,!w?&|C TʃA)RrpbC$4> / `^(́*%Ym S)!}N铘EOw.;ܹTYPqsi΅F* 0J&'%MHLkI) mr))-0݊\T8cF2L4E`z|%2 ~5VB#cYƩ"qd*2k-UR[(A@rp3ΒK$I28/ʪ۵=rOkcW xHn{#pȞ|Fy8IzU]nsqd7^}0uu@†^XsQH㞻OϏݻ.0zX7YM^O! D̸3"U(N lHs$YyY1JN1 8ph Szu5`޺oMoAރsXs'#FvAΠk7# VJd 窭9MkӣW$%;WxLijT$A˖VG`ga?O9Z?-lfRٯ_z \z b9΋olHa_MDqC9i*xinLU/ EN^2-yTB3rM Bȋx6-A5DR'X|o|6 # |{QV܁DZnK9T]ju2:,2SJ 2 Z895a%&8I_ .d(9oVS\ WaX`6,Kጮ %gw-ZE0tevkti-)G!\ \ѴϑU2& g=cH\䎅RùęVc"z2[Yp^X#PLRFq eJ!Bg3iVVAOkR ^ q6n!6U>/uxHx)=7/bs8> E-yfx=~ތ/ m#gLg "svqI_-{34rg緷|া!۽Si`Fz_.K7T`%)<^ʮOuҪ"sxU@rϪ_;:"SX_70zMZTQH>55e/#H~pjQ1\`k +!QeSų: Uq)tњ ϬeЯxM k2f(X 3D#^iJ$)v)?EHq%|,/@Hiuw9P.`wҽ;M!+5yAIl5k;^;J/?:Y"?Vv+fE7IoVQ{Jo][oܺ+y 9k5- EQ༵($Jܓz]<^RkDzV%E/H]Y~33cI%)ӉⴅzՎ+q5ѥu7"CirLK?C9ZB)1qH NluVd3s ̀Bil(7Qn9d팁;eL]gdm,R%JhC4 -+hB8@+f"S(Qm[q4egزmJ5hݯn[ 4%G49YpxŬ*k{&:jLo&C:B˭OsunN<~/uˏ g̷7C 9܅=;UEzn ,-]z#ӢjF2AQ7+G()fCHv'm}>"SVVQ{J)Sage "88VR{o\bԚ1SBilOsCjRkJC{Ezqr+n?uM߯V쌽isXeq>__tJ>Fu>a 򽿹7GWVN'Z^V;s5:E>x{_@8X] %쾧Ce& y&bS@CձwKxT BL'1mk[:wwa!D6~IiwTo"Sa§Ο ?i b- w:Yzf.o<-ǁHL9ԧoM wZ4r?k!S4"GK>%r* 0Geiz-L;9`WԧJtiu;'7YxNTR7t@Eu kVn71A͉g:!k`e1( rio{ԻP '_Nw5jγ:2kP* 1w8͜탟7_Ea6k68"/s#fZoNV+!-pqPo60/Ħ!Ҁ XS PaݲF6k])^T*E[+b+!qp$*I=AН1LQ5ª5( ӭ&mAQ@ ~5)EFZVE"wKJ(9!% 4y2=PT7`%%6m-VhS'(szZPl.ʺjjMiaQZ6X"R 5n)T wLsE+Q18H4P.M;}7 eH{@ !OHOl g2 ~%h5aawFn瞢5O#(GMƅA Y0tY tqy9()jw/LIA93#+r 0u:+!'IP+@l9}H`uBwMO7ݏMw'ci(ЕwϬT]Dn V'Ŧȁ /Jh2v5,ϓ Wy>Oy?<:K{w rӝv0-Z# -EFM޿:@:#xvZULsu48HAI: ge6@8jRS10B^Ĝ͒{ B^qHDdO ҫQIL䉱ŖT#d?Y\ DrdտRKh&R1O}F r>zsL 49,䕛hM; D EXuT9ڳP M4ɦY}y4]wK tRݖEy*@B^lJڔa [u~",'( :Eajw߇OIq(B IY ײw&4C8VBX\%w_(EZB))QK2}io')EJIңLfZu!eX0LHp־ل7 HW۲j*댒)c!PPU.TڿR$zo!\+ qy°W4moDmpN#=2Xt؋p+$h"X`SX 8k|RC꺒h iՍ"Ƙ mP *MNGAtB㡾w8?TxDø!ЊtVU$rhtk%VRcI`+l1 I%| غiݷlnF!$osWVamVhƗe? }Yv d`b+ W|0hj"GluUb[^E,s.6RP:CPc)CulRP 5Nżm'{1'¦xHŀq1J1)BTS [ .>2ثSMG$9F*qw즌 p9Ƚy`ͻɃF]YCNmb{ rNDfiIⶉI"QRH%A$q __`)8>(B,Mwx{q12EhP{")^9-D?(Ϧf(,EsQ7~>j=8g?_ØډZ0]j@L>JKZ2jp\L^E-zcQ\2:O>i?k:OtXTtF=!m%<}6=<ײ]ago~w'ٻ.:;䡽1M+_uͅko|}|t9ۥT.6a7^<^\o~ݻ y~}V{d\V_.+Eo.EK]hoojx6W2?>,!'|nxݷ>~P4 S陦ق4täTm9۸+u1h'dw~ f~)ǺN &FښdVFvkEϦPcjkm֜m.p @ iȺfxMm_var/home/core/zuul-output/logs/kubelet.log0000644000000000000000005007676615145307347017727 0ustar rootrootFeb 18 09:04:02 crc systemd[1]: Starting Kubernetes Kubelet... Feb 18 09:04:02 crc restorecon[4555]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 18 09:04:02 crc restorecon[4555]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 18 09:04:02 crc restorecon[4555]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Feb 18 09:04:03 crc kubenswrapper[4556]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 18 09:04:03 crc kubenswrapper[4556]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Feb 18 09:04:03 crc kubenswrapper[4556]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 18 09:04:03 crc kubenswrapper[4556]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 18 09:04:03 crc kubenswrapper[4556]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Feb 18 09:04:03 crc kubenswrapper[4556]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.162927 4556 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168075 4556 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168100 4556 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168105 4556 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168109 4556 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168113 4556 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168118 4556 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168125 4556 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168129 4556 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168133 4556 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168136 4556 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168140 4556 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168144 4556 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168147 4556 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168168 4556 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168172 4556 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168176 4556 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168179 4556 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168183 4556 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168186 4556 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168189 4556 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168192 4556 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168196 4556 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168199 4556 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168202 4556 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168205 4556 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168209 4556 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168213 4556 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168216 4556 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168219 4556 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168222 4556 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168225 4556 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168229 4556 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168232 4556 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168236 4556 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168244 4556 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168247 4556 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168251 4556 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168254 4556 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168259 4556 feature_gate.go:330] unrecognized feature gate: Example Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168263 4556 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168267 4556 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168270 4556 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168273 4556 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168278 4556 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168282 4556 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168285 4556 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168289 4556 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168293 4556 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168298 4556 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168302 4556 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168305 4556 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168309 4556 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168312 4556 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168316 4556 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168319 4556 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168323 4556 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168326 4556 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168330 4556 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168333 4556 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168336 4556 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168339 4556 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168342 4556 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168345 4556 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168349 4556 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168352 4556 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168356 4556 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168359 4556 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168362 4556 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168365 4556 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168371 4556 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.168374 4556 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168446 4556 flags.go:64] FLAG: --address="0.0.0.0" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168455 4556 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168462 4556 flags.go:64] FLAG: --anonymous-auth="true" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168467 4556 flags.go:64] FLAG: --application-metrics-count-limit="100" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168472 4556 flags.go:64] FLAG: --authentication-token-webhook="false" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168476 4556 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168481 4556 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168487 4556 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168491 4556 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168495 4556 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168499 4556 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168503 4556 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168507 4556 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168511 4556 flags.go:64] FLAG: --cgroup-root="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168515 4556 flags.go:64] FLAG: --cgroups-per-qos="true" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168519 4556 flags.go:64] FLAG: --client-ca-file="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168522 4556 flags.go:64] FLAG: --cloud-config="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168526 4556 flags.go:64] FLAG: --cloud-provider="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168530 4556 flags.go:64] FLAG: --cluster-dns="[]" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168535 4556 flags.go:64] FLAG: --cluster-domain="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168539 4556 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168543 4556 flags.go:64] FLAG: --config-dir="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168547 4556 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168551 4556 flags.go:64] FLAG: --container-log-max-files="5" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168556 4556 flags.go:64] FLAG: --container-log-max-size="10Mi" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168560 4556 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168565 4556 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168570 4556 flags.go:64] FLAG: --containerd-namespace="k8s.io" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168574 4556 flags.go:64] FLAG: --contention-profiling="false" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168578 4556 flags.go:64] FLAG: --cpu-cfs-quota="true" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168584 4556 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168588 4556 flags.go:64] FLAG: --cpu-manager-policy="none" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168593 4556 flags.go:64] FLAG: --cpu-manager-policy-options="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168604 4556 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168608 4556 flags.go:64] FLAG: --enable-controller-attach-detach="true" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168612 4556 flags.go:64] FLAG: --enable-debugging-handlers="true" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168615 4556 flags.go:64] FLAG: --enable-load-reader="false" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168619 4556 flags.go:64] FLAG: --enable-server="true" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168624 4556 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168629 4556 flags.go:64] FLAG: --event-burst="100" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168633 4556 flags.go:64] FLAG: --event-qps="50" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168637 4556 flags.go:64] FLAG: --event-storage-age-limit="default=0" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168640 4556 flags.go:64] FLAG: --event-storage-event-limit="default=0" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168644 4556 flags.go:64] FLAG: --eviction-hard="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168649 4556 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168653 4556 flags.go:64] FLAG: --eviction-minimum-reclaim="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168656 4556 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168661 4556 flags.go:64] FLAG: --eviction-soft="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168665 4556 flags.go:64] FLAG: --eviction-soft-grace-period="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168669 4556 flags.go:64] FLAG: --exit-on-lock-contention="false" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168673 4556 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168676 4556 flags.go:64] FLAG: --experimental-mounter-path="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168680 4556 flags.go:64] FLAG: --fail-cgroupv1="false" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168684 4556 flags.go:64] FLAG: --fail-swap-on="true" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168688 4556 flags.go:64] FLAG: --feature-gates="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168692 4556 flags.go:64] FLAG: --file-check-frequency="20s" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168706 4556 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168711 4556 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168715 4556 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168719 4556 flags.go:64] FLAG: --healthz-port="10248" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168723 4556 flags.go:64] FLAG: --help="false" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168727 4556 flags.go:64] FLAG: --hostname-override="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168732 4556 flags.go:64] FLAG: --housekeeping-interval="10s" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168736 4556 flags.go:64] FLAG: --http-check-frequency="20s" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168740 4556 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168744 4556 flags.go:64] FLAG: --image-credential-provider-config="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168748 4556 flags.go:64] FLAG: --image-gc-high-threshold="85" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168752 4556 flags.go:64] FLAG: --image-gc-low-threshold="80" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168756 4556 flags.go:64] FLAG: --image-service-endpoint="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168760 4556 flags.go:64] FLAG: --kernel-memcg-notification="false" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168764 4556 flags.go:64] FLAG: --kube-api-burst="100" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168768 4556 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168772 4556 flags.go:64] FLAG: --kube-api-qps="50" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168775 4556 flags.go:64] FLAG: --kube-reserved="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168779 4556 flags.go:64] FLAG: --kube-reserved-cgroup="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168783 4556 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168787 4556 flags.go:64] FLAG: --kubelet-cgroups="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168790 4556 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168794 4556 flags.go:64] FLAG: --lock-file="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168798 4556 flags.go:64] FLAG: --log-cadvisor-usage="false" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168802 4556 flags.go:64] FLAG: --log-flush-frequency="5s" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168806 4556 flags.go:64] FLAG: --log-json-info-buffer-size="0" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168811 4556 flags.go:64] FLAG: --log-json-split-stream="false" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168816 4556 flags.go:64] FLAG: --log-text-info-buffer-size="0" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168820 4556 flags.go:64] FLAG: --log-text-split-stream="false" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168823 4556 flags.go:64] FLAG: --logging-format="text" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168827 4556 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168831 4556 flags.go:64] FLAG: --make-iptables-util-chains="true" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168835 4556 flags.go:64] FLAG: --manifest-url="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168839 4556 flags.go:64] FLAG: --manifest-url-header="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168844 4556 flags.go:64] FLAG: --max-housekeeping-interval="15s" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168848 4556 flags.go:64] FLAG: --max-open-files="1000000" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168853 4556 flags.go:64] FLAG: --max-pods="110" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168856 4556 flags.go:64] FLAG: --maximum-dead-containers="-1" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168861 4556 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168865 4556 flags.go:64] FLAG: --memory-manager-policy="None" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168869 4556 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168873 4556 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168885 4556 flags.go:64] FLAG: --node-ip="192.168.126.11" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168889 4556 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168899 4556 flags.go:64] FLAG: --node-status-max-images="50" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168903 4556 flags.go:64] FLAG: --node-status-update-frequency="10s" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168907 4556 flags.go:64] FLAG: --oom-score-adj="-999" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168911 4556 flags.go:64] FLAG: --pod-cidr="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168915 4556 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168921 4556 flags.go:64] FLAG: --pod-manifest-path="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168924 4556 flags.go:64] FLAG: --pod-max-pids="-1" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168928 4556 flags.go:64] FLAG: --pods-per-core="0" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168932 4556 flags.go:64] FLAG: --port="10250" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168936 4556 flags.go:64] FLAG: --protect-kernel-defaults="false" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168940 4556 flags.go:64] FLAG: --provider-id="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168944 4556 flags.go:64] FLAG: --qos-reserved="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168947 4556 flags.go:64] FLAG: --read-only-port="10255" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168951 4556 flags.go:64] FLAG: --register-node="true" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168955 4556 flags.go:64] FLAG: --register-schedulable="true" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168958 4556 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168968 4556 flags.go:64] FLAG: --registry-burst="10" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168972 4556 flags.go:64] FLAG: --registry-qps="5" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168976 4556 flags.go:64] FLAG: --reserved-cpus="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168982 4556 flags.go:64] FLAG: --reserved-memory="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168987 4556 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168991 4556 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168995 4556 flags.go:64] FLAG: --rotate-certificates="false" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.168999 4556 flags.go:64] FLAG: --rotate-server-certificates="false" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.169002 4556 flags.go:64] FLAG: --runonce="false" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.169006 4556 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.169011 4556 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.169014 4556 flags.go:64] FLAG: --seccomp-default="false" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.169018 4556 flags.go:64] FLAG: --serialize-image-pulls="true" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.169022 4556 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.169026 4556 flags.go:64] FLAG: --storage-driver-db="cadvisor" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.169031 4556 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.169035 4556 flags.go:64] FLAG: --storage-driver-password="root" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.169039 4556 flags.go:64] FLAG: --storage-driver-secure="false" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.169043 4556 flags.go:64] FLAG: --storage-driver-table="stats" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.169046 4556 flags.go:64] FLAG: --storage-driver-user="root" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.169050 4556 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.169054 4556 flags.go:64] FLAG: --sync-frequency="1m0s" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.169058 4556 flags.go:64] FLAG: --system-cgroups="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.169062 4556 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.169067 4556 flags.go:64] FLAG: --system-reserved-cgroup="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.169071 4556 flags.go:64] FLAG: --tls-cert-file="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.169075 4556 flags.go:64] FLAG: --tls-cipher-suites="[]" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.169080 4556 flags.go:64] FLAG: --tls-min-version="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.169084 4556 flags.go:64] FLAG: --tls-private-key-file="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.169087 4556 flags.go:64] FLAG: --topology-manager-policy="none" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.169091 4556 flags.go:64] FLAG: --topology-manager-policy-options="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.169094 4556 flags.go:64] FLAG: --topology-manager-scope="container" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.169098 4556 flags.go:64] FLAG: --v="2" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.169103 4556 flags.go:64] FLAG: --version="false" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.169108 4556 flags.go:64] FLAG: --vmodule="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.169113 4556 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.169118 4556 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169219 4556 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169225 4556 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169229 4556 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169233 4556 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169236 4556 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169240 4556 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169244 4556 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169248 4556 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169252 4556 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169255 4556 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169260 4556 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169263 4556 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169266 4556 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169270 4556 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169273 4556 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169276 4556 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169280 4556 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169283 4556 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169287 4556 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169291 4556 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169295 4556 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169299 4556 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169302 4556 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169305 4556 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169308 4556 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169312 4556 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169315 4556 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169318 4556 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169321 4556 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169325 4556 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169328 4556 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169331 4556 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169334 4556 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169337 4556 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169340 4556 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169343 4556 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169347 4556 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169350 4556 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169354 4556 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169357 4556 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169362 4556 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169366 4556 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169369 4556 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169372 4556 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169375 4556 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169378 4556 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169382 4556 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169385 4556 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169389 4556 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169392 4556 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169396 4556 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169400 4556 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169403 4556 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169406 4556 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169410 4556 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169413 4556 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169417 4556 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169420 4556 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169424 4556 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169427 4556 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169430 4556 feature_gate.go:330] unrecognized feature gate: Example Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169433 4556 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169437 4556 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169441 4556 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169445 4556 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169448 4556 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169452 4556 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169455 4556 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169459 4556 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169466 4556 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.169469 4556 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.169475 4556 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.176447 4556 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.176482 4556 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176550 4556 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176564 4556 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176568 4556 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176571 4556 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176575 4556 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176578 4556 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176582 4556 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176587 4556 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176593 4556 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176598 4556 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176602 4556 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176605 4556 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176609 4556 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176612 4556 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176616 4556 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176619 4556 feature_gate.go:330] unrecognized feature gate: Example Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176622 4556 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176626 4556 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176629 4556 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176632 4556 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176635 4556 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176638 4556 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176641 4556 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176644 4556 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176648 4556 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176652 4556 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176655 4556 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176658 4556 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176662 4556 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176666 4556 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176671 4556 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176674 4556 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176678 4556 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176681 4556 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176685 4556 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176690 4556 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176694 4556 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176698 4556 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176701 4556 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176705 4556 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176708 4556 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176711 4556 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176714 4556 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176718 4556 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176721 4556 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176724 4556 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176727 4556 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176731 4556 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176734 4556 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176738 4556 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176741 4556 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176744 4556 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176748 4556 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176751 4556 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176754 4556 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176758 4556 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176761 4556 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176764 4556 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176767 4556 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176771 4556 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176774 4556 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176777 4556 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176781 4556 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176784 4556 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176787 4556 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176791 4556 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176794 4556 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176797 4556 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176800 4556 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176804 4556 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176813 4556 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.176819 4556 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176942 4556 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176947 4556 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176952 4556 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176955 4556 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176958 4556 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176963 4556 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176967 4556 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176971 4556 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176974 4556 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176978 4556 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176981 4556 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176984 4556 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176988 4556 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176991 4556 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176994 4556 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.176997 4556 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.177000 4556 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.177004 4556 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.177007 4556 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.177010 4556 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.177014 4556 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.177017 4556 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.177021 4556 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.177025 4556 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.177028 4556 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.177031 4556 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.177035 4556 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.177038 4556 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.177041 4556 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.177044 4556 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.177047 4556 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.177051 4556 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.177054 4556 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.177057 4556 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.177065 4556 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.177069 4556 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.177072 4556 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.177075 4556 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.177078 4556 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.177081 4556 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.177084 4556 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.177087 4556 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.177090 4556 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.177094 4556 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.177097 4556 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.177100 4556 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.177104 4556 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.177107 4556 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.177110 4556 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.177115 4556 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.177119 4556 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.177122 4556 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.177126 4556 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.177130 4556 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.177136 4556 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.177139 4556 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.177143 4556 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.177146 4556 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.177167 4556 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.177172 4556 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.177175 4556 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.177179 4556 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.177183 4556 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.177188 4556 feature_gate.go:330] unrecognized feature gate: Example Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.177191 4556 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.177194 4556 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.177198 4556 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.177201 4556 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.177204 4556 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.177207 4556 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.177215 4556 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.177221 4556 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.177363 4556 server.go:940] "Client rotation is on, will bootstrap in background" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.180048 4556 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.180120 4556 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.180977 4556 server.go:997] "Starting client certificate rotation" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.181002 4556 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.181187 4556 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-28 04:57:07.979159269 +0000 UTC Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.181255 4556 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.191004 4556 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 18 09:04:03 crc kubenswrapper[4556]: E0218 09:04:03.192998 4556 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 192.168.26.217:6443: connect: connection refused" logger="UnhandledError" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.194167 4556 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.202816 4556 log.go:25] "Validated CRI v1 runtime API" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.218932 4556 log.go:25] "Validated CRI v1 image API" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.220093 4556 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.222414 4556 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2026-02-18-09-00-46-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.222438 4556 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:49 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/containers/storage/overlay-containers/75d81934760b26101869fbd8e4b5954c62b019c1cc3e5a0c9f82ed8de46b3b22/userdata/shm:{mountpoint:/var/lib/containers/storage/overlay-containers/75d81934760b26101869fbd8e4b5954c62b019c1cc3e5a0c9f82ed8de46b3b22/userdata/shm major:0 minor:42 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:50 fsType:tmpfs blockSize:0} overlay_0-43:{mountpoint:/var/lib/containers/storage/overlay/94b752e0a51c0134b00ddef6dc7a933a9d7c1d9bdc88a18dae4192a0d557d623/merged major:0 minor:43 fsType:overlay blockSize:0}] Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.235817 4556 manager.go:217] Machine: {Timestamp:2026-02-18 09:04:03.234462797 +0000 UTC m=+0.251423798 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2445404 MemoryCapacity:33654116352 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:0f904183-79bf-4790-8551-43a6eb3adbe4 BootID:8758c02e-ef59-46a7-908c-d5c97feb8ceb Filesystems:[{Device:/var/lib/containers/storage/overlay-containers/75d81934760b26101869fbd8e4b5954c62b019c1cc3e5a0c9f82ed8de46b3b22/userdata/shm DeviceMajor:0 DeviceMinor:42 Capacity:65536000 Type:vfs Inodes:4108168 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:50 Capacity:1073741824 Type:vfs Inodes:4108168 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827056128 Type:vfs Inodes:4108168 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827060224 Type:vfs Inodes:1048576 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:49 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:overlay_0-43 DeviceMajor:0 DeviceMinor:43 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:09:a1:2f Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:enp3s0 MacAddress:fa:16:3e:09:a1:2f Speed:-1 Mtu:1500} {Name:enp7s0 MacAddress:fa:16:3e:59:1d:cb Speed:-1 Mtu:1440} {Name:enp7s0.20 MacAddress:52:54:00:9e:76:84 Speed:-1 Mtu:1436} {Name:enp7s0.21 MacAddress:52:54:00:15:f1:87 Speed:-1 Mtu:1436} {Name:enp7s0.22 MacAddress:52:54:00:8d:27:fe Speed:-1 Mtu:1436} {Name:eth10 MacAddress:96:bd:ba:d6:30:59 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:fa:f9:2a:26:31:56 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654116352 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:65536 Type:Data Level:1} {Id:0 Size:65536 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:65536 Type:Data Level:1} {Id:1 Size:65536 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:65536 Type:Data Level:1} {Id:10 Size:65536 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:65536 Type:Data Level:1} {Id:11 Size:65536 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:65536 Type:Data Level:1} {Id:2 Size:65536 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:65536 Type:Data Level:1} {Id:3 Size:65536 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:65536 Type:Data Level:1} {Id:4 Size:65536 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:65536 Type:Data Level:1} {Id:5 Size:65536 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:65536 Type:Data Level:1} {Id:6 Size:65536 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:65536 Type:Data Level:1} {Id:7 Size:65536 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:65536 Type:Data Level:1} {Id:8 Size:65536 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:65536 Type:Data Level:1} {Id:9 Size:65536 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.235955 4556 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.236050 4556 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.236752 4556 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.236892 4556 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.236923 4556 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.237066 4556 topology_manager.go:138] "Creating topology manager with none policy" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.237075 4556 container_manager_linux.go:303] "Creating device plugin manager" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.237350 4556 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.237376 4556 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.237470 4556 state_mem.go:36] "Initialized new in-memory state store" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.237537 4556 server.go:1245] "Using root directory" path="/var/lib/kubelet" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.239130 4556 kubelet.go:418] "Attempting to sync node with API server" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.239147 4556 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.239180 4556 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.239189 4556 kubelet.go:324] "Adding apiserver pod source" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.239198 4556 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.241443 4556 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.241687 4556 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 192.168.26.217:6443: connect: connection refused Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.241694 4556 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 192.168.26.217:6443: connect: connection refused Feb 18 09:04:03 crc kubenswrapper[4556]: E0218 09:04:03.241743 4556 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 192.168.26.217:6443: connect: connection refused" logger="UnhandledError" Feb 18 09:04:03 crc kubenswrapper[4556]: E0218 09:04:03.241746 4556 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 192.168.26.217:6443: connect: connection refused" logger="UnhandledError" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.242076 4556 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.243700 4556 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.244593 4556 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.244615 4556 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.244622 4556 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.244630 4556 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.244639 4556 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.244645 4556 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.244651 4556 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.244660 4556 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.244668 4556 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.244674 4556 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.244683 4556 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.244690 4556 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.245138 4556 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.245562 4556 server.go:1280] "Started kubelet" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.245928 4556 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 192.168.26.217:6443: connect: connection refused Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.246203 4556 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.246276 4556 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.246610 4556 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Feb 18 09:04:03 crc systemd[1]: Started Kubernetes Kubelet. Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.247348 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.247369 4556 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.247422 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-23 01:12:39.281518358 +0000 UTC Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.247471 4556 volume_manager.go:287] "The desired_state_of_world populator starts" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.247477 4556 volume_manager.go:289] "Starting Kubelet Volume Manager" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.247495 4556 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Feb 18 09:04:03 crc kubenswrapper[4556]: E0218 09:04:03.247502 4556 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.248761 4556 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 192.168.26.217:6443: connect: connection refused Feb 18 09:04:03 crc kubenswrapper[4556]: E0218 09:04:03.248820 4556 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 192.168.26.217:6443: connect: connection refused" logger="UnhandledError" Feb 18 09:04:03 crc kubenswrapper[4556]: E0218 09:04:03.248932 4556 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.217:6443: connect: connection refused" interval="200ms" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.248974 4556 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.249089 4556 factory.go:55] Registering systemd factory Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.249138 4556 factory.go:221] Registration of the systemd container factory successfully Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.249488 4556 server.go:460] "Adding debug handlers to kubelet server" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.249903 4556 factory.go:153] Registering CRI-O factory Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.249923 4556 factory.go:221] Registration of the crio container factory successfully Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.249948 4556 factory.go:103] Registering Raw factory Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.249962 4556 manager.go:1196] Started watching for new ooms in manager Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.250836 4556 manager.go:319] Starting recovery of all containers Feb 18 09:04:03 crc kubenswrapper[4556]: E0218 09:04:03.250315 4556 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 192.168.26.217:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.18954bdd85d7f8dd default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-18 09:04:03.245537501 +0000 UTC m=+0.262498481,LastTimestamp:2026-02-18 09:04:03.245537501 +0000 UTC m=+0.262498481,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.260713 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.260763 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.260775 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.260788 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.260797 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.260806 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.260814 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.260821 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.260830 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.260838 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.260846 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.260854 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.260862 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.260883 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.260892 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.260902 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.260910 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.260919 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.260927 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.260935 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.260943 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.260951 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.260959 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.260966 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.260974 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.260982 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.260991 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.260999 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.261007 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.261014 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.261030 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.261039 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.261046 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.261056 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.261064 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.261071 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.261079 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.261087 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.261094 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.261102 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.261109 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.261118 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.261126 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.261133 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.261140 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.261147 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.261170 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.261179 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.261187 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.261194 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.261220 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.261229 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.261241 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.261250 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.261261 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.261270 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.261278 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.261285 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.261293 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.261300 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.261309 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.261317 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.261325 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.261333 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.261341 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.261349 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.261357 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.261365 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.261373 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262257 4556 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262277 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262288 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262297 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262305 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262312 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262320 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262327 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262336 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262346 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262354 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262364 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262373 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262382 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262390 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262397 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262405 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262412 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262420 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262427 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262445 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262454 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262461 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262470 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262477 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262485 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262492 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262501 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262510 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262518 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262526 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262534 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262541 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262549 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262556 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262565 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262577 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262586 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262595 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262603 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262611 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262620 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262629 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262638 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262646 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262654 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262663 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262670 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262677 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262685 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262692 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262699 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262706 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262713 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262720 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262746 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262753 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262760 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262797 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262805 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262813 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262842 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262850 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262857 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262865 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262876 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262894 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262902 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262909 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262916 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262923 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262932 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262939 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262946 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.262987 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263005 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263012 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263020 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263040 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263048 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263091 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263099 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263115 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263122 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263129 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263136 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263143 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263172 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263180 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263186 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263207 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263214 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263221 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263228 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263246 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263264 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263272 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263280 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263296 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263328 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263335 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263343 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263371 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263379 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263390 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263397 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263416 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263432 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263447 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263453 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263460 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263467 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263475 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263482 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263506 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263513 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263546 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263554 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263564 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263588 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263614 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263622 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263642 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263650 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263656 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263675 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263690 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263700 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263707 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263720 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263779 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263786 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263793 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263800 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263807 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263815 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263837 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263862 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263878 4556 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263916 4556 reconstruct.go:97] "Volume reconstruction finished" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.263922 4556 reconciler.go:26] "Reconciler: start to sync state" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.270356 4556 manager.go:324] Recovery completed Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.279202 4556 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.279568 4556 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.280284 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.280320 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.280328 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.280846 4556 cpu_manager.go:225] "Starting CPU manager" policy="none" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.280936 4556 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.281020 4556 state_mem.go:36] "Initialized new in-memory state store" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.280936 4556 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.281201 4556 status_manager.go:217] "Starting to sync pod status with apiserver" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.281223 4556 kubelet.go:2335] "Starting kubelet main sync loop" Feb 18 09:04:03 crc kubenswrapper[4556]: E0218 09:04:03.281273 4556 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.282363 4556 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 192.168.26.217:6443: connect: connection refused Feb 18 09:04:03 crc kubenswrapper[4556]: E0218 09:04:03.282413 4556 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 192.168.26.217:6443: connect: connection refused" logger="UnhandledError" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.285495 4556 policy_none.go:49] "None policy: Start" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.286087 4556 memory_manager.go:170] "Starting memorymanager" policy="None" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.286198 4556 state_mem.go:35] "Initializing new in-memory state store" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.323071 4556 manager.go:334] "Starting Device Plugin manager" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.323116 4556 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.323143 4556 server.go:79] "Starting device plugin registration server" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.323420 4556 eviction_manager.go:189] "Eviction manager: starting control loop" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.323436 4556 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.323571 4556 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.323634 4556 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.323650 4556 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Feb 18 09:04:03 crc kubenswrapper[4556]: E0218 09:04:03.330429 4556 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.382025 4556 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.382185 4556 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.383446 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.383474 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.383482 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.383627 4556 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.383798 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.383833 4556 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.384228 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.384279 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.384289 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.384398 4556 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.384452 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.384494 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.384498 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.384523 4556 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.384503 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.385103 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.385130 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.385141 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.385335 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.385415 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.385429 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.385518 4556 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.385603 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.385625 4556 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.386074 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.386094 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.386102 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.386910 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.386935 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.386943 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.387056 4556 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.387179 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.387211 4556 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.387641 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.387664 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.387674 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.387798 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.387822 4556 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.387852 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.387950 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.387968 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.388332 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.388356 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.388364 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.423578 4556 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.424333 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.424360 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.424369 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.424409 4556 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 18 09:04:03 crc kubenswrapper[4556]: E0218 09:04:03.424780 4556 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 192.168.26.217:6443: connect: connection refused" node="crc" Feb 18 09:04:03 crc kubenswrapper[4556]: E0218 09:04:03.450423 4556 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.217:6443: connect: connection refused" interval="400ms" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.465775 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.465811 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.465834 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.465882 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.465919 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.465943 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.465962 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.465983 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.465998 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.466012 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.466028 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.466044 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.466085 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.466102 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.466165 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.567024 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.567110 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.567127 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.567145 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.567177 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.567191 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.567206 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.567220 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.567236 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.567250 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.567267 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.567270 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.567287 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.567308 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.567304 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.567283 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.567337 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.567332 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.567363 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.567351 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.567355 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.567408 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.567389 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.567391 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.567354 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.567455 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.567458 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.567477 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.567525 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.567547 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.625029 4556 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.627171 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.627214 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.627223 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.627244 4556 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 18 09:04:03 crc kubenswrapper[4556]: E0218 09:04:03.627585 4556 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 192.168.26.217:6443: connect: connection refused" node="crc" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.714252 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.721294 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.725090 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.736580 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-32cfdeb09755ae943958e60a1d12b941b03678fa28898c2a8a15851b6caa57d6 WatchSource:0}: Error finding container 32cfdeb09755ae943958e60a1d12b941b03678fa28898c2a8a15851b6caa57d6: Status 404 returned error can't find the container with id 32cfdeb09755ae943958e60a1d12b941b03678fa28898c2a8a15851b6caa57d6 Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.737710 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-3c868945020b2ba25d90df68867c2ebcc6e5cf97331f39d022ac78181e4a0d1d WatchSource:0}: Error finding container 3c868945020b2ba25d90df68867c2ebcc6e5cf97331f39d022ac78181e4a0d1d: Status 404 returned error can't find the container with id 3c868945020b2ba25d90df68867c2ebcc6e5cf97331f39d022ac78181e4a0d1d Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.739879 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-06eb0c4f2fe251786e81df4c9c12398f3a71a09ba50d9fe60197b59c1f71540f WatchSource:0}: Error finding container 06eb0c4f2fe251786e81df4c9c12398f3a71a09ba50d9fe60197b59c1f71540f: Status 404 returned error can't find the container with id 06eb0c4f2fe251786e81df4c9c12398f3a71a09ba50d9fe60197b59c1f71540f Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.752797 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 18 09:04:03 crc kubenswrapper[4556]: I0218 09:04:03.757539 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.763969 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-bfb82f229d0a437dfa4b59534ea4128dca62f97f249b0dca30e83924d06517fb WatchSource:0}: Error finding container bfb82f229d0a437dfa4b59534ea4128dca62f97f249b0dca30e83924d06517fb: Status 404 returned error can't find the container with id bfb82f229d0a437dfa4b59534ea4128dca62f97f249b0dca30e83924d06517fb Feb 18 09:04:03 crc kubenswrapper[4556]: W0218 09:04:03.767620 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-3cd94c175504b554637950436b37dce19304f1f60c53a7ad3c6deb14e5bc78c2 WatchSource:0}: Error finding container 3cd94c175504b554637950436b37dce19304f1f60c53a7ad3c6deb14e5bc78c2: Status 404 returned error can't find the container with id 3cd94c175504b554637950436b37dce19304f1f60c53a7ad3c6deb14e5bc78c2 Feb 18 09:04:03 crc kubenswrapper[4556]: E0218 09:04:03.851443 4556 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.217:6443: connect: connection refused" interval="800ms" Feb 18 09:04:04 crc kubenswrapper[4556]: I0218 09:04:04.028612 4556 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 09:04:04 crc kubenswrapper[4556]: I0218 09:04:04.030279 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:04 crc kubenswrapper[4556]: I0218 09:04:04.030311 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:04 crc kubenswrapper[4556]: I0218 09:04:04.030320 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:04 crc kubenswrapper[4556]: I0218 09:04:04.030343 4556 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 18 09:04:04 crc kubenswrapper[4556]: E0218 09:04:04.030757 4556 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 192.168.26.217:6443: connect: connection refused" node="crc" Feb 18 09:04:04 crc kubenswrapper[4556]: I0218 09:04:04.246912 4556 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 192.168.26.217:6443: connect: connection refused Feb 18 09:04:04 crc kubenswrapper[4556]: I0218 09:04:04.247977 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-19 15:23:32.786998765 +0000 UTC Feb 18 09:04:04 crc kubenswrapper[4556]: W0218 09:04:04.256525 4556 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 192.168.26.217:6443: connect: connection refused Feb 18 09:04:04 crc kubenswrapper[4556]: E0218 09:04:04.256583 4556 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 192.168.26.217:6443: connect: connection refused" logger="UnhandledError" Feb 18 09:04:04 crc kubenswrapper[4556]: I0218 09:04:04.285595 4556 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d" exitCode=0 Feb 18 09:04:04 crc kubenswrapper[4556]: I0218 09:04:04.285650 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d"} Feb 18 09:04:04 crc kubenswrapper[4556]: I0218 09:04:04.285718 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"32cfdeb09755ae943958e60a1d12b941b03678fa28898c2a8a15851b6caa57d6"} Feb 18 09:04:04 crc kubenswrapper[4556]: I0218 09:04:04.285794 4556 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 09:04:04 crc kubenswrapper[4556]: I0218 09:04:04.288838 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:04 crc kubenswrapper[4556]: I0218 09:04:04.288873 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:04 crc kubenswrapper[4556]: I0218 09:04:04.288884 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:04 crc kubenswrapper[4556]: I0218 09:04:04.289578 4556 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="a6e0422278544bddfec087d7b0a94bedc55a7ac67ff280a36030528eb87fd72b" exitCode=0 Feb 18 09:04:04 crc kubenswrapper[4556]: I0218 09:04:04.289689 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"a6e0422278544bddfec087d7b0a94bedc55a7ac67ff280a36030528eb87fd72b"} Feb 18 09:04:04 crc kubenswrapper[4556]: I0218 09:04:04.289792 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"3cd94c175504b554637950436b37dce19304f1f60c53a7ad3c6deb14e5bc78c2"} Feb 18 09:04:04 crc kubenswrapper[4556]: I0218 09:04:04.289872 4556 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 09:04:04 crc kubenswrapper[4556]: I0218 09:04:04.290582 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:04 crc kubenswrapper[4556]: I0218 09:04:04.290615 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:04 crc kubenswrapper[4556]: I0218 09:04:04.290625 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:04 crc kubenswrapper[4556]: I0218 09:04:04.291346 4556 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="dfbda9799e2d2bc16964e79b90fb7087f9e6893d867711612c63aaaa3dd2b141" exitCode=0 Feb 18 09:04:04 crc kubenswrapper[4556]: I0218 09:04:04.291413 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"dfbda9799e2d2bc16964e79b90fb7087f9e6893d867711612c63aaaa3dd2b141"} Feb 18 09:04:04 crc kubenswrapper[4556]: I0218 09:04:04.291446 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"bfb82f229d0a437dfa4b59534ea4128dca62f97f249b0dca30e83924d06517fb"} Feb 18 09:04:04 crc kubenswrapper[4556]: I0218 09:04:04.291523 4556 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 09:04:04 crc kubenswrapper[4556]: I0218 09:04:04.292086 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:04 crc kubenswrapper[4556]: I0218 09:04:04.292109 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:04 crc kubenswrapper[4556]: I0218 09:04:04.292120 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:04 crc kubenswrapper[4556]: I0218 09:04:04.292871 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"ac09c421b3101887bfb83107dbe26f2b5856d9b32a9cea92f6233fb6ae218cc1"} Feb 18 09:04:04 crc kubenswrapper[4556]: I0218 09:04:04.292910 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"06eb0c4f2fe251786e81df4c9c12398f3a71a09ba50d9fe60197b59c1f71540f"} Feb 18 09:04:04 crc kubenswrapper[4556]: I0218 09:04:04.294003 4556 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e" exitCode=0 Feb 18 09:04:04 crc kubenswrapper[4556]: I0218 09:04:04.294029 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e"} Feb 18 09:04:04 crc kubenswrapper[4556]: I0218 09:04:04.294046 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3c868945020b2ba25d90df68867c2ebcc6e5cf97331f39d022ac78181e4a0d1d"} Feb 18 09:04:04 crc kubenswrapper[4556]: I0218 09:04:04.294144 4556 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 09:04:04 crc kubenswrapper[4556]: I0218 09:04:04.294762 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:04 crc kubenswrapper[4556]: I0218 09:04:04.294801 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:04 crc kubenswrapper[4556]: I0218 09:04:04.294815 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:04 crc kubenswrapper[4556]: I0218 09:04:04.297385 4556 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 09:04:04 crc kubenswrapper[4556]: I0218 09:04:04.298168 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:04 crc kubenswrapper[4556]: I0218 09:04:04.298193 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:04 crc kubenswrapper[4556]: I0218 09:04:04.298203 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:04 crc kubenswrapper[4556]: W0218 09:04:04.314848 4556 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 192.168.26.217:6443: connect: connection refused Feb 18 09:04:04 crc kubenswrapper[4556]: E0218 09:04:04.314903 4556 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 192.168.26.217:6443: connect: connection refused" logger="UnhandledError" Feb 18 09:04:04 crc kubenswrapper[4556]: W0218 09:04:04.419963 4556 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 192.168.26.217:6443: connect: connection refused Feb 18 09:04:04 crc kubenswrapper[4556]: E0218 09:04:04.420030 4556 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 192.168.26.217:6443: connect: connection refused" logger="UnhandledError" Feb 18 09:04:04 crc kubenswrapper[4556]: E0218 09:04:04.652539 4556 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.217:6443: connect: connection refused" interval="1.6s" Feb 18 09:04:04 crc kubenswrapper[4556]: W0218 09:04:04.817034 4556 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 192.168.26.217:6443: connect: connection refused Feb 18 09:04:04 crc kubenswrapper[4556]: E0218 09:04:04.817102 4556 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 192.168.26.217:6443: connect: connection refused" logger="UnhandledError" Feb 18 09:04:04 crc kubenswrapper[4556]: I0218 09:04:04.831686 4556 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 09:04:04 crc kubenswrapper[4556]: I0218 09:04:04.832622 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:04 crc kubenswrapper[4556]: I0218 09:04:04.832652 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:04 crc kubenswrapper[4556]: I0218 09:04:04.832662 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:04 crc kubenswrapper[4556]: I0218 09:04:04.832682 4556 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 18 09:04:04 crc kubenswrapper[4556]: E0218 09:04:04.832997 4556 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 192.168.26.217:6443: connect: connection refused" node="crc" Feb 18 09:04:05 crc kubenswrapper[4556]: I0218 09:04:05.223363 4556 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 18 09:04:05 crc kubenswrapper[4556]: I0218 09:04:05.248205 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-11 07:46:31.996844797 +0000 UTC Feb 18 09:04:05 crc kubenswrapper[4556]: I0218 09:04:05.298112 4556 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0" exitCode=0 Feb 18 09:04:05 crc kubenswrapper[4556]: I0218 09:04:05.298195 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0"} Feb 18 09:04:05 crc kubenswrapper[4556]: I0218 09:04:05.298308 4556 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 09:04:05 crc kubenswrapper[4556]: I0218 09:04:05.298991 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:05 crc kubenswrapper[4556]: I0218 09:04:05.299016 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:05 crc kubenswrapper[4556]: I0218 09:04:05.299028 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:05 crc kubenswrapper[4556]: I0218 09:04:05.300269 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"e6ac8fc80676085932e05da4ff18e892eebaec46d096c7c343b667864f6c5e3e"} Feb 18 09:04:05 crc kubenswrapper[4556]: I0218 09:04:05.300325 4556 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 09:04:05 crc kubenswrapper[4556]: I0218 09:04:05.301025 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:05 crc kubenswrapper[4556]: I0218 09:04:05.301061 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:05 crc kubenswrapper[4556]: I0218 09:04:05.301070 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:05 crc kubenswrapper[4556]: I0218 09:04:05.302933 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"a1fa733477ee6330835d52333d6e9588626b6fc7ae08ae450dd2e1821447e995"} Feb 18 09:04:05 crc kubenswrapper[4556]: I0218 09:04:05.302958 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"28a600e8daaa2796b2f0cc6ce4112f8d41c611b5b6ca5a7eb0f260dab615caa9"} Feb 18 09:04:05 crc kubenswrapper[4556]: I0218 09:04:05.302969 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"52ea6a334e8cad154e46eb2f917cdac914557556bae88f922433a9545cb29eb9"} Feb 18 09:04:05 crc kubenswrapper[4556]: I0218 09:04:05.303032 4556 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 09:04:05 crc kubenswrapper[4556]: I0218 09:04:05.303674 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:05 crc kubenswrapper[4556]: I0218 09:04:05.303714 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:05 crc kubenswrapper[4556]: I0218 09:04:05.303728 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:05 crc kubenswrapper[4556]: I0218 09:04:05.305437 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"095ecf7e8763fa098b06eb9a8aac9f39451f7762bf20196e0803bbc841610e02"} Feb 18 09:04:05 crc kubenswrapper[4556]: I0218 09:04:05.305478 4556 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 09:04:05 crc kubenswrapper[4556]: I0218 09:04:05.305483 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"ce79cfe0242ecf403f989dc6b1c1b8be5ab54eee75ac413a8f9d03cef8bca0ed"} Feb 18 09:04:05 crc kubenswrapper[4556]: I0218 09:04:05.305496 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"77b0269625b7c6baab44a40373618f4ccfe9e1a028a1bc746b7d60c20ac4efaf"} Feb 18 09:04:05 crc kubenswrapper[4556]: I0218 09:04:05.306145 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:05 crc kubenswrapper[4556]: I0218 09:04:05.306186 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:05 crc kubenswrapper[4556]: I0218 09:04:05.306195 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:05 crc kubenswrapper[4556]: I0218 09:04:05.308249 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"072e865e6a50c03519ce22958ac4fcbb911fc140c330b6d7ffe5c9d7dd381e96"} Feb 18 09:04:05 crc kubenswrapper[4556]: I0218 09:04:05.308276 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"974e17ae7082631df2d1bcee60c8a5f142bae2659e9d200d537a4d650f401b4b"} Feb 18 09:04:05 crc kubenswrapper[4556]: I0218 09:04:05.308286 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"14f794063db9d33b82b79acc7134b5a2a3435e09d7cffe49d49251d1f7d20402"} Feb 18 09:04:05 crc kubenswrapper[4556]: I0218 09:04:05.308295 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5eb3d3728da247903bece56bc2481547317e386e9e97b6c69926a3b55dc89b8e"} Feb 18 09:04:05 crc kubenswrapper[4556]: I0218 09:04:05.308302 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c3a99dc228d1c4cb8aeab2140ec5f37358b65684739f833fbde45822782430a8"} Feb 18 09:04:05 crc kubenswrapper[4556]: I0218 09:04:05.308364 4556 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 09:04:05 crc kubenswrapper[4556]: I0218 09:04:05.308890 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:05 crc kubenswrapper[4556]: I0218 09:04:05.308940 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:05 crc kubenswrapper[4556]: I0218 09:04:05.308952 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:05 crc kubenswrapper[4556]: I0218 09:04:05.498938 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 09:04:06 crc kubenswrapper[4556]: I0218 09:04:06.248314 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-08 17:20:08.555204356 +0000 UTC Feb 18 09:04:06 crc kubenswrapper[4556]: I0218 09:04:06.312521 4556 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25" exitCode=0 Feb 18 09:04:06 crc kubenswrapper[4556]: I0218 09:04:06.312852 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25"} Feb 18 09:04:06 crc kubenswrapper[4556]: I0218 09:04:06.312963 4556 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 18 09:04:06 crc kubenswrapper[4556]: I0218 09:04:06.313023 4556 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 09:04:06 crc kubenswrapper[4556]: I0218 09:04:06.313035 4556 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 09:04:06 crc kubenswrapper[4556]: I0218 09:04:06.313030 4556 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 09:04:06 crc kubenswrapper[4556]: I0218 09:04:06.313026 4556 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 09:04:06 crc kubenswrapper[4556]: I0218 09:04:06.314332 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:06 crc kubenswrapper[4556]: I0218 09:04:06.314364 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:06 crc kubenswrapper[4556]: I0218 09:04:06.314382 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:06 crc kubenswrapper[4556]: I0218 09:04:06.314916 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:06 crc kubenswrapper[4556]: I0218 09:04:06.314944 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:06 crc kubenswrapper[4556]: I0218 09:04:06.314982 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:06 crc kubenswrapper[4556]: I0218 09:04:06.316211 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:06 crc kubenswrapper[4556]: I0218 09:04:06.316233 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:06 crc kubenswrapper[4556]: I0218 09:04:06.316242 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:06 crc kubenswrapper[4556]: I0218 09:04:06.316264 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:06 crc kubenswrapper[4556]: I0218 09:04:06.316287 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:06 crc kubenswrapper[4556]: I0218 09:04:06.316302 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:06 crc kubenswrapper[4556]: I0218 09:04:06.433192 4556 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 09:04:06 crc kubenswrapper[4556]: I0218 09:04:06.433900 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:06 crc kubenswrapper[4556]: I0218 09:04:06.433927 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:06 crc kubenswrapper[4556]: I0218 09:04:06.433936 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:06 crc kubenswrapper[4556]: I0218 09:04:06.433953 4556 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 18 09:04:07 crc kubenswrapper[4556]: I0218 09:04:07.249192 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-12 09:44:07.343157133 +0000 UTC Feb 18 09:04:07 crc kubenswrapper[4556]: I0218 09:04:07.319518 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"ef2678001ff826712da6716541210f018c74fd3d0218b388d2e7a6f16a1f4027"} Feb 18 09:04:07 crc kubenswrapper[4556]: I0218 09:04:07.319567 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"5c87bc671742a47de6c8e68aaf0c8645f9a8cae7b3d52264f36ae3827b3a8d80"} Feb 18 09:04:07 crc kubenswrapper[4556]: I0218 09:04:07.319578 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"8f59306c0d55e9d00714ce2ee2c1ed6f2ed33d8b21eecde490e86a93335a14f7"} Feb 18 09:04:07 crc kubenswrapper[4556]: I0218 09:04:07.319586 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"38f629262e27020512301a7b53ee836ee98e1e7cdee7a52b0ed506bbaaebd823"} Feb 18 09:04:07 crc kubenswrapper[4556]: I0218 09:04:07.319594 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"c6f37933281c927328ecfaf842d97ee8361726768f393726cb39517d2bdc9425"} Feb 18 09:04:07 crc kubenswrapper[4556]: I0218 09:04:07.319696 4556 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 09:04:07 crc kubenswrapper[4556]: I0218 09:04:07.320527 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:07 crc kubenswrapper[4556]: I0218 09:04:07.320561 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:07 crc kubenswrapper[4556]: I0218 09:04:07.320570 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:07 crc kubenswrapper[4556]: I0218 09:04:07.745307 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 09:04:07 crc kubenswrapper[4556]: I0218 09:04:07.745446 4556 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 18 09:04:07 crc kubenswrapper[4556]: I0218 09:04:07.745485 4556 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 09:04:07 crc kubenswrapper[4556]: I0218 09:04:07.746662 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:07 crc kubenswrapper[4556]: I0218 09:04:07.746708 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:07 crc kubenswrapper[4556]: I0218 09:04:07.746717 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:08 crc kubenswrapper[4556]: I0218 09:04:08.250144 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-05 13:34:31.846099577 +0000 UTC Feb 18 09:04:09 crc kubenswrapper[4556]: I0218 09:04:09.250309 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-17 15:44:23.421434052 +0000 UTC Feb 18 09:04:09 crc kubenswrapper[4556]: I0218 09:04:09.318219 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 18 09:04:09 crc kubenswrapper[4556]: I0218 09:04:09.318361 4556 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 09:04:09 crc kubenswrapper[4556]: I0218 09:04:09.319773 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:09 crc kubenswrapper[4556]: I0218 09:04:09.319826 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:09 crc kubenswrapper[4556]: I0218 09:04:09.319836 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:09 crc kubenswrapper[4556]: I0218 09:04:09.324196 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 18 09:04:09 crc kubenswrapper[4556]: I0218 09:04:09.324304 4556 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 09:04:09 crc kubenswrapper[4556]: I0218 09:04:09.324941 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:09 crc kubenswrapper[4556]: I0218 09:04:09.324967 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:09 crc kubenswrapper[4556]: I0218 09:04:09.324974 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:10 crc kubenswrapper[4556]: I0218 09:04:10.250722 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-05 03:41:35.565242047 +0000 UTC Feb 18 09:04:10 crc kubenswrapper[4556]: I0218 09:04:10.423994 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 18 09:04:10 crc kubenswrapper[4556]: I0218 09:04:10.424171 4556 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 09:04:10 crc kubenswrapper[4556]: I0218 09:04:10.424984 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:10 crc kubenswrapper[4556]: I0218 09:04:10.425015 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:10 crc kubenswrapper[4556]: I0218 09:04:10.425023 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:11 crc kubenswrapper[4556]: I0218 09:04:11.251828 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-11 04:40:09.336333018 +0000 UTC Feb 18 09:04:11 crc kubenswrapper[4556]: I0218 09:04:11.793852 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Feb 18 09:04:11 crc kubenswrapper[4556]: I0218 09:04:11.794012 4556 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 09:04:11 crc kubenswrapper[4556]: I0218 09:04:11.795120 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:11 crc kubenswrapper[4556]: I0218 09:04:11.795241 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:11 crc kubenswrapper[4556]: I0218 09:04:11.795261 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:12 crc kubenswrapper[4556]: I0218 09:04:12.253036 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-03 23:18:12.847549201 +0000 UTC Feb 18 09:04:12 crc kubenswrapper[4556]: I0218 09:04:12.717696 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 18 09:04:12 crc kubenswrapper[4556]: I0218 09:04:12.717878 4556 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 18 09:04:12 crc kubenswrapper[4556]: I0218 09:04:12.717927 4556 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 09:04:12 crc kubenswrapper[4556]: I0218 09:04:12.719007 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:12 crc kubenswrapper[4556]: I0218 09:04:12.719049 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:12 crc kubenswrapper[4556]: I0218 09:04:12.719059 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:13 crc kubenswrapper[4556]: I0218 09:04:13.077901 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 09:04:13 crc kubenswrapper[4556]: I0218 09:04:13.078046 4556 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 09:04:13 crc kubenswrapper[4556]: I0218 09:04:13.078855 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:13 crc kubenswrapper[4556]: I0218 09:04:13.078887 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:13 crc kubenswrapper[4556]: I0218 09:04:13.078897 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:13 crc kubenswrapper[4556]: I0218 09:04:13.210708 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 18 09:04:13 crc kubenswrapper[4556]: I0218 09:04:13.254131 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-08 00:40:30.26338672 +0000 UTC Feb 18 09:04:13 crc kubenswrapper[4556]: I0218 09:04:13.330092 4556 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 09:04:13 crc kubenswrapper[4556]: I0218 09:04:13.330654 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:13 crc kubenswrapper[4556]: I0218 09:04:13.330682 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:13 crc kubenswrapper[4556]: I0218 09:04:13.330691 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:13 crc kubenswrapper[4556]: E0218 09:04:13.330736 4556 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 18 09:04:14 crc kubenswrapper[4556]: I0218 09:04:14.002809 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Feb 18 09:04:14 crc kubenswrapper[4556]: I0218 09:04:14.002998 4556 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 09:04:14 crc kubenswrapper[4556]: I0218 09:04:14.004114 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:14 crc kubenswrapper[4556]: I0218 09:04:14.004146 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:14 crc kubenswrapper[4556]: I0218 09:04:14.004175 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:14 crc kubenswrapper[4556]: I0218 09:04:14.062009 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 18 09:04:14 crc kubenswrapper[4556]: I0218 09:04:14.255296 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-17 15:55:45.386475955 +0000 UTC Feb 18 09:04:14 crc kubenswrapper[4556]: I0218 09:04:14.332590 4556 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 09:04:14 crc kubenswrapper[4556]: I0218 09:04:14.333723 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:14 crc kubenswrapper[4556]: I0218 09:04:14.333765 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:14 crc kubenswrapper[4556]: I0218 09:04:14.333775 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:14 crc kubenswrapper[4556]: I0218 09:04:14.337270 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 18 09:04:15 crc kubenswrapper[4556]: E0218 09:04:15.225189 4556 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": net/http: TLS handshake timeout" logger="UnhandledError" Feb 18 09:04:15 crc kubenswrapper[4556]: I0218 09:04:15.247950 4556 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Feb 18 09:04:15 crc kubenswrapper[4556]: I0218 09:04:15.256190 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-17 10:16:42.137092408 +0000 UTC Feb 18 09:04:15 crc kubenswrapper[4556]: I0218 09:04:15.334596 4556 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 09:04:15 crc kubenswrapper[4556]: I0218 09:04:15.335477 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:15 crc kubenswrapper[4556]: I0218 09:04:15.335509 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:15 crc kubenswrapper[4556]: I0218 09:04:15.335519 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:15 crc kubenswrapper[4556]: I0218 09:04:15.686221 4556 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Feb 18 09:04:15 crc kubenswrapper[4556]: I0218 09:04:15.686289 4556 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Feb 18 09:04:15 crc kubenswrapper[4556]: I0218 09:04:15.689798 4556 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Feb 18 09:04:15 crc kubenswrapper[4556]: I0218 09:04:15.689847 4556 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Feb 18 09:04:15 crc kubenswrapper[4556]: I0218 09:04:15.718216 4556 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 18 09:04:15 crc kubenswrapper[4556]: I0218 09:04:15.718294 4556 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 18 09:04:16 crc kubenswrapper[4556]: I0218 09:04:16.256960 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 15:01:18.109491115 +0000 UTC Feb 18 09:04:17 crc kubenswrapper[4556]: I0218 09:04:17.257259 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 17:48:11.640815364 +0000 UTC Feb 18 09:04:17 crc kubenswrapper[4556]: I0218 09:04:17.750376 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 09:04:17 crc kubenswrapper[4556]: I0218 09:04:17.750507 4556 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 09:04:17 crc kubenswrapper[4556]: I0218 09:04:17.751199 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:17 crc kubenswrapper[4556]: I0218 09:04:17.751222 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:17 crc kubenswrapper[4556]: I0218 09:04:17.751230 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:17 crc kubenswrapper[4556]: I0218 09:04:17.753752 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 09:04:18 crc kubenswrapper[4556]: I0218 09:04:18.257547 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 18:08:17.406334848 +0000 UTC Feb 18 09:04:18 crc kubenswrapper[4556]: I0218 09:04:18.339328 4556 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 09:04:18 crc kubenswrapper[4556]: I0218 09:04:18.340020 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:18 crc kubenswrapper[4556]: I0218 09:04:18.340051 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:18 crc kubenswrapper[4556]: I0218 09:04:18.340059 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:19 crc kubenswrapper[4556]: I0218 09:04:19.257832 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-08 21:45:12.206615193 +0000 UTC Feb 18 09:04:19 crc kubenswrapper[4556]: I0218 09:04:19.558584 4556 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 18 09:04:19 crc kubenswrapper[4556]: I0218 09:04:19.568934 4556 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Feb 18 09:04:20 crc kubenswrapper[4556]: I0218 09:04:20.258184 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-08 14:52:11.523025622 +0000 UTC Feb 18 09:04:20 crc kubenswrapper[4556]: E0218 09:04:20.676946 4556 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="3.2s" Feb 18 09:04:20 crc kubenswrapper[4556]: I0218 09:04:20.678198 4556 trace.go:236] Trace[1645763544]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (18-Feb-2026 09:04:07.551) (total time: 13126ms): Feb 18 09:04:20 crc kubenswrapper[4556]: Trace[1645763544]: ---"Objects listed" error: 13126ms (09:04:20.678) Feb 18 09:04:20 crc kubenswrapper[4556]: Trace[1645763544]: [13.126880546s] [13.126880546s] END Feb 18 09:04:20 crc kubenswrapper[4556]: I0218 09:04:20.678210 4556 trace.go:236] Trace[1554406421]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (18-Feb-2026 09:04:07.515) (total time: 13162ms): Feb 18 09:04:20 crc kubenswrapper[4556]: Trace[1554406421]: ---"Objects listed" error: 13162ms (09:04:20.678) Feb 18 09:04:20 crc kubenswrapper[4556]: Trace[1554406421]: [13.162750171s] [13.162750171s] END Feb 18 09:04:20 crc kubenswrapper[4556]: I0218 09:04:20.678234 4556 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Feb 18 09:04:20 crc kubenswrapper[4556]: I0218 09:04:20.678220 4556 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Feb 18 09:04:20 crc kubenswrapper[4556]: I0218 09:04:20.679245 4556 trace.go:236] Trace[74097934]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (18-Feb-2026 09:04:05.937) (total time: 14741ms): Feb 18 09:04:20 crc kubenswrapper[4556]: Trace[74097934]: ---"Objects listed" error: 14741ms (09:04:20.679) Feb 18 09:04:20 crc kubenswrapper[4556]: Trace[74097934]: [14.741953693s] [14.741953693s] END Feb 18 09:04:20 crc kubenswrapper[4556]: I0218 09:04:20.679262 4556 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Feb 18 09:04:20 crc kubenswrapper[4556]: I0218 09:04:20.680428 4556 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Feb 18 09:04:20 crc kubenswrapper[4556]: I0218 09:04:20.680443 4556 trace.go:236] Trace[1533432025]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (18-Feb-2026 09:04:07.783) (total time: 12897ms): Feb 18 09:04:20 crc kubenswrapper[4556]: Trace[1533432025]: ---"Objects listed" error: 12897ms (09:04:20.680) Feb 18 09:04:20 crc kubenswrapper[4556]: Trace[1533432025]: [12.897196643s] [12.897196643s] END Feb 18 09:04:20 crc kubenswrapper[4556]: I0218 09:04:20.680458 4556 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Feb 18 09:04:20 crc kubenswrapper[4556]: E0218 09:04:20.681245 4556 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Feb 18 09:04:20 crc kubenswrapper[4556]: I0218 09:04:20.695005 4556 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:43308->192.168.126.11:17697: read: connection reset by peer" start-of-body= Feb 18 09:04:20 crc kubenswrapper[4556]: I0218 09:04:20.695053 4556 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:43308->192.168.126.11:17697: read: connection reset by peer" Feb 18 09:04:20 crc kubenswrapper[4556]: I0218 09:04:20.695318 4556 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Feb 18 09:04:20 crc kubenswrapper[4556]: I0218 09:04:20.695375 4556 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.249096 4556 apiserver.go:52] "Watching apiserver" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.251071 4556 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.251215 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"] Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.251521 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.251548 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:04:21 crc kubenswrapper[4556]: E0218 09:04:21.251573 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 09:04:21 crc kubenswrapper[4556]: E0218 09:04:21.251590 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.251624 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.251731 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.251838 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.251851 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:04:21 crc kubenswrapper[4556]: E0218 09:04:21.252088 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.253017 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.253053 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.253011 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.253316 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.253893 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.254017 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.254076 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.254238 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.254300 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.258286 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-07 21:32:13.731336236 +0000 UTC Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.273222 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.281088 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.288454 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.295400 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.303207 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.309202 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.317055 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.322995 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.344654 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.345894 4556 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="072e865e6a50c03519ce22958ac4fcbb911fc140c330b6d7ffe5c9d7dd381e96" exitCode=255 Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.345933 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"072e865e6a50c03519ce22958ac4fcbb911fc140c330b6d7ffe5c9d7dd381e96"} Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.347977 4556 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.352218 4556 scope.go:117] "RemoveContainer" containerID="072e865e6a50c03519ce22958ac4fcbb911fc140c330b6d7ffe5c9d7dd381e96" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.352543 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.353097 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.359124 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.366074 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.373397 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.380872 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.383205 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.383241 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.383260 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.383276 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.383291 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.383310 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.383327 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.383343 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.383360 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.383376 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.383390 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.383405 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.383422 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.383437 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.383452 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.383467 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.383480 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.383543 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.383559 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.383564 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.383574 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.383662 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.383686 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.383716 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.383733 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.383750 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.383765 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.383782 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.383797 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.383816 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.383834 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.383848 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.383863 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.383876 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.383883 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.383893 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.383940 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.383942 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.383964 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.383967 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.383982 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.383978 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.383985 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.383981 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.384000 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.384059 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.384080 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.384096 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.384323 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.384340 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.384357 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.384375 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.384390 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.384404 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.384419 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.384423 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.384439 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.384454 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.384448 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.384501 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.384519 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.384522 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.384535 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.384551 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.384567 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.384587 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.384601 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.384601 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.384681 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.384699 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.384725 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.384742 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.384758 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.384776 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.384791 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.384806 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.384822 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.384861 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.384890 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.384907 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.384922 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.384939 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.384954 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.384970 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.384988 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385004 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385019 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385048 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385064 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385081 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385115 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385129 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385146 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385175 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385190 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385206 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385224 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385238 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385263 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385292 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385307 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385322 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385338 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385353 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385372 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385390 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385407 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385422 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385440 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385456 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385472 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385491 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385508 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385524 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385539 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385555 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385571 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385588 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385602 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385619 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385635 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385649 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385664 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385679 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385695 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385720 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385738 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385756 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385773 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385789 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385807 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385831 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385848 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385863 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385879 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385894 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385909 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385924 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385937 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385952 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385967 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385980 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385996 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386014 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386030 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386044 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386060 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386074 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386088 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386104 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386120 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386135 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386193 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386209 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386226 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386241 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386256 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386272 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386291 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386307 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386324 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386340 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386356 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386370 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386385 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386402 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386417 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386433 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386450 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386469 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386485 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386500 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386515 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386532 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386548 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386564 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386579 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386595 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386611 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386626 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386640 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386671 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386686 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386712 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386728 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386744 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386760 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386776 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386794 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386810 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386826 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386842 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386859 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386873 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386889 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386903 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386918 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386933 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386949 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386964 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386980 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386996 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.387010 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.387026 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.387041 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.387056 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.387092 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.387117 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.387137 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.387175 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.387194 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.387215 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.387232 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.387249 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.387269 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.387286 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.387302 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.387321 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.387338 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.387353 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.387779 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.388445 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.384752 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.394081 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.384765 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.384849 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385009 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385479 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385670 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385694 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385722 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385783 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.385974 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386104 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386125 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386165 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386272 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386410 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386517 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386758 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386805 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386822 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.386925 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.387073 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.387256 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: E0218 09:04:21.387395 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 09:04:21.887372448 +0000 UTC m=+18.904333428 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.394317 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.394499 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.394691 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.394770 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.394769 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.387480 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.387508 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.387593 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.394809 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.387736 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.387853 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.388051 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.388103 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.388131 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.387795 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.387781 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.388277 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.388394 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.388400 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.388414 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.388496 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.388721 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.388731 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.388860 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.388868 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.389016 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.388737 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.394907 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.389129 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.389791 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.389836 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.389883 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.390177 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.390181 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.390204 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.390205 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.390266 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.390375 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.390490 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.390516 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.390498 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.390575 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.391213 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.391390 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.391407 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.391417 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.391580 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.391590 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.391621 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.391688 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.391857 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.391955 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.392432 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.392573 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.392687 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.392878 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.393066 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.393266 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.393382 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.393532 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.393876 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.393986 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.394000 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.394831 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.395040 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.395359 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.395414 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.395608 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.395743 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.395751 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.395768 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.395785 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.396078 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.396138 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.396203 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.396564 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.396982 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.397176 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.397496 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.397534 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.397555 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.397574 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.397656 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.397897 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.387444 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.397710 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.397939 4556 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.397978 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.398044 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.398061 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.398426 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.398507 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.398517 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.398732 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.398837 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.398999 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.399008 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.399107 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.399239 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.399288 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.399343 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.399361 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.399400 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.399789 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.399864 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.400079 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.400085 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.400161 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.401167 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 18 09:04:21 crc kubenswrapper[4556]: E0218 09:04:21.401535 4556 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 18 09:04:21 crc kubenswrapper[4556]: E0218 09:04:21.401912 4556 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.401975 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.401994 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.402008 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.402022 4556 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.402035 4556 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.402047 4556 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.402062 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.402074 4556 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.402087 4556 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.402087 4556 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Feb 18 09:04:21 crc kubenswrapper[4556]: E0218 09:04:21.402320 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-18 09:04:21.902306491 +0000 UTC m=+18.919267471 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 18 09:04:21 crc kubenswrapper[4556]: E0218 09:04:21.402334 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-18 09:04:21.902329114 +0000 UTC m=+18.919290094 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.402466 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.402841 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.403748 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.403813 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.404757 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.405340 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.405588 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.405978 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.408263 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.408407 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.408337 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.408626 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.409320 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.410077 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.410372 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.410429 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.411256 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.411530 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.411558 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.411744 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: E0218 09:04:21.411926 4556 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 18 09:04:21 crc kubenswrapper[4556]: E0218 09:04:21.411941 4556 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 18 09:04:21 crc kubenswrapper[4556]: E0218 09:04:21.411952 4556 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 09:04:21 crc kubenswrapper[4556]: E0218 09:04:21.411992 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-18 09:04:21.911979825 +0000 UTC m=+18.928940805 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 09:04:21 crc kubenswrapper[4556]: E0218 09:04:21.412055 4556 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 18 09:04:21 crc kubenswrapper[4556]: E0218 09:04:21.412070 4556 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 18 09:04:21 crc kubenswrapper[4556]: E0218 09:04:21.412081 4556 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 09:04:21 crc kubenswrapper[4556]: E0218 09:04:21.412114 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-18 09:04:21.912103458 +0000 UTC m=+18.929064437 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.412183 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.412311 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.413528 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.413614 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.414301 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.414823 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.414903 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.414941 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.415135 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.415682 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.415721 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.415834 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.416016 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.416045 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.416166 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.416210 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.416369 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.417535 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.417933 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.422305 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.422316 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.422328 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.422310 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.422444 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.422556 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.422602 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.422741 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.422760 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.422939 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.423104 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.423666 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.423785 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.423779 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.423983 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.424127 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.426401 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.427363 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.429662 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.432491 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.437763 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.502851 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503050 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503105 4556 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503115 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503124 4556 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503132 4556 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503140 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503148 4556 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503172 4556 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503180 4556 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503188 4556 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503194 4556 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503201 4556 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503209 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503215 4556 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503222 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503231 4556 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503238 4556 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503247 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503255 4556 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503262 4556 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503269 4556 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503276 4556 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503283 4556 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503291 4556 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503299 4556 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503306 4556 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503314 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503323 4556 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503331 4556 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503338 4556 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503345 4556 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503352 4556 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503359 4556 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503366 4556 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503373 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503381 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503388 4556 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503395 4556 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503401 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503408 4556 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503415 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503422 4556 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503429 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503436 4556 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503444 4556 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503451 4556 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503462 4556 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503470 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503477 4556 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503484 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503491 4556 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503498 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503505 4556 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503513 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503519 4556 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503526 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503534 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503540 4556 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503547 4556 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503554 4556 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503561 4556 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503568 4556 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503576 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503585 4556 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503592 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503599 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503606 4556 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503613 4556 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503613 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503620 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503656 4556 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503665 4556 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503674 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503682 4556 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503689 4556 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503697 4556 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.502969 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503716 4556 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503724 4556 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503731 4556 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503739 4556 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503747 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503758 4556 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503766 4556 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503775 4556 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503784 4556 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503791 4556 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503798 4556 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503805 4556 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503812 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503820 4556 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503827 4556 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503835 4556 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503842 4556 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503850 4556 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503860 4556 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503867 4556 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503875 4556 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503882 4556 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503889 4556 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503896 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503904 4556 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503911 4556 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503918 4556 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503926 4556 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503934 4556 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503941 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503948 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503955 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503962 4556 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503968 4556 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503975 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503983 4556 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503989 4556 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.503997 4556 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504004 4556 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504011 4556 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504018 4556 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504024 4556 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504031 4556 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504038 4556 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504045 4556 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504052 4556 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504059 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504067 4556 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504074 4556 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504081 4556 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504090 4556 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504097 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504104 4556 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504111 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504119 4556 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504126 4556 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504134 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504142 4556 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504162 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504170 4556 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504177 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504185 4556 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504192 4556 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504200 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504208 4556 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504215 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504223 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504231 4556 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504239 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504246 4556 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504253 4556 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504261 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504268 4556 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504275 4556 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504282 4556 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504289 4556 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504296 4556 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504304 4556 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504311 4556 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504319 4556 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504326 4556 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504335 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504343 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504351 4556 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504358 4556 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504365 4556 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504372 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504381 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504388 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504395 4556 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504402 4556 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504409 4556 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504417 4556 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504424 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504432 4556 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504439 4556 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504445 4556 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504452 4556 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504459 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504466 4556 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504474 4556 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504483 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504490 4556 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504497 4556 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504504 4556 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504511 4556 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504519 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504527 4556 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.504536 4556 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.560885 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.566787 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 18 09:04:21 crc kubenswrapper[4556]: W0218 09:04:21.569163 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-217486d7a22661dc49c15488a508855f313133a191ef2ad713e015269dfb283d WatchSource:0}: Error finding container 217486d7a22661dc49c15488a508855f313133a191ef2ad713e015269dfb283d: Status 404 returned error can't find the container with id 217486d7a22661dc49c15488a508855f313133a191ef2ad713e015269dfb283d Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.571996 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 18 09:04:21 crc kubenswrapper[4556]: W0218 09:04:21.575765 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-04c67be0af45a31b2fb507ef4921624ec52160060316c9f934cb365b9396a61b WatchSource:0}: Error finding container 04c67be0af45a31b2fb507ef4921624ec52160060316c9f934cb365b9396a61b: Status 404 returned error can't find the container with id 04c67be0af45a31b2fb507ef4921624ec52160060316c9f934cb365b9396a61b Feb 18 09:04:21 crc kubenswrapper[4556]: W0218 09:04:21.581228 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-5b3dbdbe7540f05dbe8990066edf08f642cb8f4253bf75f0b5220e0c57210169 WatchSource:0}: Error finding container 5b3dbdbe7540f05dbe8990066edf08f642cb8f4253bf75f0b5220e0c57210169: Status 404 returned error can't find the container with id 5b3dbdbe7540f05dbe8990066edf08f642cb8f4253bf75f0b5220e0c57210169 Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.907044 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.907115 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:04:21 crc kubenswrapper[4556]: I0218 09:04:21.907146 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:04:21 crc kubenswrapper[4556]: E0218 09:04:21.907257 4556 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 18 09:04:21 crc kubenswrapper[4556]: E0218 09:04:21.907301 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-18 09:04:22.90728943 +0000 UTC m=+19.924250410 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 18 09:04:21 crc kubenswrapper[4556]: E0218 09:04:21.907597 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 09:04:22.907588512 +0000 UTC m=+19.924549482 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:04:21 crc kubenswrapper[4556]: E0218 09:04:21.907640 4556 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 18 09:04:21 crc kubenswrapper[4556]: E0218 09:04:21.907665 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-18 09:04:22.907658684 +0000 UTC m=+19.924619664 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 18 09:04:22 crc kubenswrapper[4556]: I0218 09:04:22.007837 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:04:22 crc kubenswrapper[4556]: E0218 09:04:22.007986 4556 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 18 09:04:22 crc kubenswrapper[4556]: E0218 09:04:22.008007 4556 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 18 09:04:22 crc kubenswrapper[4556]: E0218 09:04:22.008019 4556 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 09:04:22 crc kubenswrapper[4556]: I0218 09:04:22.008357 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:04:22 crc kubenswrapper[4556]: E0218 09:04:22.008412 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-18 09:04:23.008399532 +0000 UTC m=+20.025360512 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 09:04:22 crc kubenswrapper[4556]: E0218 09:04:22.008500 4556 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 18 09:04:22 crc kubenswrapper[4556]: E0218 09:04:22.008517 4556 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 18 09:04:22 crc kubenswrapper[4556]: E0218 09:04:22.008526 4556 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 09:04:22 crc kubenswrapper[4556]: E0218 09:04:22.008561 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-18 09:04:23.008552979 +0000 UTC m=+20.025513960 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 09:04:22 crc kubenswrapper[4556]: I0218 09:04:22.259209 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-04 15:13:10.006299665 +0000 UTC Feb 18 09:04:22 crc kubenswrapper[4556]: I0218 09:04:22.349936 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 18 09:04:22 crc kubenswrapper[4556]: I0218 09:04:22.352008 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b42d13c1767c29d22e52537ff45561a0fbf8f2069b2c3612e903e8c127e12485"} Feb 18 09:04:22 crc kubenswrapper[4556]: I0218 09:04:22.352146 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 09:04:22 crc kubenswrapper[4556]: I0218 09:04:22.353223 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"6387f58543cf2601a53b0bbe9be3ec7816ba90d5ffdc4465594b37ccac0f6dea"} Feb 18 09:04:22 crc kubenswrapper[4556]: I0218 09:04:22.353259 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"5b3dbdbe7540f05dbe8990066edf08f642cb8f4253bf75f0b5220e0c57210169"} Feb 18 09:04:22 crc kubenswrapper[4556]: I0218 09:04:22.354372 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"04c67be0af45a31b2fb507ef4921624ec52160060316c9f934cb365b9396a61b"} Feb 18 09:04:22 crc kubenswrapper[4556]: I0218 09:04:22.355592 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"6d653c21d3d7c875bed37ed6da700dc8c801b95c80d7e361121dfbbd3d05a922"} Feb 18 09:04:22 crc kubenswrapper[4556]: I0218 09:04:22.355622 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"dd5211d0cd5172d45b92b003e2bf97d7969a65ea448b0491f8f6ea67364925ea"} Feb 18 09:04:22 crc kubenswrapper[4556]: I0218 09:04:22.355632 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"217486d7a22661dc49c15488a508855f313133a191ef2ad713e015269dfb283d"} Feb 18 09:04:22 crc kubenswrapper[4556]: I0218 09:04:22.363755 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"302eb215-a4c2-4324-836d-f37debfd620f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a99dc228d1c4cb8aeab2140ec5f37358b65684739f833fbde45822782430a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14f794063db9d33b82b79acc7134b5a2a3435e09d7cffe49d49251d1f7d20402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5eb3d3728da247903bece56bc2481547317e386e9e97b6c69926a3b55dc89b8e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b42d13c1767c29d22e52537ff45561a0fbf8f2069b2c3612e903e8c127e12485\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e865e6a50c03519ce22958ac4fcbb911fc140c330b6d7ffe5c9d7dd381e96\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T09:04:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0218 09:04:15.143085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 09:04:15.144650 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2160144045/tls.crt::/tmp/serving-cert-2160144045/tls.key\\\\\\\"\\\\nI0218 09:04:20.680799 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 09:04:20.682555 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 09:04:20.682573 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 09:04:20.682590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 09:04:20.682594 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 09:04:20.686822 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 09:04:20.686847 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686852 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686856 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 09:04:20.686859 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 09:04:20.686861 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 09:04:20.686864 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 09:04:20.688257 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 09:04:20.690726 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://974e17ae7082631df2d1bcee60c8a5f142bae2659e9d200d537a4d650f401b4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:22Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:22 crc kubenswrapper[4556]: I0218 09:04:22.372499 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:22Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:22 crc kubenswrapper[4556]: I0218 09:04:22.381020 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:22Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:22 crc kubenswrapper[4556]: I0218 09:04:22.391743 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:22Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:22 crc kubenswrapper[4556]: I0218 09:04:22.401478 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:22Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:22 crc kubenswrapper[4556]: I0218 09:04:22.410165 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:22Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:22 crc kubenswrapper[4556]: I0218 09:04:22.418859 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:22Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:22 crc kubenswrapper[4556]: I0218 09:04:22.429706 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"302eb215-a4c2-4324-836d-f37debfd620f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a99dc228d1c4cb8aeab2140ec5f37358b65684739f833fbde45822782430a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14f794063db9d33b82b79acc7134b5a2a3435e09d7cffe49d49251d1f7d20402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5eb3d3728da247903bece56bc2481547317e386e9e97b6c69926a3b55dc89b8e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b42d13c1767c29d22e52537ff45561a0fbf8f2069b2c3612e903e8c127e12485\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e865e6a50c03519ce22958ac4fcbb911fc140c330b6d7ffe5c9d7dd381e96\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T09:04:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0218 09:04:15.143085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 09:04:15.144650 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2160144045/tls.crt::/tmp/serving-cert-2160144045/tls.key\\\\\\\"\\\\nI0218 09:04:20.680799 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 09:04:20.682555 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 09:04:20.682573 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 09:04:20.682590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 09:04:20.682594 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 09:04:20.686822 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 09:04:20.686847 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686852 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686856 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 09:04:20.686859 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 09:04:20.686861 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 09:04:20.686864 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 09:04:20.688257 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 09:04:20.690726 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://974e17ae7082631df2d1bcee60c8a5f142bae2659e9d200d537a4d650f401b4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:22Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:22 crc kubenswrapper[4556]: I0218 09:04:22.439090 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:22Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:22 crc kubenswrapper[4556]: I0218 09:04:22.447233 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:22Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:22 crc kubenswrapper[4556]: I0218 09:04:22.456789 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d653c21d3d7c875bed37ed6da700dc8c801b95c80d7e361121dfbbd3d05a922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5211d0cd5172d45b92b003e2bf97d7969a65ea448b0491f8f6ea67364925ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:22Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:22 crc kubenswrapper[4556]: I0218 09:04:22.464194 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:22Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:22 crc kubenswrapper[4556]: I0218 09:04:22.473133 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6387f58543cf2601a53b0bbe9be3ec7816ba90d5ffdc4465594b37ccac0f6dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:22Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:22 crc kubenswrapper[4556]: I0218 09:04:22.480812 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:22Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:22 crc kubenswrapper[4556]: I0218 09:04:22.720921 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 18 09:04:22 crc kubenswrapper[4556]: I0218 09:04:22.723588 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 18 09:04:22 crc kubenswrapper[4556]: I0218 09:04:22.727660 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Feb 18 09:04:22 crc kubenswrapper[4556]: I0218 09:04:22.730837 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"302eb215-a4c2-4324-836d-f37debfd620f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a99dc228d1c4cb8aeab2140ec5f37358b65684739f833fbde45822782430a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14f794063db9d33b82b79acc7134b5a2a3435e09d7cffe49d49251d1f7d20402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5eb3d3728da247903bece56bc2481547317e386e9e97b6c69926a3b55dc89b8e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b42d13c1767c29d22e52537ff45561a0fbf8f2069b2c3612e903e8c127e12485\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e865e6a50c03519ce22958ac4fcbb911fc140c330b6d7ffe5c9d7dd381e96\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T09:04:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0218 09:04:15.143085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 09:04:15.144650 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2160144045/tls.crt::/tmp/serving-cert-2160144045/tls.key\\\\\\\"\\\\nI0218 09:04:20.680799 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 09:04:20.682555 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 09:04:20.682573 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 09:04:20.682590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 09:04:20.682594 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 09:04:20.686822 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 09:04:20.686847 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686852 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686856 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 09:04:20.686859 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 09:04:20.686861 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 09:04:20.686864 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 09:04:20.688257 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 09:04:20.690726 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://974e17ae7082631df2d1bcee60c8a5f142bae2659e9d200d537a4d650f401b4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:22Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:22 crc kubenswrapper[4556]: I0218 09:04:22.738819 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:22Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:22 crc kubenswrapper[4556]: I0218 09:04:22.746792 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:22Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:22 crc kubenswrapper[4556]: I0218 09:04:22.755735 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6387f58543cf2601a53b0bbe9be3ec7816ba90d5ffdc4465594b37ccac0f6dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:22Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:22 crc kubenswrapper[4556]: I0218 09:04:22.776288 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:22Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:22 crc kubenswrapper[4556]: I0218 09:04:22.790819 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d653c21d3d7c875bed37ed6da700dc8c801b95c80d7e361121dfbbd3d05a922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5211d0cd5172d45b92b003e2bf97d7969a65ea448b0491f8f6ea67364925ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:22Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:22 crc kubenswrapper[4556]: I0218 09:04:22.800396 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:22Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:22 crc kubenswrapper[4556]: I0218 09:04:22.809325 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8f6014-2544-4173-a94d-441cf42ba891\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77b0269625b7c6baab44a40373618f4ccfe9e1a028a1bc746b7d60c20ac4efaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac09c421b3101887bfb83107dbe26f2b5856d9b32a9cea92f6233fb6ae218cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce79cfe0242ecf403f989dc6b1c1b8be5ab54eee75ac413a8f9d03cef8bca0ed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://095ecf7e8763fa098b06eb9a8aac9f39451f7762bf20196e0803bbc841610e02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:22Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:22 crc kubenswrapper[4556]: I0218 09:04:22.817879 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:22Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:22 crc kubenswrapper[4556]: I0218 09:04:22.826009 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d653c21d3d7c875bed37ed6da700dc8c801b95c80d7e361121dfbbd3d05a922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5211d0cd5172d45b92b003e2bf97d7969a65ea448b0491f8f6ea67364925ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:22Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:22 crc kubenswrapper[4556]: I0218 09:04:22.833444 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:22Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:22 crc kubenswrapper[4556]: I0218 09:04:22.841979 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6387f58543cf2601a53b0bbe9be3ec7816ba90d5ffdc4465594b37ccac0f6dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:22Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:22 crc kubenswrapper[4556]: I0218 09:04:22.850109 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:22Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:22 crc kubenswrapper[4556]: I0218 09:04:22.858650 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"302eb215-a4c2-4324-836d-f37debfd620f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a99dc228d1c4cb8aeab2140ec5f37358b65684739f833fbde45822782430a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14f794063db9d33b82b79acc7134b5a2a3435e09d7cffe49d49251d1f7d20402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5eb3d3728da247903bece56bc2481547317e386e9e97b6c69926a3b55dc89b8e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b42d13c1767c29d22e52537ff45561a0fbf8f2069b2c3612e903e8c127e12485\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e865e6a50c03519ce22958ac4fcbb911fc140c330b6d7ffe5c9d7dd381e96\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T09:04:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0218 09:04:15.143085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 09:04:15.144650 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2160144045/tls.crt::/tmp/serving-cert-2160144045/tls.key\\\\\\\"\\\\nI0218 09:04:20.680799 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 09:04:20.682555 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 09:04:20.682573 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 09:04:20.682590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 09:04:20.682594 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 09:04:20.686822 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 09:04:20.686847 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686852 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686856 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 09:04:20.686859 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 09:04:20.686861 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 09:04:20.686864 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 09:04:20.688257 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 09:04:20.690726 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://974e17ae7082631df2d1bcee60c8a5f142bae2659e9d200d537a4d650f401b4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:22Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:22 crc kubenswrapper[4556]: I0218 09:04:22.867090 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:22Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:22 crc kubenswrapper[4556]: I0218 09:04:22.914359 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 09:04:22 crc kubenswrapper[4556]: I0218 09:04:22.914440 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:04:22 crc kubenswrapper[4556]: I0218 09:04:22.914466 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:04:22 crc kubenswrapper[4556]: E0218 09:04:22.914528 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 09:04:24.914507328 +0000 UTC m=+21.931468308 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:04:22 crc kubenswrapper[4556]: E0218 09:04:22.914551 4556 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 18 09:04:22 crc kubenswrapper[4556]: E0218 09:04:22.914585 4556 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 18 09:04:22 crc kubenswrapper[4556]: E0218 09:04:22.914601 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-18 09:04:24.914589603 +0000 UTC m=+21.931550583 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 18 09:04:22 crc kubenswrapper[4556]: E0218 09:04:22.914629 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-18 09:04:24.914618848 +0000 UTC m=+21.931579828 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.015315 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.015353 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:04:23 crc kubenswrapper[4556]: E0218 09:04:23.015440 4556 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 18 09:04:23 crc kubenswrapper[4556]: E0218 09:04:23.015452 4556 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 18 09:04:23 crc kubenswrapper[4556]: E0218 09:04:23.015462 4556 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 09:04:23 crc kubenswrapper[4556]: E0218 09:04:23.015502 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-18 09:04:25.015492595 +0000 UTC m=+22.032453575 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 09:04:23 crc kubenswrapper[4556]: E0218 09:04:23.015547 4556 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 18 09:04:23 crc kubenswrapper[4556]: E0218 09:04:23.015578 4556 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 18 09:04:23 crc kubenswrapper[4556]: E0218 09:04:23.015591 4556 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 09:04:23 crc kubenswrapper[4556]: E0218 09:04:23.015647 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-18 09:04:25.015632308 +0000 UTC m=+22.032593287 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.259509 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-14 10:58:20.131845974 +0000 UTC Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.281928 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.281970 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.282002 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:04:23 crc kubenswrapper[4556]: E0218 09:04:23.282071 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 09:04:23 crc kubenswrapper[4556]: E0218 09:04:23.282144 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 09:04:23 crc kubenswrapper[4556]: E0218 09:04:23.282223 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.284641 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.285096 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.285789 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.286319 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.286800 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.287250 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.287749 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.288208 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.288734 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.290497 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.290926 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.291609 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:23Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.291815 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.292284 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.292725 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.293476 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.293923 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.294718 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.295035 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.295520 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.296343 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.296730 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.297202 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.297878 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.298436 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.299120 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.299637 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.299971 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d653c21d3d7c875bed37ed6da700dc8c801b95c80d7e361121dfbbd3d05a922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5211d0cd5172d45b92b003e2bf97d7969a65ea448b0491f8f6ea67364925ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:23Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.300509 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.300936 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.301752 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.302183 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.302566 4556 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.302970 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.304346 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.304799 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.305492 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.306709 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.307390 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.307854 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:23Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.308139 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.308659 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.309545 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.309938 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.310751 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.311276 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.312066 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.312547 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.313500 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.313923 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.314988 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.315406 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.316108 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.316509 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.316961 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.317787 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.317870 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6387f58543cf2601a53b0bbe9be3ec7816ba90d5ffdc4465594b37ccac0f6dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:23Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.318317 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.326711 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8f6014-2544-4173-a94d-441cf42ba891\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77b0269625b7c6baab44a40373618f4ccfe9e1a028a1bc746b7d60c20ac4efaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac09c421b3101887bfb83107dbe26f2b5856d9b32a9cea92f6233fb6ae218cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce79cfe0242ecf403f989dc6b1c1b8be5ab54eee75ac413a8f9d03cef8bca0ed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://095ecf7e8763fa098b06eb9a8aac9f39451f7762bf20196e0803bbc841610e02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:23Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.338911 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:23Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.347886 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"302eb215-a4c2-4324-836d-f37debfd620f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a99dc228d1c4cb8aeab2140ec5f37358b65684739f833fbde45822782430a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14f794063db9d33b82b79acc7134b5a2a3435e09d7cffe49d49251d1f7d20402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5eb3d3728da247903bece56bc2481547317e386e9e97b6c69926a3b55dc89b8e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b42d13c1767c29d22e52537ff45561a0fbf8f2069b2c3612e903e8c127e12485\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e865e6a50c03519ce22958ac4fcbb911fc140c330b6d7ffe5c9d7dd381e96\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T09:04:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0218 09:04:15.143085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 09:04:15.144650 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2160144045/tls.crt::/tmp/serving-cert-2160144045/tls.key\\\\\\\"\\\\nI0218 09:04:20.680799 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 09:04:20.682555 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 09:04:20.682573 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 09:04:20.682590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 09:04:20.682594 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 09:04:20.686822 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 09:04:20.686847 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686852 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686856 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 09:04:20.686859 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 09:04:20.686861 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 09:04:20.686864 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 09:04:20.688257 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 09:04:20.690726 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://974e17ae7082631df2d1bcee60c8a5f142bae2659e9d200d537a4d650f401b4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:23Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.358554 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:23Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.358585 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"47341f31f94eaf5dd86efbe78ff62873e733b1f86ea895f281b6797770519459"} Feb 18 09:04:23 crc kubenswrapper[4556]: E0218 09:04:23.362377 4556 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.368833 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"302eb215-a4c2-4324-836d-f37debfd620f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a99dc228d1c4cb8aeab2140ec5f37358b65684739f833fbde45822782430a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14f794063db9d33b82b79acc7134b5a2a3435e09d7cffe49d49251d1f7d20402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5eb3d3728da247903bece56bc2481547317e386e9e97b6c69926a3b55dc89b8e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b42d13c1767c29d22e52537ff45561a0fbf8f2069b2c3612e903e8c127e12485\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e865e6a50c03519ce22958ac4fcbb911fc140c330b6d7ffe5c9d7dd381e96\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T09:04:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0218 09:04:15.143085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 09:04:15.144650 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2160144045/tls.crt::/tmp/serving-cert-2160144045/tls.key\\\\\\\"\\\\nI0218 09:04:20.680799 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 09:04:20.682555 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 09:04:20.682573 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 09:04:20.682590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 09:04:20.682594 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 09:04:20.686822 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 09:04:20.686847 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686852 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686856 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 09:04:20.686859 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 09:04:20.686861 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 09:04:20.686864 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 09:04:20.688257 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 09:04:20.690726 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://974e17ae7082631df2d1bcee60c8a5f142bae2659e9d200d537a4d650f401b4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:23Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.380054 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:23Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.388091 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:23Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.396186 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d653c21d3d7c875bed37ed6da700dc8c801b95c80d7e361121dfbbd3d05a922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5211d0cd5172d45b92b003e2bf97d7969a65ea448b0491f8f6ea67364925ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:23Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.404343 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47341f31f94eaf5dd86efbe78ff62873e733b1f86ea895f281b6797770519459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:23Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.413282 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6387f58543cf2601a53b0bbe9be3ec7816ba90d5ffdc4465594b37ccac0f6dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:23Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.421622 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8f6014-2544-4173-a94d-441cf42ba891\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77b0269625b7c6baab44a40373618f4ccfe9e1a028a1bc746b7d60c20ac4efaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac09c421b3101887bfb83107dbe26f2b5856d9b32a9cea92f6233fb6ae218cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce79cfe0242ecf403f989dc6b1c1b8be5ab54eee75ac413a8f9d03cef8bca0ed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://095ecf7e8763fa098b06eb9a8aac9f39451f7762bf20196e0803bbc841610e02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:23Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.430224 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:23Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.881497 4556 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.882799 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.882833 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.882843 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.882876 4556 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.887496 4556 kubelet_node_status.go:115] "Node was previously registered" node="crc" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.887645 4556 kubelet_node_status.go:79] "Successfully registered node" node="crc" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.888395 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.888421 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.888429 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.888438 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.888446 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:23Z","lastTransitionTime":"2026-02-18T09:04:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:23 crc kubenswrapper[4556]: E0218 09:04:23.900586 4556 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8758c02e-ef59-46a7-908c-d5c97feb8ceb\\\",\\\"systemUUID\\\":\\\"0f904183-79bf-4790-8551-43a6eb3adbe4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:23Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.902977 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.903007 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.903016 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.903027 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.903037 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:23Z","lastTransitionTime":"2026-02-18T09:04:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:23 crc kubenswrapper[4556]: E0218 09:04:23.910781 4556 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8758c02e-ef59-46a7-908c-d5c97feb8ceb\\\",\\\"systemUUID\\\":\\\"0f904183-79bf-4790-8551-43a6eb3adbe4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:23Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.912805 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.912829 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.912837 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.912847 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.912854 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:23Z","lastTransitionTime":"2026-02-18T09:04:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:23 crc kubenswrapper[4556]: E0218 09:04:23.920308 4556 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8758c02e-ef59-46a7-908c-d5c97feb8ceb\\\",\\\"systemUUID\\\":\\\"0f904183-79bf-4790-8551-43a6eb3adbe4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:23Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.922474 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.922498 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.922507 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.922517 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.922525 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:23Z","lastTransitionTime":"2026-02-18T09:04:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:23 crc kubenswrapper[4556]: E0218 09:04:23.932443 4556 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8758c02e-ef59-46a7-908c-d5c97feb8ceb\\\",\\\"systemUUID\\\":\\\"0f904183-79bf-4790-8551-43a6eb3adbe4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:23Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.934920 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.934946 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.934956 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.934965 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.934972 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:23Z","lastTransitionTime":"2026-02-18T09:04:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:23 crc kubenswrapper[4556]: E0218 09:04:23.943222 4556 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8758c02e-ef59-46a7-908c-d5c97feb8ceb\\\",\\\"systemUUID\\\":\\\"0f904183-79bf-4790-8551-43a6eb3adbe4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:23Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:23 crc kubenswrapper[4556]: E0218 09:04:23.943322 4556 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.944414 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.944456 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.944465 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.944477 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:23 crc kubenswrapper[4556]: I0218 09:04:23.944484 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:23Z","lastTransitionTime":"2026-02-18T09:04:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.021022 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.029795 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.031459 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.031847 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"302eb215-a4c2-4324-836d-f37debfd620f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a99dc228d1c4cb8aeab2140ec5f37358b65684739f833fbde45822782430a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14f794063db9d33b82b79acc7134b5a2a3435e09d7cffe49d49251d1f7d20402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5eb3d3728da247903bece56bc2481547317e386e9e97b6c69926a3b55dc89b8e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b42d13c1767c29d22e52537ff45561a0fbf8f2069b2c3612e903e8c127e12485\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e865e6a50c03519ce22958ac4fcbb911fc140c330b6d7ffe5c9d7dd381e96\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T09:04:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0218 09:04:15.143085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 09:04:15.144650 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2160144045/tls.crt::/tmp/serving-cert-2160144045/tls.key\\\\\\\"\\\\nI0218 09:04:20.680799 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 09:04:20.682555 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 09:04:20.682573 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 09:04:20.682590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 09:04:20.682594 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 09:04:20.686822 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 09:04:20.686847 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686852 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686856 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 09:04:20.686859 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 09:04:20.686861 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 09:04:20.686864 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 09:04:20.688257 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 09:04:20.690726 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://974e17ae7082631df2d1bcee60c8a5f142bae2659e9d200d537a4d650f401b4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:24Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.040050 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:24Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.046005 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.046036 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.046047 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.046057 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.046066 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:24Z","lastTransitionTime":"2026-02-18T09:04:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.048663 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47341f31f94eaf5dd86efbe78ff62873e733b1f86ea895f281b6797770519459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:24Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.059912 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6387f58543cf2601a53b0bbe9be3ec7816ba90d5ffdc4465594b37ccac0f6dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:24Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.068328 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8f6014-2544-4173-a94d-441cf42ba891\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77b0269625b7c6baab44a40373618f4ccfe9e1a028a1bc746b7d60c20ac4efaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac09c421b3101887bfb83107dbe26f2b5856d9b32a9cea92f6233fb6ae218cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce79cfe0242ecf403f989dc6b1c1b8be5ab54eee75ac413a8f9d03cef8bca0ed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://095ecf7e8763fa098b06eb9a8aac9f39451f7762bf20196e0803bbc841610e02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:24Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.077290 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:24Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.084812 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d653c21d3d7c875bed37ed6da700dc8c801b95c80d7e361121dfbbd3d05a922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5211d0cd5172d45b92b003e2bf97d7969a65ea448b0491f8f6ea67364925ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:24Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.092049 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:24Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.099364 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47341f31f94eaf5dd86efbe78ff62873e733b1f86ea895f281b6797770519459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:24Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.107463 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6387f58543cf2601a53b0bbe9be3ec7816ba90d5ffdc4465594b37ccac0f6dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:24Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.123897 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ae60f3-adab-4740-821a-1570059c7101\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f629262e27020512301a7b53ee836ee98e1e7cdee7a52b0ed506bbaaebd823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f59306c0d55e9d00714ce2ee2c1ed6f2ed33d8b21eecde490e86a93335a14f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c87bc671742a47de6c8e68aaf0c8645f9a8cae7b3d52264f36ae3827b3a8d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2678001ff826712da6716541210f018c74fd3d0218b388d2e7a6f16a1f4027\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6f37933281c927328ecfaf842d97ee8361726768f393726cb39517d2bdc9425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:24Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.138653 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8f6014-2544-4173-a94d-441cf42ba891\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77b0269625b7c6baab44a40373618f4ccfe9e1a028a1bc746b7d60c20ac4efaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac09c421b3101887bfb83107dbe26f2b5856d9b32a9cea92f6233fb6ae218cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce79cfe0242ecf403f989dc6b1c1b8be5ab54eee75ac413a8f9d03cef8bca0ed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://095ecf7e8763fa098b06eb9a8aac9f39451f7762bf20196e0803bbc841610e02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:24Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.147660 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.147704 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.147714 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.147727 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.147735 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:24Z","lastTransitionTime":"2026-02-18T09:04:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.155214 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:24Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.165674 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d653c21d3d7c875bed37ed6da700dc8c801b95c80d7e361121dfbbd3d05a922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5211d0cd5172d45b92b003e2bf97d7969a65ea448b0491f8f6ea67364925ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:24Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.174248 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:24Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.183355 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"302eb215-a4c2-4324-836d-f37debfd620f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a99dc228d1c4cb8aeab2140ec5f37358b65684739f833fbde45822782430a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14f794063db9d33b82b79acc7134b5a2a3435e09d7cffe49d49251d1f7d20402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5eb3d3728da247903bece56bc2481547317e386e9e97b6c69926a3b55dc89b8e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b42d13c1767c29d22e52537ff45561a0fbf8f2069b2c3612e903e8c127e12485\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e865e6a50c03519ce22958ac4fcbb911fc140c330b6d7ffe5c9d7dd381e96\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T09:04:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0218 09:04:15.143085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 09:04:15.144650 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2160144045/tls.crt::/tmp/serving-cert-2160144045/tls.key\\\\\\\"\\\\nI0218 09:04:20.680799 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 09:04:20.682555 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 09:04:20.682573 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 09:04:20.682590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 09:04:20.682594 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 09:04:20.686822 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 09:04:20.686847 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686852 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686856 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 09:04:20.686859 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 09:04:20.686861 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 09:04:20.686864 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 09:04:20.688257 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 09:04:20.690726 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://974e17ae7082631df2d1bcee60c8a5f142bae2659e9d200d537a4d650f401b4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:24Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.192641 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:24Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.250396 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.250431 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.250440 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.250454 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.250464 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:24Z","lastTransitionTime":"2026-02-18T09:04:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.259926 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-06 16:54:18.290290129 +0000 UTC Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.352241 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.352272 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.352281 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.352296 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.352307 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:24Z","lastTransitionTime":"2026-02-18T09:04:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.453964 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.453998 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.454009 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.454021 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.454031 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:24Z","lastTransitionTime":"2026-02-18T09:04:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.555599 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.555631 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.555642 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.555653 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.555660 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:24Z","lastTransitionTime":"2026-02-18T09:04:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.657734 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.657775 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.657784 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.657797 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.657807 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:24Z","lastTransitionTime":"2026-02-18T09:04:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.760089 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.760126 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.760135 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.760147 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.760176 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:24Z","lastTransitionTime":"2026-02-18T09:04:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.861584 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.861616 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.861625 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.861637 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.861645 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:24Z","lastTransitionTime":"2026-02-18T09:04:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.929474 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.929530 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.929554 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:04:24 crc kubenswrapper[4556]: E0218 09:04:24.929625 4556 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 18 09:04:24 crc kubenswrapper[4556]: E0218 09:04:24.929648 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 09:04:28.929623153 +0000 UTC m=+25.946584133 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:04:24 crc kubenswrapper[4556]: E0218 09:04:24.929687 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-18 09:04:28.929668459 +0000 UTC m=+25.946629438 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 18 09:04:24 crc kubenswrapper[4556]: E0218 09:04:24.929723 4556 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 18 09:04:24 crc kubenswrapper[4556]: E0218 09:04:24.929793 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-18 09:04:28.929779838 +0000 UTC m=+25.946740818 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.962815 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.962841 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.962850 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.962860 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:24 crc kubenswrapper[4556]: I0218 09:04:24.962871 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:24Z","lastTransitionTime":"2026-02-18T09:04:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:25 crc kubenswrapper[4556]: I0218 09:04:25.030390 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:04:25 crc kubenswrapper[4556]: I0218 09:04:25.030430 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:04:25 crc kubenswrapper[4556]: E0218 09:04:25.030514 4556 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 18 09:04:25 crc kubenswrapper[4556]: E0218 09:04:25.030535 4556 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 18 09:04:25 crc kubenswrapper[4556]: E0218 09:04:25.030546 4556 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 09:04:25 crc kubenswrapper[4556]: E0218 09:04:25.030591 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-18 09:04:29.030567693 +0000 UTC m=+26.047528673 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 09:04:25 crc kubenswrapper[4556]: E0218 09:04:25.030514 4556 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 18 09:04:25 crc kubenswrapper[4556]: E0218 09:04:25.030627 4556 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 18 09:04:25 crc kubenswrapper[4556]: E0218 09:04:25.030637 4556 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 09:04:25 crc kubenswrapper[4556]: E0218 09:04:25.030663 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-18 09:04:29.030655268 +0000 UTC m=+26.047616248 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 09:04:25 crc kubenswrapper[4556]: I0218 09:04:25.064512 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:25 crc kubenswrapper[4556]: I0218 09:04:25.064552 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:25 crc kubenswrapper[4556]: I0218 09:04:25.064561 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:25 crc kubenswrapper[4556]: I0218 09:04:25.064575 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:25 crc kubenswrapper[4556]: I0218 09:04:25.064584 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:25Z","lastTransitionTime":"2026-02-18T09:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:25 crc kubenswrapper[4556]: I0218 09:04:25.166244 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:25 crc kubenswrapper[4556]: I0218 09:04:25.166276 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:25 crc kubenswrapper[4556]: I0218 09:04:25.166285 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:25 crc kubenswrapper[4556]: I0218 09:04:25.166297 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:25 crc kubenswrapper[4556]: I0218 09:04:25.166306 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:25Z","lastTransitionTime":"2026-02-18T09:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:25 crc kubenswrapper[4556]: I0218 09:04:25.260682 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-31 01:28:24.763444889 +0000 UTC Feb 18 09:04:25 crc kubenswrapper[4556]: I0218 09:04:25.268128 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:25 crc kubenswrapper[4556]: I0218 09:04:25.268184 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:25 crc kubenswrapper[4556]: I0218 09:04:25.268197 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:25 crc kubenswrapper[4556]: I0218 09:04:25.268214 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:25 crc kubenswrapper[4556]: I0218 09:04:25.268223 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:25Z","lastTransitionTime":"2026-02-18T09:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:25 crc kubenswrapper[4556]: I0218 09:04:25.281590 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:04:25 crc kubenswrapper[4556]: I0218 09:04:25.281608 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:04:25 crc kubenswrapper[4556]: E0218 09:04:25.281727 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 09:04:25 crc kubenswrapper[4556]: I0218 09:04:25.281768 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:04:25 crc kubenswrapper[4556]: E0218 09:04:25.281849 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 09:04:25 crc kubenswrapper[4556]: E0218 09:04:25.281927 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 09:04:25 crc kubenswrapper[4556]: I0218 09:04:25.369799 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:25 crc kubenswrapper[4556]: I0218 09:04:25.369832 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:25 crc kubenswrapper[4556]: I0218 09:04:25.369842 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:25 crc kubenswrapper[4556]: I0218 09:04:25.369855 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:25 crc kubenswrapper[4556]: I0218 09:04:25.369862 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:25Z","lastTransitionTime":"2026-02-18T09:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:25 crc kubenswrapper[4556]: I0218 09:04:25.471511 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:25 crc kubenswrapper[4556]: I0218 09:04:25.471551 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:25 crc kubenswrapper[4556]: I0218 09:04:25.471561 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:25 crc kubenswrapper[4556]: I0218 09:04:25.471573 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:25 crc kubenswrapper[4556]: I0218 09:04:25.471586 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:25Z","lastTransitionTime":"2026-02-18T09:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:25 crc kubenswrapper[4556]: I0218 09:04:25.573444 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:25 crc kubenswrapper[4556]: I0218 09:04:25.573479 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:25 crc kubenswrapper[4556]: I0218 09:04:25.573490 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:25 crc kubenswrapper[4556]: I0218 09:04:25.573502 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:25 crc kubenswrapper[4556]: I0218 09:04:25.573510 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:25Z","lastTransitionTime":"2026-02-18T09:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:25 crc kubenswrapper[4556]: I0218 09:04:25.675064 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:25 crc kubenswrapper[4556]: I0218 09:04:25.675106 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:25 crc kubenswrapper[4556]: I0218 09:04:25.675117 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:25 crc kubenswrapper[4556]: I0218 09:04:25.675131 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:25 crc kubenswrapper[4556]: I0218 09:04:25.675141 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:25Z","lastTransitionTime":"2026-02-18T09:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:25 crc kubenswrapper[4556]: I0218 09:04:25.777323 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:25 crc kubenswrapper[4556]: I0218 09:04:25.777354 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:25 crc kubenswrapper[4556]: I0218 09:04:25.777363 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:25 crc kubenswrapper[4556]: I0218 09:04:25.777375 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:25 crc kubenswrapper[4556]: I0218 09:04:25.777385 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:25Z","lastTransitionTime":"2026-02-18T09:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:25 crc kubenswrapper[4556]: I0218 09:04:25.879910 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:25 crc kubenswrapper[4556]: I0218 09:04:25.879940 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:25 crc kubenswrapper[4556]: I0218 09:04:25.879949 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:25 crc kubenswrapper[4556]: I0218 09:04:25.879960 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:25 crc kubenswrapper[4556]: I0218 09:04:25.879967 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:25Z","lastTransitionTime":"2026-02-18T09:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:25 crc kubenswrapper[4556]: I0218 09:04:25.983533 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:25 crc kubenswrapper[4556]: I0218 09:04:25.983586 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:25 crc kubenswrapper[4556]: I0218 09:04:25.983599 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:25 crc kubenswrapper[4556]: I0218 09:04:25.983620 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:25 crc kubenswrapper[4556]: I0218 09:04:25.983631 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:25Z","lastTransitionTime":"2026-02-18T09:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.085254 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.085292 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.085300 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.085311 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.085319 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:26Z","lastTransitionTime":"2026-02-18T09:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.187338 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.187368 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.187376 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.187388 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.187396 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:26Z","lastTransitionTime":"2026-02-18T09:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.261693 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-12 02:49:52.817885265 +0000 UTC Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.267276 4556 csr.go:261] certificate signing request csr-8ppp4 is approved, waiting to be issued Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.275522 4556 csr.go:257] certificate signing request csr-8ppp4 is issued Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.289469 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.289503 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.289513 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.289527 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.289538 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:26Z","lastTransitionTime":"2026-02-18T09:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.301263 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-2jflp"] Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.301526 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-2jflp" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.303201 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.303638 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.304661 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.305901 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.314292 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"302eb215-a4c2-4324-836d-f37debfd620f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a99dc228d1c4cb8aeab2140ec5f37358b65684739f833fbde45822782430a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14f794063db9d33b82b79acc7134b5a2a3435e09d7cffe49d49251d1f7d20402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5eb3d3728da247903bece56bc2481547317e386e9e97b6c69926a3b55dc89b8e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b42d13c1767c29d22e52537ff45561a0fbf8f2069b2c3612e903e8c127e12485\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e865e6a50c03519ce22958ac4fcbb911fc140c330b6d7ffe5c9d7dd381e96\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T09:04:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0218 09:04:15.143085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 09:04:15.144650 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2160144045/tls.crt::/tmp/serving-cert-2160144045/tls.key\\\\\\\"\\\\nI0218 09:04:20.680799 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 09:04:20.682555 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 09:04:20.682573 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 09:04:20.682590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 09:04:20.682594 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 09:04:20.686822 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 09:04:20.686847 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686852 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686856 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 09:04:20.686859 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 09:04:20.686861 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 09:04:20.686864 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 09:04:20.688257 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 09:04:20.690726 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://974e17ae7082631df2d1bcee60c8a5f142bae2659e9d200d537a4d650f401b4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:26Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.323484 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:26Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.334335 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jflp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689d3d82-4071-4fe5-b3e3-b5b4f511580a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bm85s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jflp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:26Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.343190 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:26Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.351129 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d653c21d3d7c875bed37ed6da700dc8c801b95c80d7e361121dfbbd3d05a922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5211d0cd5172d45b92b003e2bf97d7969a65ea448b0491f8f6ea67364925ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:26Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.359531 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47341f31f94eaf5dd86efbe78ff62873e733b1f86ea895f281b6797770519459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:26Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.367513 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6387f58543cf2601a53b0bbe9be3ec7816ba90d5ffdc4465594b37ccac0f6dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:26Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.380100 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ae60f3-adab-4740-821a-1570059c7101\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f629262e27020512301a7b53ee836ee98e1e7cdee7a52b0ed506bbaaebd823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f59306c0d55e9d00714ce2ee2c1ed6f2ed33d8b21eecde490e86a93335a14f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c87bc671742a47de6c8e68aaf0c8645f9a8cae7b3d52264f36ae3827b3a8d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2678001ff826712da6716541210f018c74fd3d0218b388d2e7a6f16a1f4027\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6f37933281c927328ecfaf842d97ee8361726768f393726cb39517d2bdc9425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:26Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.388360 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8f6014-2544-4173-a94d-441cf42ba891\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77b0269625b7c6baab44a40373618f4ccfe9e1a028a1bc746b7d60c20ac4efaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac09c421b3101887bfb83107dbe26f2b5856d9b32a9cea92f6233fb6ae218cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce79cfe0242ecf403f989dc6b1c1b8be5ab54eee75ac413a8f9d03cef8bca0ed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://095ecf7e8763fa098b06eb9a8aac9f39451f7762bf20196e0803bbc841610e02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:26Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.390808 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.390843 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.390854 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.390869 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.390878 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:26Z","lastTransitionTime":"2026-02-18T09:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.396488 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:26Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.441753 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/689d3d82-4071-4fe5-b3e3-b5b4f511580a-host\") pod \"node-ca-2jflp\" (UID: \"689d3d82-4071-4fe5-b3e3-b5b4f511580a\") " pod="openshift-image-registry/node-ca-2jflp" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.441783 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bm85s\" (UniqueName: \"kubernetes.io/projected/689d3d82-4071-4fe5-b3e3-b5b4f511580a-kube-api-access-bm85s\") pod \"node-ca-2jflp\" (UID: \"689d3d82-4071-4fe5-b3e3-b5b4f511580a\") " pod="openshift-image-registry/node-ca-2jflp" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.441814 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/689d3d82-4071-4fe5-b3e3-b5b4f511580a-serviceca\") pod \"node-ca-2jflp\" (UID: \"689d3d82-4071-4fe5-b3e3-b5b4f511580a\") " pod="openshift-image-registry/node-ca-2jflp" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.493318 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.493341 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.493349 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.493361 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.493368 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:26Z","lastTransitionTime":"2026-02-18T09:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.542325 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/689d3d82-4071-4fe5-b3e3-b5b4f511580a-serviceca\") pod \"node-ca-2jflp\" (UID: \"689d3d82-4071-4fe5-b3e3-b5b4f511580a\") " pod="openshift-image-registry/node-ca-2jflp" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.542393 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/689d3d82-4071-4fe5-b3e3-b5b4f511580a-host\") pod \"node-ca-2jflp\" (UID: \"689d3d82-4071-4fe5-b3e3-b5b4f511580a\") " pod="openshift-image-registry/node-ca-2jflp" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.542411 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bm85s\" (UniqueName: \"kubernetes.io/projected/689d3d82-4071-4fe5-b3e3-b5b4f511580a-kube-api-access-bm85s\") pod \"node-ca-2jflp\" (UID: \"689d3d82-4071-4fe5-b3e3-b5b4f511580a\") " pod="openshift-image-registry/node-ca-2jflp" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.543272 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/689d3d82-4071-4fe5-b3e3-b5b4f511580a-host\") pod \"node-ca-2jflp\" (UID: \"689d3d82-4071-4fe5-b3e3-b5b4f511580a\") " pod="openshift-image-registry/node-ca-2jflp" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.544272 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/689d3d82-4071-4fe5-b3e3-b5b4f511580a-serviceca\") pod \"node-ca-2jflp\" (UID: \"689d3d82-4071-4fe5-b3e3-b5b4f511580a\") " pod="openshift-image-registry/node-ca-2jflp" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.560194 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bm85s\" (UniqueName: \"kubernetes.io/projected/689d3d82-4071-4fe5-b3e3-b5b4f511580a-kube-api-access-bm85s\") pod \"node-ca-2jflp\" (UID: \"689d3d82-4071-4fe5-b3e3-b5b4f511580a\") " pod="openshift-image-registry/node-ca-2jflp" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.595258 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.595296 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.595305 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.595317 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.595326 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:26Z","lastTransitionTime":"2026-02-18T09:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.610490 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-2jflp" Feb 18 09:04:26 crc kubenswrapper[4556]: W0218 09:04:26.622651 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod689d3d82_4071_4fe5_b3e3_b5b4f511580a.slice/crio-239ce6934d18ccf363d93e23154ce1d625ec96ed464059eb45049f8d96d72c09 WatchSource:0}: Error finding container 239ce6934d18ccf363d93e23154ce1d625ec96ed464059eb45049f8d96d72c09: Status 404 returned error can't find the container with id 239ce6934d18ccf363d93e23154ce1d625ec96ed464059eb45049f8d96d72c09 Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.697880 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.697912 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.697921 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.697933 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.697940 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:26Z","lastTransitionTime":"2026-02-18T09:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.753092 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-5vcgz"] Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.753394 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-5vcgz" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.754261 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-qfksk"] Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.754511 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-qfksk" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.754935 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.755105 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.755105 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.755305 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-f76hs"] Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.755504 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-sgcq7"] Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.755640 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.755824 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.755975 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-sgcq7" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.755994 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.756081 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.756310 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.756376 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Feb 18 09:04:26 crc kubenswrapper[4556]: W0218 09:04:26.756779 4556 reflector.go:561] object-"openshift-machine-config-operator"/"kube-rbac-proxy": failed to list *v1.ConfigMap: configmaps "kube-rbac-proxy" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-machine-config-operator": no relationship found between node 'crc' and this object Feb 18 09:04:26 crc kubenswrapper[4556]: E0218 09:04:26.756806 4556 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-config-operator\"/\"kube-rbac-proxy\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-rbac-proxy\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-machine-config-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.757908 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.757914 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.757995 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.758266 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.758811 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.758846 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.763768 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47341f31f94eaf5dd86efbe78ff62873e733b1f86ea895f281b6797770519459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:26Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.773749 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6387f58543cf2601a53b0bbe9be3ec7816ba90d5ffdc4465594b37ccac0f6dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:26Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.787736 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ae60f3-adab-4740-821a-1570059c7101\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f629262e27020512301a7b53ee836ee98e1e7cdee7a52b0ed506bbaaebd823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f59306c0d55e9d00714ce2ee2c1ed6f2ed33d8b21eecde490e86a93335a14f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c87bc671742a47de6c8e68aaf0c8645f9a8cae7b3d52264f36ae3827b3a8d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2678001ff826712da6716541210f018c74fd3d0218b388d2e7a6f16a1f4027\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6f37933281c927328ecfaf842d97ee8361726768f393726cb39517d2bdc9425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:26Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.799099 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8f6014-2544-4173-a94d-441cf42ba891\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77b0269625b7c6baab44a40373618f4ccfe9e1a028a1bc746b7d60c20ac4efaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac09c421b3101887bfb83107dbe26f2b5856d9b32a9cea92f6233fb6ae218cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce79cfe0242ecf403f989dc6b1c1b8be5ab54eee75ac413a8f9d03cef8bca0ed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://095ecf7e8763fa098b06eb9a8aac9f39451f7762bf20196e0803bbc841610e02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:26Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.799511 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.799543 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.799554 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.799566 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.799574 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:26Z","lastTransitionTime":"2026-02-18T09:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.808842 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:26Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.817265 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d653c21d3d7c875bed37ed6da700dc8c801b95c80d7e361121dfbbd3d05a922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5211d0cd5172d45b92b003e2bf97d7969a65ea448b0491f8f6ea67364925ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:26Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.827790 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:26Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.834110 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5vcgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d50c39aa-f956-441a-8abe-1d7247a7fd86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4llw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5vcgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:26Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.843528 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/e226b189-cfa9-47fc-b94d-19b5cbe0859f-multus-socket-dir-parent\") pod \"multus-qfksk\" (UID: \"e226b189-cfa9-47fc-b94d-19b5cbe0859f\") " pod="openshift-multus/multus-qfksk" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.843562 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/e226b189-cfa9-47fc-b94d-19b5cbe0859f-host-run-k8s-cni-cncf-io\") pod \"multus-qfksk\" (UID: \"e226b189-cfa9-47fc-b94d-19b5cbe0859f\") " pod="openshift-multus/multus-qfksk" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.843583 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e226b189-cfa9-47fc-b94d-19b5cbe0859f-system-cni-dir\") pod \"multus-qfksk\" (UID: \"e226b189-cfa9-47fc-b94d-19b5cbe0859f\") " pod="openshift-multus/multus-qfksk" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.843601 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e226b189-cfa9-47fc-b94d-19b5cbe0859f-host-var-lib-cni-bin\") pod \"multus-qfksk\" (UID: \"e226b189-cfa9-47fc-b94d-19b5cbe0859f\") " pod="openshift-multus/multus-qfksk" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.843617 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtvs9\" (UniqueName: \"kubernetes.io/projected/8dac7f27-d3d1-4778-9e54-f273035a1d37-kube-api-access-dtvs9\") pod \"machine-config-daemon-f76hs\" (UID: \"8dac7f27-d3d1-4778-9e54-f273035a1d37\") " pod="openshift-machine-config-operator/machine-config-daemon-f76hs" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.843632 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e226b189-cfa9-47fc-b94d-19b5cbe0859f-multus-conf-dir\") pod \"multus-qfksk\" (UID: \"e226b189-cfa9-47fc-b94d-19b5cbe0859f\") " pod="openshift-multus/multus-qfksk" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.843647 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e226b189-cfa9-47fc-b94d-19b5cbe0859f-etc-kubernetes\") pod \"multus-qfksk\" (UID: \"e226b189-cfa9-47fc-b94d-19b5cbe0859f\") " pod="openshift-multus/multus-qfksk" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.843691 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e226b189-cfa9-47fc-b94d-19b5cbe0859f-cnibin\") pod \"multus-qfksk\" (UID: \"e226b189-cfa9-47fc-b94d-19b5cbe0859f\") " pod="openshift-multus/multus-qfksk" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.843705 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e226b189-cfa9-47fc-b94d-19b5cbe0859f-multus-cni-dir\") pod \"multus-qfksk\" (UID: \"e226b189-cfa9-47fc-b94d-19b5cbe0859f\") " pod="openshift-multus/multus-qfksk" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.843720 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e226b189-cfa9-47fc-b94d-19b5cbe0859f-cni-binary-copy\") pod \"multus-qfksk\" (UID: \"e226b189-cfa9-47fc-b94d-19b5cbe0859f\") " pod="openshift-multus/multus-qfksk" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.843737 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/e226b189-cfa9-47fc-b94d-19b5cbe0859f-host-var-lib-cni-multus\") pod \"multus-qfksk\" (UID: \"e226b189-cfa9-47fc-b94d-19b5cbe0859f\") " pod="openshift-multus/multus-qfksk" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.843754 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mx8j8\" (UniqueName: \"kubernetes.io/projected/9945b792-abbd-4103-a7b4-9496e1cc1b56-kube-api-access-mx8j8\") pod \"multus-additional-cni-plugins-sgcq7\" (UID: \"9945b792-abbd-4103-a7b4-9496e1cc1b56\") " pod="openshift-multus/multus-additional-cni-plugins-sgcq7" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.843769 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/e226b189-cfa9-47fc-b94d-19b5cbe0859f-host-run-multus-certs\") pod \"multus-qfksk\" (UID: \"e226b189-cfa9-47fc-b94d-19b5cbe0859f\") " pod="openshift-multus/multus-qfksk" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.843782 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/9945b792-abbd-4103-a7b4-9496e1cc1b56-cni-binary-copy\") pod \"multus-additional-cni-plugins-sgcq7\" (UID: \"9945b792-abbd-4103-a7b4-9496e1cc1b56\") " pod="openshift-multus/multus-additional-cni-plugins-sgcq7" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.843795 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fx4r\" (UniqueName: \"kubernetes.io/projected/e226b189-cfa9-47fc-b94d-19b5cbe0859f-kube-api-access-2fx4r\") pod \"multus-qfksk\" (UID: \"e226b189-cfa9-47fc-b94d-19b5cbe0859f\") " pod="openshift-multus/multus-qfksk" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.843807 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/9945b792-abbd-4103-a7b4-9496e1cc1b56-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-sgcq7\" (UID: \"9945b792-abbd-4103-a7b4-9496e1cc1b56\") " pod="openshift-multus/multus-additional-cni-plugins-sgcq7" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.843862 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8dac7f27-d3d1-4778-9e54-f273035a1d37-proxy-tls\") pod \"machine-config-daemon-f76hs\" (UID: \"8dac7f27-d3d1-4778-9e54-f273035a1d37\") " pod="openshift-machine-config-operator/machine-config-daemon-f76hs" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.843895 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e226b189-cfa9-47fc-b94d-19b5cbe0859f-host-run-netns\") pod \"multus-qfksk\" (UID: \"e226b189-cfa9-47fc-b94d-19b5cbe0859f\") " pod="openshift-multus/multus-qfksk" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.843911 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9945b792-abbd-4103-a7b4-9496e1cc1b56-system-cni-dir\") pod \"multus-additional-cni-plugins-sgcq7\" (UID: \"9945b792-abbd-4103-a7b4-9496e1cc1b56\") " pod="openshift-multus/multus-additional-cni-plugins-sgcq7" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.843927 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9945b792-abbd-4103-a7b4-9496e1cc1b56-tuning-conf-dir\") pod \"multus-additional-cni-plugins-sgcq7\" (UID: \"9945b792-abbd-4103-a7b4-9496e1cc1b56\") " pod="openshift-multus/multus-additional-cni-plugins-sgcq7" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.843968 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/e226b189-cfa9-47fc-b94d-19b5cbe0859f-hostroot\") pod \"multus-qfksk\" (UID: \"e226b189-cfa9-47fc-b94d-19b5cbe0859f\") " pod="openshift-multus/multus-qfksk" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.843991 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/9945b792-abbd-4103-a7b4-9496e1cc1b56-os-release\") pod \"multus-additional-cni-plugins-sgcq7\" (UID: \"9945b792-abbd-4103-a7b4-9496e1cc1b56\") " pod="openshift-multus/multus-additional-cni-plugins-sgcq7" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.844003 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/e226b189-cfa9-47fc-b94d-19b5cbe0859f-host-var-lib-kubelet\") pod \"multus-qfksk\" (UID: \"e226b189-cfa9-47fc-b94d-19b5cbe0859f\") " pod="openshift-multus/multus-qfksk" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.844026 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/8dac7f27-d3d1-4778-9e54-f273035a1d37-rootfs\") pod \"machine-config-daemon-f76hs\" (UID: \"8dac7f27-d3d1-4778-9e54-f273035a1d37\") " pod="openshift-machine-config-operator/machine-config-daemon-f76hs" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.844051 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/d50c39aa-f956-441a-8abe-1d7247a7fd86-hosts-file\") pod \"node-resolver-5vcgz\" (UID: \"d50c39aa-f956-441a-8abe-1d7247a7fd86\") " pod="openshift-dns/node-resolver-5vcgz" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.844064 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8dac7f27-d3d1-4778-9e54-f273035a1d37-mcd-auth-proxy-config\") pod \"machine-config-daemon-f76hs\" (UID: \"8dac7f27-d3d1-4778-9e54-f273035a1d37\") " pod="openshift-machine-config-operator/machine-config-daemon-f76hs" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.844087 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/9945b792-abbd-4103-a7b4-9496e1cc1b56-cnibin\") pod \"multus-additional-cni-plugins-sgcq7\" (UID: \"9945b792-abbd-4103-a7b4-9496e1cc1b56\") " pod="openshift-multus/multus-additional-cni-plugins-sgcq7" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.844116 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4llw\" (UniqueName: \"kubernetes.io/projected/d50c39aa-f956-441a-8abe-1d7247a7fd86-kube-api-access-m4llw\") pod \"node-resolver-5vcgz\" (UID: \"d50c39aa-f956-441a-8abe-1d7247a7fd86\") " pod="openshift-dns/node-resolver-5vcgz" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.844131 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e226b189-cfa9-47fc-b94d-19b5cbe0859f-os-release\") pod \"multus-qfksk\" (UID: \"e226b189-cfa9-47fc-b94d-19b5cbe0859f\") " pod="openshift-multus/multus-qfksk" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.844146 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/e226b189-cfa9-47fc-b94d-19b5cbe0859f-multus-daemon-config\") pod \"multus-qfksk\" (UID: \"e226b189-cfa9-47fc-b94d-19b5cbe0859f\") " pod="openshift-multus/multus-qfksk" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.844967 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"302eb215-a4c2-4324-836d-f37debfd620f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a99dc228d1c4cb8aeab2140ec5f37358b65684739f833fbde45822782430a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14f794063db9d33b82b79acc7134b5a2a3435e09d7cffe49d49251d1f7d20402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5eb3d3728da247903bece56bc2481547317e386e9e97b6c69926a3b55dc89b8e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b42d13c1767c29d22e52537ff45561a0fbf8f2069b2c3612e903e8c127e12485\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e865e6a50c03519ce22958ac4fcbb911fc140c330b6d7ffe5c9d7dd381e96\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T09:04:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0218 09:04:15.143085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 09:04:15.144650 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2160144045/tls.crt::/tmp/serving-cert-2160144045/tls.key\\\\\\\"\\\\nI0218 09:04:20.680799 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 09:04:20.682555 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 09:04:20.682573 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 09:04:20.682590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 09:04:20.682594 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 09:04:20.686822 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 09:04:20.686847 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686852 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686856 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 09:04:20.686859 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 09:04:20.686861 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 09:04:20.686864 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 09:04:20.688257 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 09:04:20.690726 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://974e17ae7082631df2d1bcee60c8a5f142bae2659e9d200d537a4d650f401b4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:26Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.852926 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:26Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.859711 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jflp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689d3d82-4071-4fe5-b3e3-b5b4f511580a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bm85s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jflp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:26Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.869923 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47341f31f94eaf5dd86efbe78ff62873e733b1f86ea895f281b6797770519459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:26Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.879562 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5vcgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d50c39aa-f956-441a-8abe-1d7247a7fd86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4llw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5vcgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:26Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.890005 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jflp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689d3d82-4071-4fe5-b3e3-b5b4f511580a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bm85s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jflp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:26Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.901334 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.901374 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.901383 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.901396 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.901405 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:26Z","lastTransitionTime":"2026-02-18T09:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.923317 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ae60f3-adab-4740-821a-1570059c7101\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f629262e27020512301a7b53ee836ee98e1e7cdee7a52b0ed506bbaaebd823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f59306c0d55e9d00714ce2ee2c1ed6f2ed33d8b21eecde490e86a93335a14f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c87bc671742a47de6c8e68aaf0c8645f9a8cae7b3d52264f36ae3827b3a8d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2678001ff826712da6716541210f018c74fd3d0218b388d2e7a6f16a1f4027\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6f37933281c927328ecfaf842d97ee8361726768f393726cb39517d2bdc9425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:26Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.937714 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8f6014-2544-4173-a94d-441cf42ba891\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77b0269625b7c6baab44a40373618f4ccfe9e1a028a1bc746b7d60c20ac4efaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac09c421b3101887bfb83107dbe26f2b5856d9b32a9cea92f6233fb6ae218cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce79cfe0242ecf403f989dc6b1c1b8be5ab54eee75ac413a8f9d03cef8bca0ed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://095ecf7e8763fa098b06eb9a8aac9f39451f7762bf20196e0803bbc841610e02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:26Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.944737 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/e226b189-cfa9-47fc-b94d-19b5cbe0859f-multus-daemon-config\") pod \"multus-qfksk\" (UID: \"e226b189-cfa9-47fc-b94d-19b5cbe0859f\") " pod="openshift-multus/multus-qfksk" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.944778 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4llw\" (UniqueName: \"kubernetes.io/projected/d50c39aa-f956-441a-8abe-1d7247a7fd86-kube-api-access-m4llw\") pod \"node-resolver-5vcgz\" (UID: \"d50c39aa-f956-441a-8abe-1d7247a7fd86\") " pod="openshift-dns/node-resolver-5vcgz" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.944796 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e226b189-cfa9-47fc-b94d-19b5cbe0859f-os-release\") pod \"multus-qfksk\" (UID: \"e226b189-cfa9-47fc-b94d-19b5cbe0859f\") " pod="openshift-multus/multus-qfksk" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.944812 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e226b189-cfa9-47fc-b94d-19b5cbe0859f-system-cni-dir\") pod \"multus-qfksk\" (UID: \"e226b189-cfa9-47fc-b94d-19b5cbe0859f\") " pod="openshift-multus/multus-qfksk" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.944826 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/e226b189-cfa9-47fc-b94d-19b5cbe0859f-multus-socket-dir-parent\") pod \"multus-qfksk\" (UID: \"e226b189-cfa9-47fc-b94d-19b5cbe0859f\") " pod="openshift-multus/multus-qfksk" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.944840 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/e226b189-cfa9-47fc-b94d-19b5cbe0859f-host-run-k8s-cni-cncf-io\") pod \"multus-qfksk\" (UID: \"e226b189-cfa9-47fc-b94d-19b5cbe0859f\") " pod="openshift-multus/multus-qfksk" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.944859 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e226b189-cfa9-47fc-b94d-19b5cbe0859f-host-var-lib-cni-bin\") pod \"multus-qfksk\" (UID: \"e226b189-cfa9-47fc-b94d-19b5cbe0859f\") " pod="openshift-multus/multus-qfksk" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.944873 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtvs9\" (UniqueName: \"kubernetes.io/projected/8dac7f27-d3d1-4778-9e54-f273035a1d37-kube-api-access-dtvs9\") pod \"machine-config-daemon-f76hs\" (UID: \"8dac7f27-d3d1-4778-9e54-f273035a1d37\") " pod="openshift-machine-config-operator/machine-config-daemon-f76hs" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.944886 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e226b189-cfa9-47fc-b94d-19b5cbe0859f-multus-conf-dir\") pod \"multus-qfksk\" (UID: \"e226b189-cfa9-47fc-b94d-19b5cbe0859f\") " pod="openshift-multus/multus-qfksk" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.944922 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e226b189-cfa9-47fc-b94d-19b5cbe0859f-etc-kubernetes\") pod \"multus-qfksk\" (UID: \"e226b189-cfa9-47fc-b94d-19b5cbe0859f\") " pod="openshift-multus/multus-qfksk" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.944950 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e226b189-cfa9-47fc-b94d-19b5cbe0859f-cnibin\") pod \"multus-qfksk\" (UID: \"e226b189-cfa9-47fc-b94d-19b5cbe0859f\") " pod="openshift-multus/multus-qfksk" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.944965 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e226b189-cfa9-47fc-b94d-19b5cbe0859f-multus-cni-dir\") pod \"multus-qfksk\" (UID: \"e226b189-cfa9-47fc-b94d-19b5cbe0859f\") " pod="openshift-multus/multus-qfksk" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.944978 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e226b189-cfa9-47fc-b94d-19b5cbe0859f-cni-binary-copy\") pod \"multus-qfksk\" (UID: \"e226b189-cfa9-47fc-b94d-19b5cbe0859f\") " pod="openshift-multus/multus-qfksk" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.944993 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/e226b189-cfa9-47fc-b94d-19b5cbe0859f-host-var-lib-cni-multus\") pod \"multus-qfksk\" (UID: \"e226b189-cfa9-47fc-b94d-19b5cbe0859f\") " pod="openshift-multus/multus-qfksk" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.945010 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mx8j8\" (UniqueName: \"kubernetes.io/projected/9945b792-abbd-4103-a7b4-9496e1cc1b56-kube-api-access-mx8j8\") pod \"multus-additional-cni-plugins-sgcq7\" (UID: \"9945b792-abbd-4103-a7b4-9496e1cc1b56\") " pod="openshift-multus/multus-additional-cni-plugins-sgcq7" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.945024 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/e226b189-cfa9-47fc-b94d-19b5cbe0859f-host-run-multus-certs\") pod \"multus-qfksk\" (UID: \"e226b189-cfa9-47fc-b94d-19b5cbe0859f\") " pod="openshift-multus/multus-qfksk" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.945038 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/9945b792-abbd-4103-a7b4-9496e1cc1b56-cni-binary-copy\") pod \"multus-additional-cni-plugins-sgcq7\" (UID: \"9945b792-abbd-4103-a7b4-9496e1cc1b56\") " pod="openshift-multus/multus-additional-cni-plugins-sgcq7" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.945054 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/9945b792-abbd-4103-a7b4-9496e1cc1b56-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-sgcq7\" (UID: \"9945b792-abbd-4103-a7b4-9496e1cc1b56\") " pod="openshift-multus/multus-additional-cni-plugins-sgcq7" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.945069 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fx4r\" (UniqueName: \"kubernetes.io/projected/e226b189-cfa9-47fc-b94d-19b5cbe0859f-kube-api-access-2fx4r\") pod \"multus-qfksk\" (UID: \"e226b189-cfa9-47fc-b94d-19b5cbe0859f\") " pod="openshift-multus/multus-qfksk" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.945082 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9945b792-abbd-4103-a7b4-9496e1cc1b56-system-cni-dir\") pod \"multus-additional-cni-plugins-sgcq7\" (UID: \"9945b792-abbd-4103-a7b4-9496e1cc1b56\") " pod="openshift-multus/multus-additional-cni-plugins-sgcq7" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.945096 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9945b792-abbd-4103-a7b4-9496e1cc1b56-tuning-conf-dir\") pod \"multus-additional-cni-plugins-sgcq7\" (UID: \"9945b792-abbd-4103-a7b4-9496e1cc1b56\") " pod="openshift-multus/multus-additional-cni-plugins-sgcq7" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.945110 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8dac7f27-d3d1-4778-9e54-f273035a1d37-proxy-tls\") pod \"machine-config-daemon-f76hs\" (UID: \"8dac7f27-d3d1-4778-9e54-f273035a1d37\") " pod="openshift-machine-config-operator/machine-config-daemon-f76hs" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.945124 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e226b189-cfa9-47fc-b94d-19b5cbe0859f-host-run-netns\") pod \"multus-qfksk\" (UID: \"e226b189-cfa9-47fc-b94d-19b5cbe0859f\") " pod="openshift-multus/multus-qfksk" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.945138 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/9945b792-abbd-4103-a7b4-9496e1cc1b56-os-release\") pod \"multus-additional-cni-plugins-sgcq7\" (UID: \"9945b792-abbd-4103-a7b4-9496e1cc1b56\") " pod="openshift-multus/multus-additional-cni-plugins-sgcq7" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.945175 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/e226b189-cfa9-47fc-b94d-19b5cbe0859f-hostroot\") pod \"multus-qfksk\" (UID: \"e226b189-cfa9-47fc-b94d-19b5cbe0859f\") " pod="openshift-multus/multus-qfksk" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.945193 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/e226b189-cfa9-47fc-b94d-19b5cbe0859f-host-var-lib-kubelet\") pod \"multus-qfksk\" (UID: \"e226b189-cfa9-47fc-b94d-19b5cbe0859f\") " pod="openshift-multus/multus-qfksk" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.945209 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/8dac7f27-d3d1-4778-9e54-f273035a1d37-rootfs\") pod \"machine-config-daemon-f76hs\" (UID: \"8dac7f27-d3d1-4778-9e54-f273035a1d37\") " pod="openshift-machine-config-operator/machine-config-daemon-f76hs" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.945226 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8dac7f27-d3d1-4778-9e54-f273035a1d37-mcd-auth-proxy-config\") pod \"machine-config-daemon-f76hs\" (UID: \"8dac7f27-d3d1-4778-9e54-f273035a1d37\") " pod="openshift-machine-config-operator/machine-config-daemon-f76hs" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.945239 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/9945b792-abbd-4103-a7b4-9496e1cc1b56-cnibin\") pod \"multus-additional-cni-plugins-sgcq7\" (UID: \"9945b792-abbd-4103-a7b4-9496e1cc1b56\") " pod="openshift-multus/multus-additional-cni-plugins-sgcq7" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.945255 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/d50c39aa-f956-441a-8abe-1d7247a7fd86-hosts-file\") pod \"node-resolver-5vcgz\" (UID: \"d50c39aa-f956-441a-8abe-1d7247a7fd86\") " pod="openshift-dns/node-resolver-5vcgz" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.945321 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/d50c39aa-f956-441a-8abe-1d7247a7fd86-hosts-file\") pod \"node-resolver-5vcgz\" (UID: \"d50c39aa-f956-441a-8abe-1d7247a7fd86\") " pod="openshift-dns/node-resolver-5vcgz" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.945887 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/e226b189-cfa9-47fc-b94d-19b5cbe0859f-multus-daemon-config\") pod \"multus-qfksk\" (UID: \"e226b189-cfa9-47fc-b94d-19b5cbe0859f\") " pod="openshift-multus/multus-qfksk" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.946102 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e226b189-cfa9-47fc-b94d-19b5cbe0859f-os-release\") pod \"multus-qfksk\" (UID: \"e226b189-cfa9-47fc-b94d-19b5cbe0859f\") " pod="openshift-multus/multus-qfksk" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.946204 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e226b189-cfa9-47fc-b94d-19b5cbe0859f-system-cni-dir\") pod \"multus-qfksk\" (UID: \"e226b189-cfa9-47fc-b94d-19b5cbe0859f\") " pod="openshift-multus/multus-qfksk" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.946245 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/e226b189-cfa9-47fc-b94d-19b5cbe0859f-multus-socket-dir-parent\") pod \"multus-qfksk\" (UID: \"e226b189-cfa9-47fc-b94d-19b5cbe0859f\") " pod="openshift-multus/multus-qfksk" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.946266 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/e226b189-cfa9-47fc-b94d-19b5cbe0859f-host-run-k8s-cni-cncf-io\") pod \"multus-qfksk\" (UID: \"e226b189-cfa9-47fc-b94d-19b5cbe0859f\") " pod="openshift-multus/multus-qfksk" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.946286 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e226b189-cfa9-47fc-b94d-19b5cbe0859f-host-var-lib-cni-bin\") pod \"multus-qfksk\" (UID: \"e226b189-cfa9-47fc-b94d-19b5cbe0859f\") " pod="openshift-multus/multus-qfksk" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.946413 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e226b189-cfa9-47fc-b94d-19b5cbe0859f-multus-conf-dir\") pod \"multus-qfksk\" (UID: \"e226b189-cfa9-47fc-b94d-19b5cbe0859f\") " pod="openshift-multus/multus-qfksk" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.946442 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e226b189-cfa9-47fc-b94d-19b5cbe0859f-etc-kubernetes\") pod \"multus-qfksk\" (UID: \"e226b189-cfa9-47fc-b94d-19b5cbe0859f\") " pod="openshift-multus/multus-qfksk" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.946473 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e226b189-cfa9-47fc-b94d-19b5cbe0859f-cnibin\") pod \"multus-qfksk\" (UID: \"e226b189-cfa9-47fc-b94d-19b5cbe0859f\") " pod="openshift-multus/multus-qfksk" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.946508 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e226b189-cfa9-47fc-b94d-19b5cbe0859f-multus-cni-dir\") pod \"multus-qfksk\" (UID: \"e226b189-cfa9-47fc-b94d-19b5cbe0859f\") " pod="openshift-multus/multus-qfksk" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.946893 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e226b189-cfa9-47fc-b94d-19b5cbe0859f-cni-binary-copy\") pod \"multus-qfksk\" (UID: \"e226b189-cfa9-47fc-b94d-19b5cbe0859f\") " pod="openshift-multus/multus-qfksk" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.946934 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/e226b189-cfa9-47fc-b94d-19b5cbe0859f-host-var-lib-cni-multus\") pod \"multus-qfksk\" (UID: \"e226b189-cfa9-47fc-b94d-19b5cbe0859f\") " pod="openshift-multus/multus-qfksk" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.947067 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/e226b189-cfa9-47fc-b94d-19b5cbe0859f-host-run-multus-certs\") pod \"multus-qfksk\" (UID: \"e226b189-cfa9-47fc-b94d-19b5cbe0859f\") " pod="openshift-multus/multus-qfksk" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.947516 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/9945b792-abbd-4103-a7b4-9496e1cc1b56-cni-binary-copy\") pod \"multus-additional-cni-plugins-sgcq7\" (UID: \"9945b792-abbd-4103-a7b4-9496e1cc1b56\") " pod="openshift-multus/multus-additional-cni-plugins-sgcq7" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.947896 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/9945b792-abbd-4103-a7b4-9496e1cc1b56-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-sgcq7\" (UID: \"9945b792-abbd-4103-a7b4-9496e1cc1b56\") " pod="openshift-multus/multus-additional-cni-plugins-sgcq7" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.948037 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9945b792-abbd-4103-a7b4-9496e1cc1b56-system-cni-dir\") pod \"multus-additional-cni-plugins-sgcq7\" (UID: \"9945b792-abbd-4103-a7b4-9496e1cc1b56\") " pod="openshift-multus/multus-additional-cni-plugins-sgcq7" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.948141 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/e226b189-cfa9-47fc-b94d-19b5cbe0859f-hostroot\") pod \"multus-qfksk\" (UID: \"e226b189-cfa9-47fc-b94d-19b5cbe0859f\") " pod="openshift-multus/multus-qfksk" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.948241 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/8dac7f27-d3d1-4778-9e54-f273035a1d37-rootfs\") pod \"machine-config-daemon-f76hs\" (UID: \"8dac7f27-d3d1-4778-9e54-f273035a1d37\") " pod="openshift-machine-config-operator/machine-config-daemon-f76hs" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.948280 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e226b189-cfa9-47fc-b94d-19b5cbe0859f-host-run-netns\") pod \"multus-qfksk\" (UID: \"e226b189-cfa9-47fc-b94d-19b5cbe0859f\") " pod="openshift-multus/multus-qfksk" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.948259 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/9945b792-abbd-4103-a7b4-9496e1cc1b56-cnibin\") pod \"multus-additional-cni-plugins-sgcq7\" (UID: \"9945b792-abbd-4103-a7b4-9496e1cc1b56\") " pod="openshift-multus/multus-additional-cni-plugins-sgcq7" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.948314 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/9945b792-abbd-4103-a7b4-9496e1cc1b56-os-release\") pod \"multus-additional-cni-plugins-sgcq7\" (UID: \"9945b792-abbd-4103-a7b4-9496e1cc1b56\") " pod="openshift-multus/multus-additional-cni-plugins-sgcq7" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.948347 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/e226b189-cfa9-47fc-b94d-19b5cbe0859f-host-var-lib-kubelet\") pod \"multus-qfksk\" (UID: \"e226b189-cfa9-47fc-b94d-19b5cbe0859f\") " pod="openshift-multus/multus-qfksk" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.948571 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9945b792-abbd-4103-a7b4-9496e1cc1b56-tuning-conf-dir\") pod \"multus-additional-cni-plugins-sgcq7\" (UID: \"9945b792-abbd-4103-a7b4-9496e1cc1b56\") " pod="openshift-multus/multus-additional-cni-plugins-sgcq7" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.952539 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8dac7f27-d3d1-4778-9e54-f273035a1d37-proxy-tls\") pod \"machine-config-daemon-f76hs\" (UID: \"8dac7f27-d3d1-4778-9e54-f273035a1d37\") " pod="openshift-machine-config-operator/machine-config-daemon-f76hs" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.960591 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:26Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.964223 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtvs9\" (UniqueName: \"kubernetes.io/projected/8dac7f27-d3d1-4778-9e54-f273035a1d37-kube-api-access-dtvs9\") pod \"machine-config-daemon-f76hs\" (UID: \"8dac7f27-d3d1-4778-9e54-f273035a1d37\") " pod="openshift-machine-config-operator/machine-config-daemon-f76hs" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.966318 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4llw\" (UniqueName: \"kubernetes.io/projected/d50c39aa-f956-441a-8abe-1d7247a7fd86-kube-api-access-m4llw\") pod \"node-resolver-5vcgz\" (UID: \"d50c39aa-f956-441a-8abe-1d7247a7fd86\") " pod="openshift-dns/node-resolver-5vcgz" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.971174 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mx8j8\" (UniqueName: \"kubernetes.io/projected/9945b792-abbd-4103-a7b4-9496e1cc1b56-kube-api-access-mx8j8\") pod \"multus-additional-cni-plugins-sgcq7\" (UID: \"9945b792-abbd-4103-a7b4-9496e1cc1b56\") " pod="openshift-multus/multus-additional-cni-plugins-sgcq7" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.976515 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fx4r\" (UniqueName: \"kubernetes.io/projected/e226b189-cfa9-47fc-b94d-19b5cbe0859f-kube-api-access-2fx4r\") pod \"multus-qfksk\" (UID: \"e226b189-cfa9-47fc-b94d-19b5cbe0859f\") " pod="openshift-multus/multus-qfksk" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.979020 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d653c21d3d7c875bed37ed6da700dc8c801b95c80d7e361121dfbbd3d05a922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5211d0cd5172d45b92b003e2bf97d7969a65ea448b0491f8f6ea67364925ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:26Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:26 crc kubenswrapper[4556]: I0218 09:04:26.992388 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6387f58543cf2601a53b0bbe9be3ec7816ba90d5ffdc4465594b37ccac0f6dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:26Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.003298 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.003319 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.003328 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.003340 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.003348 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:27Z","lastTransitionTime":"2026-02-18T09:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.004088 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:27Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.019299 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"302eb215-a4c2-4324-836d-f37debfd620f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a99dc228d1c4cb8aeab2140ec5f37358b65684739f833fbde45822782430a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14f794063db9d33b82b79acc7134b5a2a3435e09d7cffe49d49251d1f7d20402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5eb3d3728da247903bece56bc2481547317e386e9e97b6c69926a3b55dc89b8e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b42d13c1767c29d22e52537ff45561a0fbf8f2069b2c3612e903e8c127e12485\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e865e6a50c03519ce22958ac4fcbb911fc140c330b6d7ffe5c9d7dd381e96\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T09:04:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0218 09:04:15.143085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 09:04:15.144650 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2160144045/tls.crt::/tmp/serving-cert-2160144045/tls.key\\\\\\\"\\\\nI0218 09:04:20.680799 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 09:04:20.682555 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 09:04:20.682573 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 09:04:20.682590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 09:04:20.682594 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 09:04:20.686822 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 09:04:20.686847 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686852 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686856 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 09:04:20.686859 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 09:04:20.686861 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 09:04:20.686864 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 09:04:20.688257 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 09:04:20.690726 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://974e17ae7082631df2d1bcee60c8a5f142bae2659e9d200d537a4d650f401b4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:27Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.034697 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:27Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.047699 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qfksk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e226b189-cfa9-47fc-b94d-19b5cbe0859f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2fx4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qfksk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:27Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.056256 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dac7f27-d3d1-4778-9e54-f273035a1d37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f76hs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:27Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.062794 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-5vcgz" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.066175 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sgcq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9945b792-abbd-4103-a7b4-9496e1cc1b56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sgcq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:27Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.068884 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-qfksk" Feb 18 09:04:27 crc kubenswrapper[4556]: W0218 09:04:27.074798 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd50c39aa_f956_441a_8abe_1d7247a7fd86.slice/crio-a67adf872d4eafb530e5a64c82997b7765859cc72ba1ed00e3a6b32b574d81a5 WatchSource:0}: Error finding container a67adf872d4eafb530e5a64c82997b7765859cc72ba1ed00e3a6b32b574d81a5: Status 404 returned error can't find the container with id a67adf872d4eafb530e5a64c82997b7765859cc72ba1ed00e3a6b32b574d81a5 Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.077245 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-sgcq7" Feb 18 09:04:27 crc kubenswrapper[4556]: W0218 09:04:27.079065 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode226b189_cfa9_47fc_b94d_19b5cbe0859f.slice/crio-56f34ce3ff374cf5ddd18753c814d3ec175bc7886a229c254d76d83b3f7980df WatchSource:0}: Error finding container 56f34ce3ff374cf5ddd18753c814d3ec175bc7886a229c254d76d83b3f7980df: Status 404 returned error can't find the container with id 56f34ce3ff374cf5ddd18753c814d3ec175bc7886a229c254d76d83b3f7980df Feb 18 09:04:27 crc kubenswrapper[4556]: W0218 09:04:27.085177 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9945b792_abbd_4103_a7b4_9496e1cc1b56.slice/crio-acc4b1cd7cdd69d71738d3df0f12fb2f9192397593af1b559591847588cc0c8a WatchSource:0}: Error finding container acc4b1cd7cdd69d71738d3df0f12fb2f9192397593af1b559591847588cc0c8a: Status 404 returned error can't find the container with id acc4b1cd7cdd69d71738d3df0f12fb2f9192397593af1b559591847588cc0c8a Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.105513 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.105546 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.105554 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.105567 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.105575 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:27Z","lastTransitionTime":"2026-02-18T09:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.131238 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-b7fsj"] Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.132064 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.135406 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.135466 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.135502 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.135599 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.136058 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.136499 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.137132 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.147975 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jflp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689d3d82-4071-4fe5-b3e3-b5b4f511580a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bm85s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jflp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:27Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.159097 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8f6014-2544-4173-a94d-441cf42ba891\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77b0269625b7c6baab44a40373618f4ccfe9e1a028a1bc746b7d60c20ac4efaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac09c421b3101887bfb83107dbe26f2b5856d9b32a9cea92f6233fb6ae218cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce79cfe0242ecf403f989dc6b1c1b8be5ab54eee75ac413a8f9d03cef8bca0ed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://095ecf7e8763fa098b06eb9a8aac9f39451f7762bf20196e0803bbc841610e02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:27Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.168408 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:27Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.179839 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d653c21d3d7c875bed37ed6da700dc8c801b95c80d7e361121dfbbd3d05a922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5211d0cd5172d45b92b003e2bf97d7969a65ea448b0491f8f6ea67364925ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:27Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.193949 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6387f58543cf2601a53b0bbe9be3ec7816ba90d5ffdc4465594b37ccac0f6dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:27Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.202887 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:27Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.207068 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.207100 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.207109 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.207123 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.207131 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:27Z","lastTransitionTime":"2026-02-18T09:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.216702 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ae60f3-adab-4740-821a-1570059c7101\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f629262e27020512301a7b53ee836ee98e1e7cdee7a52b0ed506bbaaebd823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f59306c0d55e9d00714ce2ee2c1ed6f2ed33d8b21eecde490e86a93335a14f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c87bc671742a47de6c8e68aaf0c8645f9a8cae7b3d52264f36ae3827b3a8d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2678001ff826712da6716541210f018c74fd3d0218b388d2e7a6f16a1f4027\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6f37933281c927328ecfaf842d97ee8361726768f393726cb39517d2bdc9425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:27Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.225510 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:27Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.235198 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qfksk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e226b189-cfa9-47fc-b94d-19b5cbe0859f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2fx4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qfksk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:27Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.242778 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dac7f27-d3d1-4778-9e54-f273035a1d37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f76hs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:27Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.248241 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-var-lib-openvswitch\") pod \"ovnkube-node-b7fsj\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.248269 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-host-run-ovn-kubernetes\") pod \"ovnkube-node-b7fsj\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.248285 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-b7fsj\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.248321 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-log-socket\") pod \"ovnkube-node-b7fsj\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.248340 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-host-kubelet\") pod \"ovnkube-node-b7fsj\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.248361 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-run-systemd\") pod \"ovnkube-node-b7fsj\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.248376 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-etc-openvswitch\") pod \"ovnkube-node-b7fsj\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.248390 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bcd67\" (UniqueName: \"kubernetes.io/projected/09971522-31bc-4e10-8831-4fb62675e220-kube-api-access-bcd67\") pod \"ovnkube-node-b7fsj\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.248418 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-host-slash\") pod \"ovnkube-node-b7fsj\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.248433 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/09971522-31bc-4e10-8831-4fb62675e220-ovnkube-script-lib\") pod \"ovnkube-node-b7fsj\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.248446 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/09971522-31bc-4e10-8831-4fb62675e220-ovnkube-config\") pod \"ovnkube-node-b7fsj\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.248460 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/09971522-31bc-4e10-8831-4fb62675e220-env-overrides\") pod \"ovnkube-node-b7fsj\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.248544 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-run-openvswitch\") pod \"ovnkube-node-b7fsj\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.248616 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-node-log\") pod \"ovnkube-node-b7fsj\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.248680 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-host-run-netns\") pod \"ovnkube-node-b7fsj\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.248774 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-host-cni-bin\") pod \"ovnkube-node-b7fsj\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.248818 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-host-cni-netd\") pod \"ovnkube-node-b7fsj\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.248839 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/09971522-31bc-4e10-8831-4fb62675e220-ovn-node-metrics-cert\") pod \"ovnkube-node-b7fsj\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.248896 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-systemd-units\") pod \"ovnkube-node-b7fsj\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.248928 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-run-ovn\") pod \"ovnkube-node-b7fsj\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.253691 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sgcq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9945b792-abbd-4103-a7b4-9496e1cc1b56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sgcq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:27Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.262329 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-03 15:56:15.600606448 +0000 UTC Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.264188 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"302eb215-a4c2-4324-836d-f37debfd620f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a99dc228d1c4cb8aeab2140ec5f37358b65684739f833fbde45822782430a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14f794063db9d33b82b79acc7134b5a2a3435e09d7cffe49d49251d1f7d20402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5eb3d3728da247903bece56bc2481547317e386e9e97b6c69926a3b55dc89b8e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b42d13c1767c29d22e52537ff45561a0fbf8f2069b2c3612e903e8c127e12485\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e865e6a50c03519ce22958ac4fcbb911fc140c330b6d7ffe5c9d7dd381e96\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T09:04:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0218 09:04:15.143085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 09:04:15.144650 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2160144045/tls.crt::/tmp/serving-cert-2160144045/tls.key\\\\\\\"\\\\nI0218 09:04:20.680799 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 09:04:20.682555 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 09:04:20.682573 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 09:04:20.682590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 09:04:20.682594 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 09:04:20.686822 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 09:04:20.686847 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686852 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686856 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 09:04:20.686859 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 09:04:20.686861 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 09:04:20.686864 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 09:04:20.688257 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 09:04:20.690726 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://974e17ae7082631df2d1bcee60c8a5f142bae2659e9d200d537a4d650f401b4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:27Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.272337 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47341f31f94eaf5dd86efbe78ff62873e733b1f86ea895f281b6797770519459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:27Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.276855 4556 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-02-18 08:59:26 +0000 UTC, rotation deadline is 2026-11-24 15:07:35.854414742 +0000 UTC Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.276914 4556 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6702h3m8.577515867s for next certificate rotation Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.279512 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5vcgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d50c39aa-f956-441a-8abe-1d7247a7fd86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4llw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5vcgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:27Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.281693 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.281746 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:04:27 crc kubenswrapper[4556]: E0218 09:04:27.281780 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 09:04:27 crc kubenswrapper[4556]: E0218 09:04:27.281848 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.281929 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:04:27 crc kubenswrapper[4556]: E0218 09:04:27.281986 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.293868 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09971522-31bc-4e10-8831-4fb62675e220\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7fsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:27Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.309412 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.309443 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.309451 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.309466 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.309476 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:27Z","lastTransitionTime":"2026-02-18T09:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.349487 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-node-log\") pod \"ovnkube-node-b7fsj\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.349611 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-node-log\") pod \"ovnkube-node-b7fsj\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.349792 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/09971522-31bc-4e10-8831-4fb62675e220-ovnkube-config\") pod \"ovnkube-node-b7fsj\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.349844 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/09971522-31bc-4e10-8831-4fb62675e220-env-overrides\") pod \"ovnkube-node-b7fsj\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.349867 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-run-openvswitch\") pod \"ovnkube-node-b7fsj\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.349884 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-host-run-netns\") pod \"ovnkube-node-b7fsj\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.349902 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-host-cni-bin\") pod \"ovnkube-node-b7fsj\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.349935 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-host-cni-netd\") pod \"ovnkube-node-b7fsj\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.349950 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/09971522-31bc-4e10-8831-4fb62675e220-ovn-node-metrics-cert\") pod \"ovnkube-node-b7fsj\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.349970 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-systemd-units\") pod \"ovnkube-node-b7fsj\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.349986 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-run-ovn\") pod \"ovnkube-node-b7fsj\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.350006 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-var-lib-openvswitch\") pod \"ovnkube-node-b7fsj\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.350024 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-host-run-ovn-kubernetes\") pod \"ovnkube-node-b7fsj\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.350040 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-b7fsj\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.350045 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-host-cni-bin\") pod \"ovnkube-node-b7fsj\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.350069 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-run-openvswitch\") pod \"ovnkube-node-b7fsj\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.350085 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-log-socket\") pod \"ovnkube-node-b7fsj\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.350064 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-log-socket\") pod \"ovnkube-node-b7fsj\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.350089 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-systemd-units\") pod \"ovnkube-node-b7fsj\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.350108 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-host-run-ovn-kubernetes\") pod \"ovnkube-node-b7fsj\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.350119 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-host-run-netns\") pod \"ovnkube-node-b7fsj\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.350147 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-host-kubelet\") pod \"ovnkube-node-b7fsj\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.350162 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-var-lib-openvswitch\") pod \"ovnkube-node-b7fsj\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.350212 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-host-cni-netd\") pod \"ovnkube-node-b7fsj\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.350256 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-run-ovn\") pod \"ovnkube-node-b7fsj\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.350265 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-host-kubelet\") pod \"ovnkube-node-b7fsj\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.350268 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-b7fsj\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.350289 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-run-systemd\") pod \"ovnkube-node-b7fsj\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.350310 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-run-systemd\") pod \"ovnkube-node-b7fsj\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.350320 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-etc-openvswitch\") pod \"ovnkube-node-b7fsj\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.350339 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bcd67\" (UniqueName: \"kubernetes.io/projected/09971522-31bc-4e10-8831-4fb62675e220-kube-api-access-bcd67\") pod \"ovnkube-node-b7fsj\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.350389 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-host-slash\") pod \"ovnkube-node-b7fsj\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.350391 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-etc-openvswitch\") pod \"ovnkube-node-b7fsj\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.350406 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/09971522-31bc-4e10-8831-4fb62675e220-ovnkube-script-lib\") pod \"ovnkube-node-b7fsj\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.350414 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/09971522-31bc-4e10-8831-4fb62675e220-ovnkube-config\") pod \"ovnkube-node-b7fsj\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.350425 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-host-slash\") pod \"ovnkube-node-b7fsj\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.350531 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/09971522-31bc-4e10-8831-4fb62675e220-env-overrides\") pod \"ovnkube-node-b7fsj\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.350900 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/09971522-31bc-4e10-8831-4fb62675e220-ovnkube-script-lib\") pod \"ovnkube-node-b7fsj\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.352914 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/09971522-31bc-4e10-8831-4fb62675e220-ovn-node-metrics-cert\") pod \"ovnkube-node-b7fsj\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.363129 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bcd67\" (UniqueName: \"kubernetes.io/projected/09971522-31bc-4e10-8831-4fb62675e220-kube-api-access-bcd67\") pod \"ovnkube-node-b7fsj\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.368164 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-5vcgz" event={"ID":"d50c39aa-f956-441a-8abe-1d7247a7fd86","Type":"ContainerStarted","Data":"0ddc5d42163d1fa86f1a59e224adac434c990224e839ef3cb12c6d7689e2b439"} Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.368204 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-5vcgz" event={"ID":"d50c39aa-f956-441a-8abe-1d7247a7fd86","Type":"ContainerStarted","Data":"a67adf872d4eafb530e5a64c82997b7765859cc72ba1ed00e3a6b32b574d81a5"} Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.369352 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-qfksk" event={"ID":"e226b189-cfa9-47fc-b94d-19b5cbe0859f","Type":"ContainerStarted","Data":"fc2cb92a7735d16b8e5918797e52e1e488d9261a0418ed5ebbb94c033ca9328a"} Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.369381 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-qfksk" event={"ID":"e226b189-cfa9-47fc-b94d-19b5cbe0859f","Type":"ContainerStarted","Data":"56f34ce3ff374cf5ddd18753c814d3ec175bc7886a229c254d76d83b3f7980df"} Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.370889 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-2jflp" event={"ID":"689d3d82-4071-4fe5-b3e3-b5b4f511580a","Type":"ContainerStarted","Data":"279d56d65964aaf12feaaa3e5fa3ca779bffdb13fcebfc9f027b563c7dd01f4e"} Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.370918 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-2jflp" event={"ID":"689d3d82-4071-4fe5-b3e3-b5b4f511580a","Type":"ContainerStarted","Data":"239ce6934d18ccf363d93e23154ce1d625ec96ed464059eb45049f8d96d72c09"} Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.372328 4556 generic.go:334] "Generic (PLEG): container finished" podID="9945b792-abbd-4103-a7b4-9496e1cc1b56" containerID="e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0" exitCode=0 Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.372363 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-sgcq7" event={"ID":"9945b792-abbd-4103-a7b4-9496e1cc1b56","Type":"ContainerDied","Data":"e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0"} Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.372406 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-sgcq7" event={"ID":"9945b792-abbd-4103-a7b4-9496e1cc1b56","Type":"ContainerStarted","Data":"acc4b1cd7cdd69d71738d3df0f12fb2f9192397593af1b559591847588cc0c8a"} Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.377778 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:27Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.391858 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ae60f3-adab-4740-821a-1570059c7101\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f629262e27020512301a7b53ee836ee98e1e7cdee7a52b0ed506bbaaebd823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f59306c0d55e9d00714ce2ee2c1ed6f2ed33d8b21eecde490e86a93335a14f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c87bc671742a47de6c8e68aaf0c8645f9a8cae7b3d52264f36ae3827b3a8d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2678001ff826712da6716541210f018c74fd3d0218b388d2e7a6f16a1f4027\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6f37933281c927328ecfaf842d97ee8361726768f393726cb39517d2bdc9425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:27Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.401651 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8f6014-2544-4173-a94d-441cf42ba891\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77b0269625b7c6baab44a40373618f4ccfe9e1a028a1bc746b7d60c20ac4efaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac09c421b3101887bfb83107dbe26f2b5856d9b32a9cea92f6233fb6ae218cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce79cfe0242ecf403f989dc6b1c1b8be5ab54eee75ac413a8f9d03cef8bca0ed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://095ecf7e8763fa098b06eb9a8aac9f39451f7762bf20196e0803bbc841610e02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:27Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.410125 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:27Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.411456 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.411480 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.411489 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.411501 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.411511 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:27Z","lastTransitionTime":"2026-02-18T09:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.418467 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d653c21d3d7c875bed37ed6da700dc8c801b95c80d7e361121dfbbd3d05a922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5211d0cd5172d45b92b003e2bf97d7969a65ea448b0491f8f6ea67364925ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:27Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.428117 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6387f58543cf2601a53b0bbe9be3ec7816ba90d5ffdc4465594b37ccac0f6dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:27Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.437978 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"302eb215-a4c2-4324-836d-f37debfd620f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a99dc228d1c4cb8aeab2140ec5f37358b65684739f833fbde45822782430a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14f794063db9d33b82b79acc7134b5a2a3435e09d7cffe49d49251d1f7d20402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5eb3d3728da247903bece56bc2481547317e386e9e97b6c69926a3b55dc89b8e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b42d13c1767c29d22e52537ff45561a0fbf8f2069b2c3612e903e8c127e12485\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e865e6a50c03519ce22958ac4fcbb911fc140c330b6d7ffe5c9d7dd381e96\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T09:04:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0218 09:04:15.143085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 09:04:15.144650 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2160144045/tls.crt::/tmp/serving-cert-2160144045/tls.key\\\\\\\"\\\\nI0218 09:04:20.680799 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 09:04:20.682555 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 09:04:20.682573 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 09:04:20.682590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 09:04:20.682594 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 09:04:20.686822 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 09:04:20.686847 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686852 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686856 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 09:04:20.686859 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 09:04:20.686861 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 09:04:20.686864 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 09:04:20.688257 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 09:04:20.690726 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://974e17ae7082631df2d1bcee60c8a5f142bae2659e9d200d537a4d650f401b4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:27Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.443901 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.446491 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:27Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:27 crc kubenswrapper[4556]: W0218 09:04:27.455597 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod09971522_31bc_4e10_8831_4fb62675e220.slice/crio-67f192b1fd186997f2fc73bbdf9672b329d777c219cb80c49cfef1f0e2fbfebc WatchSource:0}: Error finding container 67f192b1fd186997f2fc73bbdf9672b329d777c219cb80c49cfef1f0e2fbfebc: Status 404 returned error can't find the container with id 67f192b1fd186997f2fc73bbdf9672b329d777c219cb80c49cfef1f0e2fbfebc Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.461792 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qfksk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e226b189-cfa9-47fc-b94d-19b5cbe0859f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2fx4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qfksk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:27Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.470056 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dac7f27-d3d1-4778-9e54-f273035a1d37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f76hs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:27Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.482364 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sgcq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9945b792-abbd-4103-a7b4-9496e1cc1b56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sgcq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:27Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.492460 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47341f31f94eaf5dd86efbe78ff62873e733b1f86ea895f281b6797770519459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:27Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.503008 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5vcgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d50c39aa-f956-441a-8abe-1d7247a7fd86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ddc5d42163d1fa86f1a59e224adac434c990224e839ef3cb12c6d7689e2b439\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4llw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5vcgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:27Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.513818 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.513847 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.513854 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.513867 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.513875 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:27Z","lastTransitionTime":"2026-02-18T09:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.519500 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09971522-31bc-4e10-8831-4fb62675e220\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7fsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:27Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.548097 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jflp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689d3d82-4071-4fe5-b3e3-b5b4f511580a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bm85s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jflp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:27Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.589113 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jflp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689d3d82-4071-4fe5-b3e3-b5b4f511580a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d56d65964aaf12feaaa3e5fa3ca779bffdb13fcebfc9f027b563c7dd01f4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bm85s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jflp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:27Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.616402 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.616438 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.616446 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.616459 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.616467 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:27Z","lastTransitionTime":"2026-02-18T09:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.630994 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d653c21d3d7c875bed37ed6da700dc8c801b95c80d7e361121dfbbd3d05a922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5211d0cd5172d45b92b003e2bf97d7969a65ea448b0491f8f6ea67364925ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:27Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.670781 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6387f58543cf2601a53b0bbe9be3ec7816ba90d5ffdc4465594b37ccac0f6dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:27Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.710115 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:27Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.718567 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.718602 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.718612 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.718625 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.718641 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:27Z","lastTransitionTime":"2026-02-18T09:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.754593 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ae60f3-adab-4740-821a-1570059c7101\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f629262e27020512301a7b53ee836ee98e1e7cdee7a52b0ed506bbaaebd823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f59306c0d55e9d00714ce2ee2c1ed6f2ed33d8b21eecde490e86a93335a14f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c87bc671742a47de6c8e68aaf0c8645f9a8cae7b3d52264f36ae3827b3a8d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2678001ff826712da6716541210f018c74fd3d0218b388d2e7a6f16a1f4027\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6f37933281c927328ecfaf842d97ee8361726768f393726cb39517d2bdc9425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:27Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.790847 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8f6014-2544-4173-a94d-441cf42ba891\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77b0269625b7c6baab44a40373618f4ccfe9e1a028a1bc746b7d60c20ac4efaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac09c421b3101887bfb83107dbe26f2b5856d9b32a9cea92f6233fb6ae218cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce79cfe0242ecf403f989dc6b1c1b8be5ab54eee75ac413a8f9d03cef8bca0ed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://095ecf7e8763fa098b06eb9a8aac9f39451f7762bf20196e0803bbc841610e02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:27Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.820587 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.820641 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.820664 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.820688 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.820701 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:27Z","lastTransitionTime":"2026-02-18T09:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.831067 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:27Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.868340 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dac7f27-d3d1-4778-9e54-f273035a1d37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f76hs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:27Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.909475 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sgcq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9945b792-abbd-4103-a7b4-9496e1cc1b56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sgcq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:27Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.923418 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.923459 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.923471 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.923493 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.923505 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:27Z","lastTransitionTime":"2026-02-18T09:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:27 crc kubenswrapper[4556]: E0218 09:04:27.949503 4556 configmap.go:193] Couldn't get configMap openshift-machine-config-operator/kube-rbac-proxy: failed to sync configmap cache: timed out waiting for the condition Feb 18 09:04:27 crc kubenswrapper[4556]: E0218 09:04:27.949629 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/8dac7f27-d3d1-4778-9e54-f273035a1d37-mcd-auth-proxy-config podName:8dac7f27-d3d1-4778-9e54-f273035a1d37 nodeName:}" failed. No retries permitted until 2026-02-18 09:04:28.449599519 +0000 UTC m=+25.466560499 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "mcd-auth-proxy-config" (UniqueName: "kubernetes.io/configmap/8dac7f27-d3d1-4778-9e54-f273035a1d37-mcd-auth-proxy-config") pod "machine-config-daemon-f76hs" (UID: "8dac7f27-d3d1-4778-9e54-f273035a1d37") : failed to sync configmap cache: timed out waiting for the condition Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.951049 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"302eb215-a4c2-4324-836d-f37debfd620f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a99dc228d1c4cb8aeab2140ec5f37358b65684739f833fbde45822782430a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14f794063db9d33b82b79acc7134b5a2a3435e09d7cffe49d49251d1f7d20402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5eb3d3728da247903bece56bc2481547317e386e9e97b6c69926a3b55dc89b8e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b42d13c1767c29d22e52537ff45561a0fbf8f2069b2c3612e903e8c127e12485\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e865e6a50c03519ce22958ac4fcbb911fc140c330b6d7ffe5c9d7dd381e96\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T09:04:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0218 09:04:15.143085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 09:04:15.144650 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2160144045/tls.crt::/tmp/serving-cert-2160144045/tls.key\\\\\\\"\\\\nI0218 09:04:20.680799 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 09:04:20.682555 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 09:04:20.682573 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 09:04:20.682590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 09:04:20.682594 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 09:04:20.686822 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 09:04:20.686847 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686852 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686856 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 09:04:20.686859 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 09:04:20.686861 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 09:04:20.686864 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 09:04:20.688257 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 09:04:20.690726 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://974e17ae7082631df2d1bcee60c8a5f142bae2659e9d200d537a4d650f401b4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:27Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:27 crc kubenswrapper[4556]: I0218 09:04:27.989928 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:27Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.026279 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.026325 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.026338 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.026356 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.026368 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:28Z","lastTransitionTime":"2026-02-18T09:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.031040 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qfksk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e226b189-cfa9-47fc-b94d-19b5cbe0859f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc2cb92a7735d16b8e5918797e52e1e488d9261a0418ed5ebbb94c033ca9328a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2fx4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qfksk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:28Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.068237 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5vcgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d50c39aa-f956-441a-8abe-1d7247a7fd86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ddc5d42163d1fa86f1a59e224adac434c990224e839ef3cb12c6d7689e2b439\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4llw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5vcgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:28Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.113497 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09971522-31bc-4e10-8831-4fb62675e220\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7fsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:28Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.129404 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.129442 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.129459 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.129476 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.129486 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:28Z","lastTransitionTime":"2026-02-18T09:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.157311 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47341f31f94eaf5dd86efbe78ff62873e733b1f86ea895f281b6797770519459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:28Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.182844 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.231660 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.231700 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.231710 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.231724 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.231735 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:28Z","lastTransitionTime":"2026-02-18T09:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.263325 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-23 15:50:19.792791236 +0000 UTC Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.333748 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.333804 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.333816 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.333833 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.333843 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:28Z","lastTransitionTime":"2026-02-18T09:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.377524 4556 generic.go:334] "Generic (PLEG): container finished" podID="09971522-31bc-4e10-8831-4fb62675e220" containerID="323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e" exitCode=0 Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.377563 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" event={"ID":"09971522-31bc-4e10-8831-4fb62675e220","Type":"ContainerDied","Data":"323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e"} Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.377616 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" event={"ID":"09971522-31bc-4e10-8831-4fb62675e220","Type":"ContainerStarted","Data":"67f192b1fd186997f2fc73bbdf9672b329d777c219cb80c49cfef1f0e2fbfebc"} Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.380235 4556 generic.go:334] "Generic (PLEG): container finished" podID="9945b792-abbd-4103-a7b4-9496e1cc1b56" containerID="0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca" exitCode=0 Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.380258 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-sgcq7" event={"ID":"9945b792-abbd-4103-a7b4-9496e1cc1b56","Type":"ContainerDied","Data":"0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca"} Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.391269 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5vcgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d50c39aa-f956-441a-8abe-1d7247a7fd86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ddc5d42163d1fa86f1a59e224adac434c990224e839ef3cb12c6d7689e2b439\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4llw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5vcgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:28Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.407828 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09971522-31bc-4e10-8831-4fb62675e220\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7fsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:28Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.422094 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47341f31f94eaf5dd86efbe78ff62873e733b1f86ea895f281b6797770519459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:28Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.431001 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jflp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689d3d82-4071-4fe5-b3e3-b5b4f511580a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d56d65964aaf12feaaa3e5fa3ca779bffdb13fcebfc9f027b563c7dd01f4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bm85s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jflp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:28Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.437175 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.437199 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.437208 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.437222 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.437232 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:28Z","lastTransitionTime":"2026-02-18T09:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.445582 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d653c21d3d7c875bed37ed6da700dc8c801b95c80d7e361121dfbbd3d05a922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5211d0cd5172d45b92b003e2bf97d7969a65ea448b0491f8f6ea67364925ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:28Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.457119 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6387f58543cf2601a53b0bbe9be3ec7816ba90d5ffdc4465594b37ccac0f6dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:28Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.460384 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8dac7f27-d3d1-4778-9e54-f273035a1d37-mcd-auth-proxy-config\") pod \"machine-config-daemon-f76hs\" (UID: \"8dac7f27-d3d1-4778-9e54-f273035a1d37\") " pod="openshift-machine-config-operator/machine-config-daemon-f76hs" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.461012 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8dac7f27-d3d1-4778-9e54-f273035a1d37-mcd-auth-proxy-config\") pod \"machine-config-daemon-f76hs\" (UID: \"8dac7f27-d3d1-4778-9e54-f273035a1d37\") " pod="openshift-machine-config-operator/machine-config-daemon-f76hs" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.469629 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:28Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.494735 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ae60f3-adab-4740-821a-1570059c7101\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f629262e27020512301a7b53ee836ee98e1e7cdee7a52b0ed506bbaaebd823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f59306c0d55e9d00714ce2ee2c1ed6f2ed33d8b21eecde490e86a93335a14f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c87bc671742a47de6c8e68aaf0c8645f9a8cae7b3d52264f36ae3827b3a8d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2678001ff826712da6716541210f018c74fd3d0218b388d2e7a6f16a1f4027\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6f37933281c927328ecfaf842d97ee8361726768f393726cb39517d2bdc9425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:28Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.529864 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8f6014-2544-4173-a94d-441cf42ba891\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77b0269625b7c6baab44a40373618f4ccfe9e1a028a1bc746b7d60c20ac4efaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac09c421b3101887bfb83107dbe26f2b5856d9b32a9cea92f6233fb6ae218cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce79cfe0242ecf403f989dc6b1c1b8be5ab54eee75ac413a8f9d03cef8bca0ed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://095ecf7e8763fa098b06eb9a8aac9f39451f7762bf20196e0803bbc841610e02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:28Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.538789 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.538811 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.538819 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.538832 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.538841 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:28Z","lastTransitionTime":"2026-02-18T09:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.570441 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:28Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.572527 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.609825 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dac7f27-d3d1-4778-9e54-f273035a1d37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f76hs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:28Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:28 crc kubenswrapper[4556]: W0218 09:04:28.637571 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8dac7f27_d3d1_4778_9e54_f273035a1d37.slice/crio-1b7594a9d7d6363a766c7412f80efa2d7bb9d4e4b664bca8d9e01f9d879cc290 WatchSource:0}: Error finding container 1b7594a9d7d6363a766c7412f80efa2d7bb9d4e4b664bca8d9e01f9d879cc290: Status 404 returned error can't find the container with id 1b7594a9d7d6363a766c7412f80efa2d7bb9d4e4b664bca8d9e01f9d879cc290 Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.640499 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.640538 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.640550 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.640568 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.640582 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:28Z","lastTransitionTime":"2026-02-18T09:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.651055 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sgcq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9945b792-abbd-4103-a7b4-9496e1cc1b56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sgcq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:28Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.691731 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"302eb215-a4c2-4324-836d-f37debfd620f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a99dc228d1c4cb8aeab2140ec5f37358b65684739f833fbde45822782430a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14f794063db9d33b82b79acc7134b5a2a3435e09d7cffe49d49251d1f7d20402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5eb3d3728da247903bece56bc2481547317e386e9e97b6c69926a3b55dc89b8e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b42d13c1767c29d22e52537ff45561a0fbf8f2069b2c3612e903e8c127e12485\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e865e6a50c03519ce22958ac4fcbb911fc140c330b6d7ffe5c9d7dd381e96\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T09:04:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0218 09:04:15.143085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 09:04:15.144650 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2160144045/tls.crt::/tmp/serving-cert-2160144045/tls.key\\\\\\\"\\\\nI0218 09:04:20.680799 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 09:04:20.682555 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 09:04:20.682573 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 09:04:20.682590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 09:04:20.682594 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 09:04:20.686822 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 09:04:20.686847 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686852 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686856 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 09:04:20.686859 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 09:04:20.686861 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 09:04:20.686864 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 09:04:20.688257 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 09:04:20.690726 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://974e17ae7082631df2d1bcee60c8a5f142bae2659e9d200d537a4d650f401b4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:28Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.733072 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:28Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.743771 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.744099 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.744109 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.744127 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.744137 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:28Z","lastTransitionTime":"2026-02-18T09:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.773063 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qfksk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e226b189-cfa9-47fc-b94d-19b5cbe0859f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc2cb92a7735d16b8e5918797e52e1e488d9261a0418ed5ebbb94c033ca9328a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2fx4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qfksk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:28Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.807802 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jflp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689d3d82-4071-4fe5-b3e3-b5b4f511580a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d56d65964aaf12feaaa3e5fa3ca779bffdb13fcebfc9f027b563c7dd01f4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bm85s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jflp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:28Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.846182 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.846216 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.846224 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.846238 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.846249 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:28Z","lastTransitionTime":"2026-02-18T09:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.850309 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:28Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.895896 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ae60f3-adab-4740-821a-1570059c7101\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f629262e27020512301a7b53ee836ee98e1e7cdee7a52b0ed506bbaaebd823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f59306c0d55e9d00714ce2ee2c1ed6f2ed33d8b21eecde490e86a93335a14f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c87bc671742a47de6c8e68aaf0c8645f9a8cae7b3d52264f36ae3827b3a8d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2678001ff826712da6716541210f018c74fd3d0218b388d2e7a6f16a1f4027\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6f37933281c927328ecfaf842d97ee8361726768f393726cb39517d2bdc9425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:28Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.929944 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8f6014-2544-4173-a94d-441cf42ba891\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77b0269625b7c6baab44a40373618f4ccfe9e1a028a1bc746b7d60c20ac4efaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac09c421b3101887bfb83107dbe26f2b5856d9b32a9cea92f6233fb6ae218cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce79cfe0242ecf403f989dc6b1c1b8be5ab54eee75ac413a8f9d03cef8bca0ed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://095ecf7e8763fa098b06eb9a8aac9f39451f7762bf20196e0803bbc841610e02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:28Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.947843 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.947874 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.947882 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.947921 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.947933 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:28Z","lastTransitionTime":"2026-02-18T09:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.964479 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.964610 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:04:28 crc kubenswrapper[4556]: E0218 09:04:28.964626 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 09:04:36.964608307 +0000 UTC m=+33.981569287 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:04:28 crc kubenswrapper[4556]: E0218 09:04:28.964791 4556 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 18 09:04:28 crc kubenswrapper[4556]: E0218 09:04:28.964853 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-18 09:04:36.964840352 +0000 UTC m=+33.981801332 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.964874 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:04:28 crc kubenswrapper[4556]: E0218 09:04:28.964934 4556 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 18 09:04:28 crc kubenswrapper[4556]: E0218 09:04:28.964961 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-18 09:04:36.964954357 +0000 UTC m=+33.981915337 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 18 09:04:28 crc kubenswrapper[4556]: I0218 09:04:28.970202 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:28Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.008922 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d653c21d3d7c875bed37ed6da700dc8c801b95c80d7e361121dfbbd3d05a922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5211d0cd5172d45b92b003e2bf97d7969a65ea448b0491f8f6ea67364925ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:29Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.049504 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6387f58543cf2601a53b0bbe9be3ec7816ba90d5ffdc4465594b37ccac0f6dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:29Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.049707 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.049770 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.049785 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.049806 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.049821 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:29Z","lastTransitionTime":"2026-02-18T09:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.066091 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.066179 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:04:29 crc kubenswrapper[4556]: E0218 09:04:29.066287 4556 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 18 09:04:29 crc kubenswrapper[4556]: E0218 09:04:29.066311 4556 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 18 09:04:29 crc kubenswrapper[4556]: E0218 09:04:29.066322 4556 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 09:04:29 crc kubenswrapper[4556]: E0218 09:04:29.066360 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-18 09:04:37.066345656 +0000 UTC m=+34.083306637 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 09:04:29 crc kubenswrapper[4556]: E0218 09:04:29.066289 4556 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 18 09:04:29 crc kubenswrapper[4556]: E0218 09:04:29.066404 4556 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 18 09:04:29 crc kubenswrapper[4556]: E0218 09:04:29.066418 4556 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 09:04:29 crc kubenswrapper[4556]: E0218 09:04:29.066461 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-18 09:04:37.066447998 +0000 UTC m=+34.083408978 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.090439 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"302eb215-a4c2-4324-836d-f37debfd620f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a99dc228d1c4cb8aeab2140ec5f37358b65684739f833fbde45822782430a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14f794063db9d33b82b79acc7134b5a2a3435e09d7cffe49d49251d1f7d20402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5eb3d3728da247903bece56bc2481547317e386e9e97b6c69926a3b55dc89b8e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b42d13c1767c29d22e52537ff45561a0fbf8f2069b2c3612e903e8c127e12485\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e865e6a50c03519ce22958ac4fcbb911fc140c330b6d7ffe5c9d7dd381e96\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T09:04:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0218 09:04:15.143085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 09:04:15.144650 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2160144045/tls.crt::/tmp/serving-cert-2160144045/tls.key\\\\\\\"\\\\nI0218 09:04:20.680799 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 09:04:20.682555 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 09:04:20.682573 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 09:04:20.682590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 09:04:20.682594 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 09:04:20.686822 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 09:04:20.686847 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686852 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686856 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 09:04:20.686859 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 09:04:20.686861 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 09:04:20.686864 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 09:04:20.688257 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 09:04:20.690726 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://974e17ae7082631df2d1bcee60c8a5f142bae2659e9d200d537a4d650f401b4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:29Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.132007 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:29Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.152834 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.152870 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.152882 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.152898 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.152908 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:29Z","lastTransitionTime":"2026-02-18T09:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.170749 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qfksk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e226b189-cfa9-47fc-b94d-19b5cbe0859f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc2cb92a7735d16b8e5918797e52e1e488d9261a0418ed5ebbb94c033ca9328a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2fx4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qfksk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:29Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.210323 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dac7f27-d3d1-4778-9e54-f273035a1d37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f76hs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:29Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.250872 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sgcq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9945b792-abbd-4103-a7b4-9496e1cc1b56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sgcq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:29Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.255139 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.255192 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.255204 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.255220 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.255232 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:29Z","lastTransitionTime":"2026-02-18T09:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.263797 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-01 06:06:29.540637939 +0000 UTC Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.282322 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.282358 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.282417 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:04:29 crc kubenswrapper[4556]: E0218 09:04:29.282509 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 09:04:29 crc kubenswrapper[4556]: E0218 09:04:29.282615 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 09:04:29 crc kubenswrapper[4556]: E0218 09:04:29.282769 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.289485 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47341f31f94eaf5dd86efbe78ff62873e733b1f86ea895f281b6797770519459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:29Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.329694 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5vcgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d50c39aa-f956-441a-8abe-1d7247a7fd86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ddc5d42163d1fa86f1a59e224adac434c990224e839ef3cb12c6d7689e2b439\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4llw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5vcgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:29Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.357527 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.357564 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.357577 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.357593 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.357602 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:29Z","lastTransitionTime":"2026-02-18T09:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.373661 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09971522-31bc-4e10-8831-4fb62675e220\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7fsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:29Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.384580 4556 generic.go:334] "Generic (PLEG): container finished" podID="9945b792-abbd-4103-a7b4-9496e1cc1b56" containerID="51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f" exitCode=0 Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.384681 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-sgcq7" event={"ID":"9945b792-abbd-4103-a7b4-9496e1cc1b56","Type":"ContainerDied","Data":"51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f"} Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.386760 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" event={"ID":"8dac7f27-d3d1-4778-9e54-f273035a1d37","Type":"ContainerStarted","Data":"18da54fe071964d8aae8a977f5ee209c31f1c75e1041c44f8d755f58f3e39d7c"} Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.386799 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" event={"ID":"8dac7f27-d3d1-4778-9e54-f273035a1d37","Type":"ContainerStarted","Data":"6d41b85076e695bbc51a41fce200e0e47c099d88b4d825b4d133d33ed96237e2"} Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.386810 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" event={"ID":"8dac7f27-d3d1-4778-9e54-f273035a1d37","Type":"ContainerStarted","Data":"1b7594a9d7d6363a766c7412f80efa2d7bb9d4e4b664bca8d9e01f9d879cc290"} Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.395029 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" event={"ID":"09971522-31bc-4e10-8831-4fb62675e220","Type":"ContainerStarted","Data":"4cd6e758ee396c9000b1bdb6ded2176ec803fa9da6201cdd9b8146940a2c00cb"} Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.395060 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" event={"ID":"09971522-31bc-4e10-8831-4fb62675e220","Type":"ContainerStarted","Data":"9b0412f3330999c3348ac726ac5a0c2d541571f4b13786ca24cb3109a44aa653"} Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.395070 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" event={"ID":"09971522-31bc-4e10-8831-4fb62675e220","Type":"ContainerStarted","Data":"01c526c0613f19e1a1a584b7129bd56001faf08d0b97fd77feeb69f764ed01f8"} Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.395079 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" event={"ID":"09971522-31bc-4e10-8831-4fb62675e220","Type":"ContainerStarted","Data":"9170bdcbaf1065995d03e6448e40bdced0741c716c9e81f37a72b015132f1161"} Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.395087 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" event={"ID":"09971522-31bc-4e10-8831-4fb62675e220","Type":"ContainerStarted","Data":"318cb69e50df4c14df9bf5ef57d5c381421a9bdb72135c9b9620f55c0547b410"} Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.395094 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" event={"ID":"09971522-31bc-4e10-8831-4fb62675e220","Type":"ContainerStarted","Data":"0c0f15d331af9bc2ca3082c1573dd167c66773eb390593844e681f8d0b455d3d"} Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.414569 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ae60f3-adab-4740-821a-1570059c7101\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f629262e27020512301a7b53ee836ee98e1e7cdee7a52b0ed506bbaaebd823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f59306c0d55e9d00714ce2ee2c1ed6f2ed33d8b21eecde490e86a93335a14f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c87bc671742a47de6c8e68aaf0c8645f9a8cae7b3d52264f36ae3827b3a8d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2678001ff826712da6716541210f018c74fd3d0218b388d2e7a6f16a1f4027\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6f37933281c927328ecfaf842d97ee8361726768f393726cb39517d2bdc9425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:29Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.450527 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8f6014-2544-4173-a94d-441cf42ba891\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77b0269625b7c6baab44a40373618f4ccfe9e1a028a1bc746b7d60c20ac4efaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac09c421b3101887bfb83107dbe26f2b5856d9b32a9cea92f6233fb6ae218cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce79cfe0242ecf403f989dc6b1c1b8be5ab54eee75ac413a8f9d03cef8bca0ed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://095ecf7e8763fa098b06eb9a8aac9f39451f7762bf20196e0803bbc841610e02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:29Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.459723 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.459755 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.459767 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.459781 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.459792 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:29Z","lastTransitionTime":"2026-02-18T09:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.489389 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:29Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.530817 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d653c21d3d7c875bed37ed6da700dc8c801b95c80d7e361121dfbbd3d05a922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5211d0cd5172d45b92b003e2bf97d7969a65ea448b0491f8f6ea67364925ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:29Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.561475 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.561514 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.561523 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.561538 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.561550 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:29Z","lastTransitionTime":"2026-02-18T09:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.570647 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6387f58543cf2601a53b0bbe9be3ec7816ba90d5ffdc4465594b37ccac0f6dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:29Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.610105 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:29Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.649312 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"302eb215-a4c2-4324-836d-f37debfd620f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a99dc228d1c4cb8aeab2140ec5f37358b65684739f833fbde45822782430a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14f794063db9d33b82b79acc7134b5a2a3435e09d7cffe49d49251d1f7d20402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5eb3d3728da247903bece56bc2481547317e386e9e97b6c69926a3b55dc89b8e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b42d13c1767c29d22e52537ff45561a0fbf8f2069b2c3612e903e8c127e12485\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e865e6a50c03519ce22958ac4fcbb911fc140c330b6d7ffe5c9d7dd381e96\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T09:04:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0218 09:04:15.143085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 09:04:15.144650 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2160144045/tls.crt::/tmp/serving-cert-2160144045/tls.key\\\\\\\"\\\\nI0218 09:04:20.680799 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 09:04:20.682555 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 09:04:20.682573 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 09:04:20.682590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 09:04:20.682594 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 09:04:20.686822 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 09:04:20.686847 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686852 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686856 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 09:04:20.686859 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 09:04:20.686861 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 09:04:20.686864 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 09:04:20.688257 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 09:04:20.690726 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://974e17ae7082631df2d1bcee60c8a5f142bae2659e9d200d537a4d650f401b4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:29Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.664025 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.664060 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.664069 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.664083 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.664091 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:29Z","lastTransitionTime":"2026-02-18T09:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.691777 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:29Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.729947 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qfksk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e226b189-cfa9-47fc-b94d-19b5cbe0859f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc2cb92a7735d16b8e5918797e52e1e488d9261a0418ed5ebbb94c033ca9328a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2fx4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qfksk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:29Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.766289 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.766321 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.766333 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.766349 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.766359 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:29Z","lastTransitionTime":"2026-02-18T09:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.769507 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dac7f27-d3d1-4778-9e54-f273035a1d37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f76hs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:29Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.810736 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sgcq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9945b792-abbd-4103-a7b4-9496e1cc1b56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sgcq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:29Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.848760 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47341f31f94eaf5dd86efbe78ff62873e733b1f86ea895f281b6797770519459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:29Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.868365 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.868399 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.868407 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.868426 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.868435 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:29Z","lastTransitionTime":"2026-02-18T09:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.888850 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5vcgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d50c39aa-f956-441a-8abe-1d7247a7fd86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ddc5d42163d1fa86f1a59e224adac434c990224e839ef3cb12c6d7689e2b439\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4llw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5vcgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:29Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.932685 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09971522-31bc-4e10-8831-4fb62675e220\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7fsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:29Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.969461 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jflp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689d3d82-4071-4fe5-b3e3-b5b4f511580a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d56d65964aaf12feaaa3e5fa3ca779bffdb13fcebfc9f027b563c7dd01f4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bm85s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jflp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:29Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.970687 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.970714 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.970725 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.970744 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:29 crc kubenswrapper[4556]: I0218 09:04:29.970752 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:29Z","lastTransitionTime":"2026-02-18T09:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.008708 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5vcgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d50c39aa-f956-441a-8abe-1d7247a7fd86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ddc5d42163d1fa86f1a59e224adac434c990224e839ef3cb12c6d7689e2b439\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4llw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5vcgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:30Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.053620 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09971522-31bc-4e10-8831-4fb62675e220\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7fsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:30Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.072686 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.072748 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.072759 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.072772 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.072781 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:30Z","lastTransitionTime":"2026-02-18T09:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.089174 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47341f31f94eaf5dd86efbe78ff62873e733b1f86ea895f281b6797770519459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:30Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.127935 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jflp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689d3d82-4071-4fe5-b3e3-b5b4f511580a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d56d65964aaf12feaaa3e5fa3ca779bffdb13fcebfc9f027b563c7dd01f4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bm85s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jflp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:30Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.168896 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d653c21d3d7c875bed37ed6da700dc8c801b95c80d7e361121dfbbd3d05a922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5211d0cd5172d45b92b003e2bf97d7969a65ea448b0491f8f6ea67364925ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:30Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.174029 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.174058 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.174067 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.174079 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.174089 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:30Z","lastTransitionTime":"2026-02-18T09:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.209296 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6387f58543cf2601a53b0bbe9be3ec7816ba90d5ffdc4465594b37ccac0f6dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:30Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.250038 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:30Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.263921 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 16:19:03.889028563 +0000 UTC Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.276514 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.276542 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.276550 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.276560 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.276568 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:30Z","lastTransitionTime":"2026-02-18T09:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.294063 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ae60f3-adab-4740-821a-1570059c7101\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f629262e27020512301a7b53ee836ee98e1e7cdee7a52b0ed506bbaaebd823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f59306c0d55e9d00714ce2ee2c1ed6f2ed33d8b21eecde490e86a93335a14f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c87bc671742a47de6c8e68aaf0c8645f9a8cae7b3d52264f36ae3827b3a8d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2678001ff826712da6716541210f018c74fd3d0218b388d2e7a6f16a1f4027\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6f37933281c927328ecfaf842d97ee8361726768f393726cb39517d2bdc9425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:30Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.330479 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8f6014-2544-4173-a94d-441cf42ba891\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77b0269625b7c6baab44a40373618f4ccfe9e1a028a1bc746b7d60c20ac4efaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac09c421b3101887bfb83107dbe26f2b5856d9b32a9cea92f6233fb6ae218cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce79cfe0242ecf403f989dc6b1c1b8be5ab54eee75ac413a8f9d03cef8bca0ed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://095ecf7e8763fa098b06eb9a8aac9f39451f7762bf20196e0803bbc841610e02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:30Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.368963 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:30Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.378394 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.378424 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.378436 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.378450 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.378459 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:30Z","lastTransitionTime":"2026-02-18T09:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.399920 4556 generic.go:334] "Generic (PLEG): container finished" podID="9945b792-abbd-4103-a7b4-9496e1cc1b56" containerID="ac0d9d7b9bc867b32cf6cbeb804c08e67df757e4b22914e2da69a376511cf75c" exitCode=0 Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.399960 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-sgcq7" event={"ID":"9945b792-abbd-4103-a7b4-9496e1cc1b56","Type":"ContainerDied","Data":"ac0d9d7b9bc867b32cf6cbeb804c08e67df757e4b22914e2da69a376511cf75c"} Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.409571 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dac7f27-d3d1-4778-9e54-f273035a1d37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18da54fe071964d8aae8a977f5ee209c31f1c75e1041c44f8d755f58f3e39d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d41b85076e695bbc51a41fce200e0e47c099d88b4d825b4d133d33ed96237e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f76hs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:30Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.450883 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sgcq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9945b792-abbd-4103-a7b4-9496e1cc1b56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sgcq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:30Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.479736 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.479768 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.479780 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.479793 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.479803 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:30Z","lastTransitionTime":"2026-02-18T09:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.489848 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"302eb215-a4c2-4324-836d-f37debfd620f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a99dc228d1c4cb8aeab2140ec5f37358b65684739f833fbde45822782430a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14f794063db9d33b82b79acc7134b5a2a3435e09d7cffe49d49251d1f7d20402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5eb3d3728da247903bece56bc2481547317e386e9e97b6c69926a3b55dc89b8e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b42d13c1767c29d22e52537ff45561a0fbf8f2069b2c3612e903e8c127e12485\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e865e6a50c03519ce22958ac4fcbb911fc140c330b6d7ffe5c9d7dd381e96\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T09:04:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0218 09:04:15.143085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 09:04:15.144650 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2160144045/tls.crt::/tmp/serving-cert-2160144045/tls.key\\\\\\\"\\\\nI0218 09:04:20.680799 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 09:04:20.682555 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 09:04:20.682573 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 09:04:20.682590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 09:04:20.682594 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 09:04:20.686822 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 09:04:20.686847 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686852 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686856 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 09:04:20.686859 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 09:04:20.686861 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 09:04:20.686864 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 09:04:20.688257 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 09:04:20.690726 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://974e17ae7082631df2d1bcee60c8a5f142bae2659e9d200d537a4d650f401b4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:30Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.529621 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:30Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.570418 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qfksk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e226b189-cfa9-47fc-b94d-19b5cbe0859f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc2cb92a7735d16b8e5918797e52e1e488d9261a0418ed5ebbb94c033ca9328a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2fx4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qfksk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:30Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.583664 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.583704 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.583714 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.583736 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.583745 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:30Z","lastTransitionTime":"2026-02-18T09:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.608242 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jflp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689d3d82-4071-4fe5-b3e3-b5b4f511580a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d56d65964aaf12feaaa3e5fa3ca779bffdb13fcebfc9f027b563c7dd01f4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bm85s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jflp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:30Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.649976 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:30Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.685761 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.685800 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.685810 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.685827 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.685837 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:30Z","lastTransitionTime":"2026-02-18T09:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.695552 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ae60f3-adab-4740-821a-1570059c7101\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f629262e27020512301a7b53ee836ee98e1e7cdee7a52b0ed506bbaaebd823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f59306c0d55e9d00714ce2ee2c1ed6f2ed33d8b21eecde490e86a93335a14f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c87bc671742a47de6c8e68aaf0c8645f9a8cae7b3d52264f36ae3827b3a8d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2678001ff826712da6716541210f018c74fd3d0218b388d2e7a6f16a1f4027\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6f37933281c927328ecfaf842d97ee8361726768f393726cb39517d2bdc9425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:30Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.730385 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8f6014-2544-4173-a94d-441cf42ba891\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77b0269625b7c6baab44a40373618f4ccfe9e1a028a1bc746b7d60c20ac4efaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac09c421b3101887bfb83107dbe26f2b5856d9b32a9cea92f6233fb6ae218cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce79cfe0242ecf403f989dc6b1c1b8be5ab54eee75ac413a8f9d03cef8bca0ed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://095ecf7e8763fa098b06eb9a8aac9f39451f7762bf20196e0803bbc841610e02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:30Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.769767 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:30Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.788092 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.788122 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.788133 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.788173 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.788187 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:30Z","lastTransitionTime":"2026-02-18T09:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.809804 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d653c21d3d7c875bed37ed6da700dc8c801b95c80d7e361121dfbbd3d05a922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5211d0cd5172d45b92b003e2bf97d7969a65ea448b0491f8f6ea67364925ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:30Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.854024 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6387f58543cf2601a53b0bbe9be3ec7816ba90d5ffdc4465594b37ccac0f6dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:30Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.889656 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.889753 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.889813 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.889897 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.889958 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:30Z","lastTransitionTime":"2026-02-18T09:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.891442 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"302eb215-a4c2-4324-836d-f37debfd620f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a99dc228d1c4cb8aeab2140ec5f37358b65684739f833fbde45822782430a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14f794063db9d33b82b79acc7134b5a2a3435e09d7cffe49d49251d1f7d20402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5eb3d3728da247903bece56bc2481547317e386e9e97b6c69926a3b55dc89b8e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b42d13c1767c29d22e52537ff45561a0fbf8f2069b2c3612e903e8c127e12485\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e865e6a50c03519ce22958ac4fcbb911fc140c330b6d7ffe5c9d7dd381e96\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T09:04:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0218 09:04:15.143085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 09:04:15.144650 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2160144045/tls.crt::/tmp/serving-cert-2160144045/tls.key\\\\\\\"\\\\nI0218 09:04:20.680799 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 09:04:20.682555 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 09:04:20.682573 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 09:04:20.682590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 09:04:20.682594 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 09:04:20.686822 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 09:04:20.686847 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686852 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686856 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 09:04:20.686859 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 09:04:20.686861 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 09:04:20.686864 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 09:04:20.688257 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 09:04:20.690726 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://974e17ae7082631df2d1bcee60c8a5f142bae2659e9d200d537a4d650f401b4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:30Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.930414 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:30Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.976442 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qfksk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e226b189-cfa9-47fc-b94d-19b5cbe0859f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc2cb92a7735d16b8e5918797e52e1e488d9261a0418ed5ebbb94c033ca9328a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2fx4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qfksk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:30Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.991825 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.991862 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.991872 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.991887 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:30 crc kubenswrapper[4556]: I0218 09:04:30.991896 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:30Z","lastTransitionTime":"2026-02-18T09:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.013454 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dac7f27-d3d1-4778-9e54-f273035a1d37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18da54fe071964d8aae8a977f5ee209c31f1c75e1041c44f8d755f58f3e39d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d41b85076e695bbc51a41fce200e0e47c099d88b4d825b4d133d33ed96237e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f76hs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:31Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.052720 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sgcq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9945b792-abbd-4103-a7b4-9496e1cc1b56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac0d9d7b9bc867b32cf6cbeb804c08e67df757e4b22914e2da69a376511cf75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac0d9d7b9bc867b32cf6cbeb804c08e67df757e4b22914e2da69a376511cf75c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sgcq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:31Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.092553 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47341f31f94eaf5dd86efbe78ff62873e733b1f86ea895f281b6797770519459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:31Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.093477 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.093511 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.093523 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.093541 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.093552 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:31Z","lastTransitionTime":"2026-02-18T09:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.129784 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5vcgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d50c39aa-f956-441a-8abe-1d7247a7fd86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ddc5d42163d1fa86f1a59e224adac434c990224e839ef3cb12c6d7689e2b439\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4llw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5vcgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:31Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.181016 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09971522-31bc-4e10-8831-4fb62675e220\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7fsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:31Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.196021 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.196063 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.196074 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.196091 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.196102 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:31Z","lastTransitionTime":"2026-02-18T09:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.264263 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 13:57:39.00542422 +0000 UTC Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.281569 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.281649 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:04:31 crc kubenswrapper[4556]: E0218 09:04:31.281691 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.281709 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:04:31 crc kubenswrapper[4556]: E0218 09:04:31.281802 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 09:04:31 crc kubenswrapper[4556]: E0218 09:04:31.281879 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.297541 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.297574 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.297585 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.297599 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.297611 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:31Z","lastTransitionTime":"2026-02-18T09:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.399602 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.399647 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.399656 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.399672 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.399681 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:31Z","lastTransitionTime":"2026-02-18T09:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.407375 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" event={"ID":"09971522-31bc-4e10-8831-4fb62675e220","Type":"ContainerStarted","Data":"aec6030df0a59ad8902d5c694d5cd30d249ca1e895155c57a458e359fa531b7c"} Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.410616 4556 generic.go:334] "Generic (PLEG): container finished" podID="9945b792-abbd-4103-a7b4-9496e1cc1b56" containerID="b807d53d620e44d3196f68492b9f9665593adac804bc7b91569796e5594c4ebb" exitCode=0 Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.410645 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-sgcq7" event={"ID":"9945b792-abbd-4103-a7b4-9496e1cc1b56","Type":"ContainerDied","Data":"b807d53d620e44d3196f68492b9f9665593adac804bc7b91569796e5594c4ebb"} Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.422726 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jflp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689d3d82-4071-4fe5-b3e3-b5b4f511580a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d56d65964aaf12feaaa3e5fa3ca779bffdb13fcebfc9f027b563c7dd01f4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bm85s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jflp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:31Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.441578 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ae60f3-adab-4740-821a-1570059c7101\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f629262e27020512301a7b53ee836ee98e1e7cdee7a52b0ed506bbaaebd823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f59306c0d55e9d00714ce2ee2c1ed6f2ed33d8b21eecde490e86a93335a14f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c87bc671742a47de6c8e68aaf0c8645f9a8cae7b3d52264f36ae3827b3a8d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2678001ff826712da6716541210f018c74fd3d0218b388d2e7a6f16a1f4027\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6f37933281c927328ecfaf842d97ee8361726768f393726cb39517d2bdc9425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:31Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.451286 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8f6014-2544-4173-a94d-441cf42ba891\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77b0269625b7c6baab44a40373618f4ccfe9e1a028a1bc746b7d60c20ac4efaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac09c421b3101887bfb83107dbe26f2b5856d9b32a9cea92f6233fb6ae218cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce79cfe0242ecf403f989dc6b1c1b8be5ab54eee75ac413a8f9d03cef8bca0ed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://095ecf7e8763fa098b06eb9a8aac9f39451f7762bf20196e0803bbc841610e02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:31Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.461291 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:31Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.472060 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d653c21d3d7c875bed37ed6da700dc8c801b95c80d7e361121dfbbd3d05a922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5211d0cd5172d45b92b003e2bf97d7969a65ea448b0491f8f6ea67364925ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:31Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.482527 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6387f58543cf2601a53b0bbe9be3ec7816ba90d5ffdc4465594b37ccac0f6dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:31Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.493710 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:31Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.501881 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.501922 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.501932 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.501948 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.501962 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:31Z","lastTransitionTime":"2026-02-18T09:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.502843 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"302eb215-a4c2-4324-836d-f37debfd620f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a99dc228d1c4cb8aeab2140ec5f37358b65684739f833fbde45822782430a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14f794063db9d33b82b79acc7134b5a2a3435e09d7cffe49d49251d1f7d20402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5eb3d3728da247903bece56bc2481547317e386e9e97b6c69926a3b55dc89b8e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b42d13c1767c29d22e52537ff45561a0fbf8f2069b2c3612e903e8c127e12485\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e865e6a50c03519ce22958ac4fcbb911fc140c330b6d7ffe5c9d7dd381e96\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T09:04:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0218 09:04:15.143085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 09:04:15.144650 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2160144045/tls.crt::/tmp/serving-cert-2160144045/tls.key\\\\\\\"\\\\nI0218 09:04:20.680799 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 09:04:20.682555 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 09:04:20.682573 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 09:04:20.682590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 09:04:20.682594 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 09:04:20.686822 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 09:04:20.686847 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686852 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686856 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 09:04:20.686859 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 09:04:20.686861 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 09:04:20.686864 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 09:04:20.688257 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 09:04:20.690726 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://974e17ae7082631df2d1bcee60c8a5f142bae2659e9d200d537a4d650f401b4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:31Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.531089 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:31Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.572392 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qfksk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e226b189-cfa9-47fc-b94d-19b5cbe0859f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc2cb92a7735d16b8e5918797e52e1e488d9261a0418ed5ebbb94c033ca9328a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2fx4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qfksk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:31Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.604634 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.604716 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.604732 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.604760 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.604794 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:31Z","lastTransitionTime":"2026-02-18T09:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.612083 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dac7f27-d3d1-4778-9e54-f273035a1d37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18da54fe071964d8aae8a977f5ee209c31f1c75e1041c44f8d755f58f3e39d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d41b85076e695bbc51a41fce200e0e47c099d88b4d825b4d133d33ed96237e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f76hs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:31Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.652429 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sgcq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9945b792-abbd-4103-a7b4-9496e1cc1b56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac0d9d7b9bc867b32cf6cbeb804c08e67df757e4b22914e2da69a376511cf75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac0d9d7b9bc867b32cf6cbeb804c08e67df757e4b22914e2da69a376511cf75c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b807d53d620e44d3196f68492b9f9665593adac804bc7b91569796e5594c4ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b807d53d620e44d3196f68492b9f9665593adac804bc7b91569796e5594c4ebb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sgcq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:31Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.689525 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47341f31f94eaf5dd86efbe78ff62873e733b1f86ea895f281b6797770519459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:31Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.707885 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.707916 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.707925 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.707941 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.707952 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:31Z","lastTransitionTime":"2026-02-18T09:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.729529 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5vcgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d50c39aa-f956-441a-8abe-1d7247a7fd86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ddc5d42163d1fa86f1a59e224adac434c990224e839ef3cb12c6d7689e2b439\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4llw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5vcgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:31Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.775120 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09971522-31bc-4e10-8831-4fb62675e220\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7fsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:31Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.810678 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.810729 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.810742 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.810766 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.810784 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:31Z","lastTransitionTime":"2026-02-18T09:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.912331 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.912368 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.912378 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.912396 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:31 crc kubenswrapper[4556]: I0218 09:04:31.912406 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:31Z","lastTransitionTime":"2026-02-18T09:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.014978 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.015024 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.015037 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.015054 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.015068 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:32Z","lastTransitionTime":"2026-02-18T09:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.120193 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.120218 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.120228 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.120240 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.120248 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:32Z","lastTransitionTime":"2026-02-18T09:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.222167 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.222209 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.222220 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.222235 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.222245 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:32Z","lastTransitionTime":"2026-02-18T09:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.264705 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-19 22:46:23.944960513 +0000 UTC Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.324533 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.324601 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.324614 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.324651 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.324667 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:32Z","lastTransitionTime":"2026-02-18T09:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.417115 4556 generic.go:334] "Generic (PLEG): container finished" podID="9945b792-abbd-4103-a7b4-9496e1cc1b56" containerID="ebc044bbd5b0c1e725e6ed46ec2e0015b254070b7518d20815539dcf8c6e4e45" exitCode=0 Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.417178 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-sgcq7" event={"ID":"9945b792-abbd-4103-a7b4-9496e1cc1b56","Type":"ContainerDied","Data":"ebc044bbd5b0c1e725e6ed46ec2e0015b254070b7518d20815539dcf8c6e4e45"} Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.426991 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.427032 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.427042 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.427056 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.427067 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:32Z","lastTransitionTime":"2026-02-18T09:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.428931 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47341f31f94eaf5dd86efbe78ff62873e733b1f86ea895f281b6797770519459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:32Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.442087 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5vcgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d50c39aa-f956-441a-8abe-1d7247a7fd86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ddc5d42163d1fa86f1a59e224adac434c990224e839ef3cb12c6d7689e2b439\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4llw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5vcgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:32Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.459630 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09971522-31bc-4e10-8831-4fb62675e220\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7fsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:32Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.468500 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jflp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689d3d82-4071-4fe5-b3e3-b5b4f511580a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d56d65964aaf12feaaa3e5fa3ca779bffdb13fcebfc9f027b563c7dd01f4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bm85s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jflp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:32Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.481581 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:32Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.493337 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d653c21d3d7c875bed37ed6da700dc8c801b95c80d7e361121dfbbd3d05a922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5211d0cd5172d45b92b003e2bf97d7969a65ea448b0491f8f6ea67364925ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:32Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.504568 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6387f58543cf2601a53b0bbe9be3ec7816ba90d5ffdc4465594b37ccac0f6dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:32Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.514635 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:32Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.529068 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.529099 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.529107 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.529120 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.529130 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:32Z","lastTransitionTime":"2026-02-18T09:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.532174 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ae60f3-adab-4740-821a-1570059c7101\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f629262e27020512301a7b53ee836ee98e1e7cdee7a52b0ed506bbaaebd823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f59306c0d55e9d00714ce2ee2c1ed6f2ed33d8b21eecde490e86a93335a14f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c87bc671742a47de6c8e68aaf0c8645f9a8cae7b3d52264f36ae3827b3a8d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2678001ff826712da6716541210f018c74fd3d0218b388d2e7a6f16a1f4027\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6f37933281c927328ecfaf842d97ee8361726768f393726cb39517d2bdc9425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:32Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.541234 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8f6014-2544-4173-a94d-441cf42ba891\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77b0269625b7c6baab44a40373618f4ccfe9e1a028a1bc746b7d60c20ac4efaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac09c421b3101887bfb83107dbe26f2b5856d9b32a9cea92f6233fb6ae218cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce79cfe0242ecf403f989dc6b1c1b8be5ab54eee75ac413a8f9d03cef8bca0ed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://095ecf7e8763fa098b06eb9a8aac9f39451f7762bf20196e0803bbc841610e02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:32Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.550780 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qfksk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e226b189-cfa9-47fc-b94d-19b5cbe0859f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc2cb92a7735d16b8e5918797e52e1e488d9261a0418ed5ebbb94c033ca9328a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2fx4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qfksk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:32Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.559146 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dac7f27-d3d1-4778-9e54-f273035a1d37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18da54fe071964d8aae8a977f5ee209c31f1c75e1041c44f8d755f58f3e39d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d41b85076e695bbc51a41fce200e0e47c099d88b4d825b4d133d33ed96237e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f76hs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:32Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.569416 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sgcq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9945b792-abbd-4103-a7b4-9496e1cc1b56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac0d9d7b9bc867b32cf6cbeb804c08e67df757e4b22914e2da69a376511cf75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac0d9d7b9bc867b32cf6cbeb804c08e67df757e4b22914e2da69a376511cf75c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b807d53d620e44d3196f68492b9f9665593adac804bc7b91569796e5594c4ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b807d53d620e44d3196f68492b9f9665593adac804bc7b91569796e5594c4ebb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebc044bbd5b0c1e725e6ed46ec2e0015b254070b7518d20815539dcf8c6e4e45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ebc044bbd5b0c1e725e6ed46ec2e0015b254070b7518d20815539dcf8c6e4e45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sgcq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:32Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.579453 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"302eb215-a4c2-4324-836d-f37debfd620f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a99dc228d1c4cb8aeab2140ec5f37358b65684739f833fbde45822782430a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14f794063db9d33b82b79acc7134b5a2a3435e09d7cffe49d49251d1f7d20402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5eb3d3728da247903bece56bc2481547317e386e9e97b6c69926a3b55dc89b8e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b42d13c1767c29d22e52537ff45561a0fbf8f2069b2c3612e903e8c127e12485\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e865e6a50c03519ce22958ac4fcbb911fc140c330b6d7ffe5c9d7dd381e96\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T09:04:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0218 09:04:15.143085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 09:04:15.144650 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2160144045/tls.crt::/tmp/serving-cert-2160144045/tls.key\\\\\\\"\\\\nI0218 09:04:20.680799 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 09:04:20.682555 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 09:04:20.682573 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 09:04:20.682590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 09:04:20.682594 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 09:04:20.686822 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 09:04:20.686847 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686852 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686856 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 09:04:20.686859 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 09:04:20.686861 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 09:04:20.686864 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 09:04:20.688257 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 09:04:20.690726 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://974e17ae7082631df2d1bcee60c8a5f142bae2659e9d200d537a4d650f401b4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:32Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.589360 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:32Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.631353 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.631386 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.631395 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.631411 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.631421 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:32Z","lastTransitionTime":"2026-02-18T09:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.734543 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.734633 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.734651 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.734692 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.734708 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:32Z","lastTransitionTime":"2026-02-18T09:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.837260 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.837300 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.837312 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.837329 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.837340 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:32Z","lastTransitionTime":"2026-02-18T09:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.940189 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.940234 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.940248 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.940266 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:32 crc kubenswrapper[4556]: I0218 09:04:32.940278 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:32Z","lastTransitionTime":"2026-02-18T09:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.042523 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.042568 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.042579 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.042598 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.042623 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:33Z","lastTransitionTime":"2026-02-18T09:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.082245 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.095742 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sgcq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9945b792-abbd-4103-a7b4-9496e1cc1b56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac0d9d7b9bc867b32cf6cbeb804c08e67df757e4b22914e2da69a376511cf75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac0d9d7b9bc867b32cf6cbeb804c08e67df757e4b22914e2da69a376511cf75c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b807d53d620e44d3196f68492b9f9665593adac804bc7b91569796e5594c4ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b807d53d620e44d3196f68492b9f9665593adac804bc7b91569796e5594c4ebb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebc044bbd5b0c1e725e6ed46ec2e0015b254070b7518d20815539dcf8c6e4e45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ebc044bbd5b0c1e725e6ed46ec2e0015b254070b7518d20815539dcf8c6e4e45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sgcq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:33Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.108118 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"302eb215-a4c2-4324-836d-f37debfd620f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a99dc228d1c4cb8aeab2140ec5f37358b65684739f833fbde45822782430a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14f794063db9d33b82b79acc7134b5a2a3435e09d7cffe49d49251d1f7d20402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5eb3d3728da247903bece56bc2481547317e386e9e97b6c69926a3b55dc89b8e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b42d13c1767c29d22e52537ff45561a0fbf8f2069b2c3612e903e8c127e12485\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e865e6a50c03519ce22958ac4fcbb911fc140c330b6d7ffe5c9d7dd381e96\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T09:04:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0218 09:04:15.143085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 09:04:15.144650 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2160144045/tls.crt::/tmp/serving-cert-2160144045/tls.key\\\\\\\"\\\\nI0218 09:04:20.680799 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 09:04:20.682555 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 09:04:20.682573 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 09:04:20.682590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 09:04:20.682594 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 09:04:20.686822 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 09:04:20.686847 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686852 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686856 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 09:04:20.686859 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 09:04:20.686861 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 09:04:20.686864 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 09:04:20.688257 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 09:04:20.690726 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://974e17ae7082631df2d1bcee60c8a5f142bae2659e9d200d537a4d650f401b4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:33Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.127215 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:33Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.137687 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qfksk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e226b189-cfa9-47fc-b94d-19b5cbe0859f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc2cb92a7735d16b8e5918797e52e1e488d9261a0418ed5ebbb94c033ca9328a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2fx4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qfksk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:33Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.144983 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.145022 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.145033 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.145051 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.145061 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:33Z","lastTransitionTime":"2026-02-18T09:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.147037 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dac7f27-d3d1-4778-9e54-f273035a1d37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18da54fe071964d8aae8a977f5ee209c31f1c75e1041c44f8d755f58f3e39d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d41b85076e695bbc51a41fce200e0e47c099d88b4d825b4d133d33ed96237e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f76hs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:33Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.161118 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09971522-31bc-4e10-8831-4fb62675e220\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7fsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:33Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.170099 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47341f31f94eaf5dd86efbe78ff62873e733b1f86ea895f281b6797770519459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:33Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.178270 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5vcgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d50c39aa-f956-441a-8abe-1d7247a7fd86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ddc5d42163d1fa86f1a59e224adac434c990224e839ef3cb12c6d7689e2b439\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4llw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5vcgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:33Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.181966 4556 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.182645 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jflp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689d3d82-4071-4fe5-b3e3-b5b4f511580a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d56d65964aaf12feaaa3e5fa3ca779bffdb13fcebfc9f027b563c7dd01f4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bm85s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jflp\": Patch \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/pods/node-ca-2jflp/status\": read tcp 192.168.26.217:35510->192.168.26.217:6443: use of closed network connection" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.201779 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6387f58543cf2601a53b0bbe9be3ec7816ba90d5ffdc4465594b37ccac0f6dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:33Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.210073 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:33Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.223298 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ae60f3-adab-4740-821a-1570059c7101\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f629262e27020512301a7b53ee836ee98e1e7cdee7a52b0ed506bbaaebd823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f59306c0d55e9d00714ce2ee2c1ed6f2ed33d8b21eecde490e86a93335a14f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c87bc671742a47de6c8e68aaf0c8645f9a8cae7b3d52264f36ae3827b3a8d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2678001ff826712da6716541210f018c74fd3d0218b388d2e7a6f16a1f4027\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6f37933281c927328ecfaf842d97ee8361726768f393726cb39517d2bdc9425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:33Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.234439 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8f6014-2544-4173-a94d-441cf42ba891\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77b0269625b7c6baab44a40373618f4ccfe9e1a028a1bc746b7d60c20ac4efaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac09c421b3101887bfb83107dbe26f2b5856d9b32a9cea92f6233fb6ae218cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce79cfe0242ecf403f989dc6b1c1b8be5ab54eee75ac413a8f9d03cef8bca0ed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://095ecf7e8763fa098b06eb9a8aac9f39451f7762bf20196e0803bbc841610e02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:33Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.244898 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:33Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.248177 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.248213 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.248224 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.248255 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.248266 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:33Z","lastTransitionTime":"2026-02-18T09:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.255485 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d653c21d3d7c875bed37ed6da700dc8c801b95c80d7e361121dfbbd3d05a922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5211d0cd5172d45b92b003e2bf97d7969a65ea448b0491f8f6ea67364925ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:33Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.265757 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-20 05:48:16.569584171 +0000 UTC Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.282065 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.282118 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:04:33 crc kubenswrapper[4556]: E0218 09:04:33.282182 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.282126 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:04:33 crc kubenswrapper[4556]: E0218 09:04:33.282257 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 09:04:33 crc kubenswrapper[4556]: E0218 09:04:33.282584 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.290572 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jflp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689d3d82-4071-4fe5-b3e3-b5b4f511580a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d56d65964aaf12feaaa3e5fa3ca779bffdb13fcebfc9f027b563c7dd01f4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bm85s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jflp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:33Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.299318 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8f6014-2544-4173-a94d-441cf42ba891\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77b0269625b7c6baab44a40373618f4ccfe9e1a028a1bc746b7d60c20ac4efaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac09c421b3101887bfb83107dbe26f2b5856d9b32a9cea92f6233fb6ae218cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce79cfe0242ecf403f989dc6b1c1b8be5ab54eee75ac413a8f9d03cef8bca0ed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://095ecf7e8763fa098b06eb9a8aac9f39451f7762bf20196e0803bbc841610e02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:33Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.308582 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:33Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.317934 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d653c21d3d7c875bed37ed6da700dc8c801b95c80d7e361121dfbbd3d05a922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5211d0cd5172d45b92b003e2bf97d7969a65ea448b0491f8f6ea67364925ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:33Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.327043 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6387f58543cf2601a53b0bbe9be3ec7816ba90d5ffdc4465594b37ccac0f6dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:33Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.341208 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:33Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.350529 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.350561 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.350571 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.350585 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.350597 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:33Z","lastTransitionTime":"2026-02-18T09:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.357027 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ae60f3-adab-4740-821a-1570059c7101\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f629262e27020512301a7b53ee836ee98e1e7cdee7a52b0ed506bbaaebd823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f59306c0d55e9d00714ce2ee2c1ed6f2ed33d8b21eecde490e86a93335a14f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c87bc671742a47de6c8e68aaf0c8645f9a8cae7b3d52264f36ae3827b3a8d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2678001ff826712da6716541210f018c74fd3d0218b388d2e7a6f16a1f4027\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6f37933281c927328ecfaf842d97ee8361726768f393726cb39517d2bdc9425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:33Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.370200 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:33Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.385801 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qfksk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e226b189-cfa9-47fc-b94d-19b5cbe0859f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc2cb92a7735d16b8e5918797e52e1e488d9261a0418ed5ebbb94c033ca9328a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2fx4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qfksk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:33Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.393982 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dac7f27-d3d1-4778-9e54-f273035a1d37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18da54fe071964d8aae8a977f5ee209c31f1c75e1041c44f8d755f58f3e39d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d41b85076e695bbc51a41fce200e0e47c099d88b4d825b4d133d33ed96237e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f76hs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:33Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.409761 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sgcq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9945b792-abbd-4103-a7b4-9496e1cc1b56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac0d9d7b9bc867b32cf6cbeb804c08e67df757e4b22914e2da69a376511cf75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac0d9d7b9bc867b32cf6cbeb804c08e67df757e4b22914e2da69a376511cf75c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b807d53d620e44d3196f68492b9f9665593adac804bc7b91569796e5594c4ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b807d53d620e44d3196f68492b9f9665593adac804bc7b91569796e5594c4ebb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebc044bbd5b0c1e725e6ed46ec2e0015b254070b7518d20815539dcf8c6e4e45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ebc044bbd5b0c1e725e6ed46ec2e0015b254070b7518d20815539dcf8c6e4e45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sgcq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:33Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.423587 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" event={"ID":"09971522-31bc-4e10-8831-4fb62675e220","Type":"ContainerStarted","Data":"75e3e104e07b685e6805dd63a14e3d79a2023787668d9c547b694b22143cfbf5"} Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.423738 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.423768 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.427770 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-sgcq7" event={"ID":"9945b792-abbd-4103-a7b4-9496e1cc1b56","Type":"ContainerStarted","Data":"e5dd402827182bc79283f1ea1c355b60c4ced53ffa9ce6781b0a0db3109359a5"} Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.445953 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.450366 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"302eb215-a4c2-4324-836d-f37debfd620f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a99dc228d1c4cb8aeab2140ec5f37358b65684739f833fbde45822782430a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14f794063db9d33b82b79acc7134b5a2a3435e09d7cffe49d49251d1f7d20402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5eb3d3728da247903bece56bc2481547317e386e9e97b6c69926a3b55dc89b8e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b42d13c1767c29d22e52537ff45561a0fbf8f2069b2c3612e903e8c127e12485\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e865e6a50c03519ce22958ac4fcbb911fc140c330b6d7ffe5c9d7dd381e96\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T09:04:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0218 09:04:15.143085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 09:04:15.144650 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2160144045/tls.crt::/tmp/serving-cert-2160144045/tls.key\\\\\\\"\\\\nI0218 09:04:20.680799 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 09:04:20.682555 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 09:04:20.682573 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 09:04:20.682590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 09:04:20.682594 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 09:04:20.686822 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 09:04:20.686847 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686852 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686856 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 09:04:20.686859 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 09:04:20.686861 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 09:04:20.686864 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 09:04:20.688257 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 09:04:20.690726 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://974e17ae7082631df2d1bcee60c8a5f142bae2659e9d200d537a4d650f401b4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:33Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.452540 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.452580 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.452590 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.452618 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.452640 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:33Z","lastTransitionTime":"2026-02-18T09:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.489575 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47341f31f94eaf5dd86efbe78ff62873e733b1f86ea895f281b6797770519459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:33Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.527905 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5vcgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d50c39aa-f956-441a-8abe-1d7247a7fd86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ddc5d42163d1fa86f1a59e224adac434c990224e839ef3cb12c6d7689e2b439\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4llw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5vcgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:33Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.555731 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.555780 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.555793 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.555807 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.555818 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:33Z","lastTransitionTime":"2026-02-18T09:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.571793 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09971522-31bc-4e10-8831-4fb62675e220\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7fsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:33Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.609200 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:33Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.656445 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ae60f3-adab-4740-821a-1570059c7101\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f629262e27020512301a7b53ee836ee98e1e7cdee7a52b0ed506bbaaebd823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f59306c0d55e9d00714ce2ee2c1ed6f2ed33d8b21eecde490e86a93335a14f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c87bc671742a47de6c8e68aaf0c8645f9a8cae7b3d52264f36ae3827b3a8d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2678001ff826712da6716541210f018c74fd3d0218b388d2e7a6f16a1f4027\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6f37933281c927328ecfaf842d97ee8361726768f393726cb39517d2bdc9425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:33Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.657924 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.657949 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.657960 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.657973 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.657985 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:33Z","lastTransitionTime":"2026-02-18T09:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.689823 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8f6014-2544-4173-a94d-441cf42ba891\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77b0269625b7c6baab44a40373618f4ccfe9e1a028a1bc746b7d60c20ac4efaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac09c421b3101887bfb83107dbe26f2b5856d9b32a9cea92f6233fb6ae218cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce79cfe0242ecf403f989dc6b1c1b8be5ab54eee75ac413a8f9d03cef8bca0ed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://095ecf7e8763fa098b06eb9a8aac9f39451f7762bf20196e0803bbc841610e02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:33Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.729904 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:33Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.759661 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.759791 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.759852 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.759913 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.759969 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:33Z","lastTransitionTime":"2026-02-18T09:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.769911 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d653c21d3d7c875bed37ed6da700dc8c801b95c80d7e361121dfbbd3d05a922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5211d0cd5172d45b92b003e2bf97d7969a65ea448b0491f8f6ea67364925ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:33Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.811616 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6387f58543cf2601a53b0bbe9be3ec7816ba90d5ffdc4465594b37ccac0f6dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:33Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.850034 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"302eb215-a4c2-4324-836d-f37debfd620f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a99dc228d1c4cb8aeab2140ec5f37358b65684739f833fbde45822782430a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14f794063db9d33b82b79acc7134b5a2a3435e09d7cffe49d49251d1f7d20402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5eb3d3728da247903bece56bc2481547317e386e9e97b6c69926a3b55dc89b8e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b42d13c1767c29d22e52537ff45561a0fbf8f2069b2c3612e903e8c127e12485\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e865e6a50c03519ce22958ac4fcbb911fc140c330b6d7ffe5c9d7dd381e96\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T09:04:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0218 09:04:15.143085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 09:04:15.144650 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2160144045/tls.crt::/tmp/serving-cert-2160144045/tls.key\\\\\\\"\\\\nI0218 09:04:20.680799 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 09:04:20.682555 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 09:04:20.682573 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 09:04:20.682590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 09:04:20.682594 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 09:04:20.686822 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 09:04:20.686847 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686852 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686856 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 09:04:20.686859 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 09:04:20.686861 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 09:04:20.686864 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 09:04:20.688257 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 09:04:20.690726 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://974e17ae7082631df2d1bcee60c8a5f142bae2659e9d200d537a4d650f401b4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:33Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.862761 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.862799 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.862811 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.862828 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.862841 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:33Z","lastTransitionTime":"2026-02-18T09:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.891060 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:33Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.930389 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qfksk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e226b189-cfa9-47fc-b94d-19b5cbe0859f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc2cb92a7735d16b8e5918797e52e1e488d9261a0418ed5ebbb94c033ca9328a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2fx4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qfksk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:33Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.964733 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.964785 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.964797 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.964811 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.964824 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:33Z","lastTransitionTime":"2026-02-18T09:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:33 crc kubenswrapper[4556]: I0218 09:04:33.969492 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dac7f27-d3d1-4778-9e54-f273035a1d37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18da54fe071964d8aae8a977f5ee209c31f1c75e1041c44f8d755f58f3e39d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d41b85076e695bbc51a41fce200e0e47c099d88b4d825b4d133d33ed96237e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f76hs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:33Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.010271 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sgcq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9945b792-abbd-4103-a7b4-9496e1cc1b56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5dd402827182bc79283f1ea1c355b60c4ced53ffa9ce6781b0a0db3109359a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac0d9d7b9bc867b32cf6cbeb804c08e67df757e4b22914e2da69a376511cf75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac0d9d7b9bc867b32cf6cbeb804c08e67df757e4b22914e2da69a376511cf75c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b807d53d620e44d3196f68492b9f9665593adac804bc7b91569796e5594c4ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b807d53d620e44d3196f68492b9f9665593adac804bc7b91569796e5594c4ebb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebc044bbd5b0c1e725e6ed46ec2e0015b254070b7518d20815539dcf8c6e4e45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ebc044bbd5b0c1e725e6ed46ec2e0015b254070b7518d20815539dcf8c6e4e45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sgcq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:34Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.039047 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.039088 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.039100 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.039116 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.039127 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:34Z","lastTransitionTime":"2026-02-18T09:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:34 crc kubenswrapper[4556]: E0218 09:04:34.049177 4556 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8758c02e-ef59-46a7-908c-d5c97feb8ceb\\\",\\\"systemUUID\\\":\\\"0f904183-79bf-4790-8551-43a6eb3adbe4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:34Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.050450 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47341f31f94eaf5dd86efbe78ff62873e733b1f86ea895f281b6797770519459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:34Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.051642 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.051665 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.051673 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.051686 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.051695 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:34Z","lastTransitionTime":"2026-02-18T09:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:34 crc kubenswrapper[4556]: E0218 09:04:34.064778 4556 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8758c02e-ef59-46a7-908c-d5c97feb8ceb\\\",\\\"systemUUID\\\":\\\"0f904183-79bf-4790-8551-43a6eb3adbe4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:34Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.068218 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.068263 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.068276 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.068291 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.068304 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:34Z","lastTransitionTime":"2026-02-18T09:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:34 crc kubenswrapper[4556]: E0218 09:04:34.077045 4556 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8758c02e-ef59-46a7-908c-d5c97feb8ceb\\\",\\\"systemUUID\\\":\\\"0f904183-79bf-4790-8551-43a6eb3adbe4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:34Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.079572 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.079622 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.079635 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.079658 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.079671 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:34Z","lastTransitionTime":"2026-02-18T09:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.087968 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5vcgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d50c39aa-f956-441a-8abe-1d7247a7fd86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ddc5d42163d1fa86f1a59e224adac434c990224e839ef3cb12c6d7689e2b439\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4llw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5vcgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:34Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:34 crc kubenswrapper[4556]: E0218 09:04:34.089639 4556 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8758c02e-ef59-46a7-908c-d5c97feb8ceb\\\",\\\"systemUUID\\\":\\\"0f904183-79bf-4790-8551-43a6eb3adbe4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:34Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.093101 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.093140 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.093165 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.093179 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.093190 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:34Z","lastTransitionTime":"2026-02-18T09:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:34 crc kubenswrapper[4556]: E0218 09:04:34.101245 4556 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8758c02e-ef59-46a7-908c-d5c97feb8ceb\\\",\\\"systemUUID\\\":\\\"0f904183-79bf-4790-8551-43a6eb3adbe4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:34Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:34 crc kubenswrapper[4556]: E0218 09:04:34.101368 4556 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.102570 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.102621 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.102631 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.102641 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.102650 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:34Z","lastTransitionTime":"2026-02-18T09:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.134031 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09971522-31bc-4e10-8831-4fb62675e220\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9170bdcbaf1065995d03e6448e40bdced0741c716c9e81f37a72b015132f1161\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c526c0613f19e1a1a584b7129bd56001faf08d0b97fd77feeb69f764ed01f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cd6e758ee396c9000b1bdb6ded2176ec803fa9da6201cdd9b8146940a2c00cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b0412f3330999c3348ac726ac5a0c2d541571f4b13786ca24cb3109a44aa653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://318cb69e50df4c14df9bf5ef57d5c381421a9bdb72135c9b9620f55c0547b410\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0f15d331af9bc2ca3082c1573dd167c66773eb390593844e681f8d0b455d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75e3e104e07b685e6805dd63a14e3d79a2023787668d9c547b694b22143cfbf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aec6030df0a59ad8902d5c694d5cd30d249ca1e895155c57a458e359fa531b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7fsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:34Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.169250 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jflp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689d3d82-4071-4fe5-b3e3-b5b4f511580a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d56d65964aaf12feaaa3e5fa3ca779bffdb13fcebfc9f027b563c7dd01f4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bm85s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jflp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:34Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.205059 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.205089 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.205097 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.205108 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.205121 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:34Z","lastTransitionTime":"2026-02-18T09:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.266802 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-30 15:38:06.740767088 +0000 UTC Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.307434 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.307461 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.307471 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.307484 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.307494 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:34Z","lastTransitionTime":"2026-02-18T09:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.409722 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.409788 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.409800 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.409831 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.409846 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:34Z","lastTransitionTime":"2026-02-18T09:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.430207 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.454404 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.469088 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jflp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689d3d82-4071-4fe5-b3e3-b5b4f511580a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d56d65964aaf12feaaa3e5fa3ca779bffdb13fcebfc9f027b563c7dd01f4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bm85s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jflp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:34Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.478949 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6387f58543cf2601a53b0bbe9be3ec7816ba90d5ffdc4465594b37ccac0f6dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:34Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.489678 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:34Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.503329 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ae60f3-adab-4740-821a-1570059c7101\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f629262e27020512301a7b53ee836ee98e1e7cdee7a52b0ed506bbaaebd823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f59306c0d55e9d00714ce2ee2c1ed6f2ed33d8b21eecde490e86a93335a14f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c87bc671742a47de6c8e68aaf0c8645f9a8cae7b3d52264f36ae3827b3a8d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2678001ff826712da6716541210f018c74fd3d0218b388d2e7a6f16a1f4027\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6f37933281c927328ecfaf842d97ee8361726768f393726cb39517d2bdc9425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:34Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.511968 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.512008 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.512019 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.512036 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.512047 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:34Z","lastTransitionTime":"2026-02-18T09:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.512810 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8f6014-2544-4173-a94d-441cf42ba891\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77b0269625b7c6baab44a40373618f4ccfe9e1a028a1bc746b7d60c20ac4efaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac09c421b3101887bfb83107dbe26f2b5856d9b32a9cea92f6233fb6ae218cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce79cfe0242ecf403f989dc6b1c1b8be5ab54eee75ac413a8f9d03cef8bca0ed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://095ecf7e8763fa098b06eb9a8aac9f39451f7762bf20196e0803bbc841610e02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:34Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.521901 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:34Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.530710 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d653c21d3d7c875bed37ed6da700dc8c801b95c80d7e361121dfbbd3d05a922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5211d0cd5172d45b92b003e2bf97d7969a65ea448b0491f8f6ea67364925ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:34Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.540907 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sgcq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9945b792-abbd-4103-a7b4-9496e1cc1b56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5dd402827182bc79283f1ea1c355b60c4ced53ffa9ce6781b0a0db3109359a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac0d9d7b9bc867b32cf6cbeb804c08e67df757e4b22914e2da69a376511cf75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac0d9d7b9bc867b32cf6cbeb804c08e67df757e4b22914e2da69a376511cf75c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b807d53d620e44d3196f68492b9f9665593adac804bc7b91569796e5594c4ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b807d53d620e44d3196f68492b9f9665593adac804bc7b91569796e5594c4ebb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebc044bbd5b0c1e725e6ed46ec2e0015b254070b7518d20815539dcf8c6e4e45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ebc044bbd5b0c1e725e6ed46ec2e0015b254070b7518d20815539dcf8c6e4e45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sgcq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:34Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.550036 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"302eb215-a4c2-4324-836d-f37debfd620f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a99dc228d1c4cb8aeab2140ec5f37358b65684739f833fbde45822782430a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14f794063db9d33b82b79acc7134b5a2a3435e09d7cffe49d49251d1f7d20402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5eb3d3728da247903bece56bc2481547317e386e9e97b6c69926a3b55dc89b8e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b42d13c1767c29d22e52537ff45561a0fbf8f2069b2c3612e903e8c127e12485\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e865e6a50c03519ce22958ac4fcbb911fc140c330b6d7ffe5c9d7dd381e96\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T09:04:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0218 09:04:15.143085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 09:04:15.144650 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2160144045/tls.crt::/tmp/serving-cert-2160144045/tls.key\\\\\\\"\\\\nI0218 09:04:20.680799 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 09:04:20.682555 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 09:04:20.682573 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 09:04:20.682590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 09:04:20.682594 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 09:04:20.686822 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 09:04:20.686847 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686852 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686856 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 09:04:20.686859 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 09:04:20.686861 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 09:04:20.686864 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 09:04:20.688257 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 09:04:20.690726 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://974e17ae7082631df2d1bcee60c8a5f142bae2659e9d200d537a4d650f401b4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:34Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.568540 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:34Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.609643 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qfksk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e226b189-cfa9-47fc-b94d-19b5cbe0859f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc2cb92a7735d16b8e5918797e52e1e488d9261a0418ed5ebbb94c033ca9328a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2fx4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qfksk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:34Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.614305 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.614342 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.614355 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.614373 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.614397 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:34Z","lastTransitionTime":"2026-02-18T09:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.649016 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dac7f27-d3d1-4778-9e54-f273035a1d37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18da54fe071964d8aae8a977f5ee209c31f1c75e1041c44f8d755f58f3e39d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d41b85076e695bbc51a41fce200e0e47c099d88b4d825b4d133d33ed96237e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f76hs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:34Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.693323 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09971522-31bc-4e10-8831-4fb62675e220\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9170bdcbaf1065995d03e6448e40bdced0741c716c9e81f37a72b015132f1161\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c526c0613f19e1a1a584b7129bd56001faf08d0b97fd77feeb69f764ed01f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cd6e758ee396c9000b1bdb6ded2176ec803fa9da6201cdd9b8146940a2c00cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b0412f3330999c3348ac726ac5a0c2d541571f4b13786ca24cb3109a44aa653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://318cb69e50df4c14df9bf5ef57d5c381421a9bdb72135c9b9620f55c0547b410\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0f15d331af9bc2ca3082c1573dd167c66773eb390593844e681f8d0b455d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75e3e104e07b685e6805dd63a14e3d79a2023787668d9c547b694b22143cfbf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aec6030df0a59ad8902d5c694d5cd30d249ca1e895155c57a458e359fa531b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7fsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:34Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.717107 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.717138 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.717148 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.717183 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.717195 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:34Z","lastTransitionTime":"2026-02-18T09:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.729392 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47341f31f94eaf5dd86efbe78ff62873e733b1f86ea895f281b6797770519459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:34Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.768369 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5vcgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d50c39aa-f956-441a-8abe-1d7247a7fd86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ddc5d42163d1fa86f1a59e224adac434c990224e839ef3cb12c6d7689e2b439\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4llw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5vcgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:34Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.819319 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.819370 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.819382 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.819400 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.819411 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:34Z","lastTransitionTime":"2026-02-18T09:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.921501 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.921562 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.921578 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.921615 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:34 crc kubenswrapper[4556]: I0218 09:04:34.921629 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:34Z","lastTransitionTime":"2026-02-18T09:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.023461 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.023513 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.023537 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.023561 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.023574 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:35Z","lastTransitionTime":"2026-02-18T09:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.125658 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.125703 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.125714 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.125731 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.125741 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:35Z","lastTransitionTime":"2026-02-18T09:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.228072 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.228112 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.228122 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.228138 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.228147 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:35Z","lastTransitionTime":"2026-02-18T09:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.267782 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-13 10:27:47.118671886 +0000 UTC Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.282194 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.282245 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.282328 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:04:35 crc kubenswrapper[4556]: E0218 09:04:35.282499 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 09:04:35 crc kubenswrapper[4556]: E0218 09:04:35.282633 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 09:04:35 crc kubenswrapper[4556]: E0218 09:04:35.282842 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.329877 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.329926 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.329939 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.329958 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.329968 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:35Z","lastTransitionTime":"2026-02-18T09:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.431676 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.431742 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.431751 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.431764 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.431775 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:35Z","lastTransitionTime":"2026-02-18T09:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.433569 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-b7fsj_09971522-31bc-4e10-8831-4fb62675e220/ovnkube-controller/0.log" Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.438702 4556 generic.go:334] "Generic (PLEG): container finished" podID="09971522-31bc-4e10-8831-4fb62675e220" containerID="75e3e104e07b685e6805dd63a14e3d79a2023787668d9c547b694b22143cfbf5" exitCode=1 Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.438778 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" event={"ID":"09971522-31bc-4e10-8831-4fb62675e220","Type":"ContainerDied","Data":"75e3e104e07b685e6805dd63a14e3d79a2023787668d9c547b694b22143cfbf5"} Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.439306 4556 scope.go:117] "RemoveContainer" containerID="75e3e104e07b685e6805dd63a14e3d79a2023787668d9c547b694b22143cfbf5" Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.455177 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dac7f27-d3d1-4778-9e54-f273035a1d37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18da54fe071964d8aae8a977f5ee209c31f1c75e1041c44f8d755f58f3e39d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d41b85076e695bbc51a41fce200e0e47c099d88b4d825b4d133d33ed96237e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f76hs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:35Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.471639 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sgcq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9945b792-abbd-4103-a7b4-9496e1cc1b56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5dd402827182bc79283f1ea1c355b60c4ced53ffa9ce6781b0a0db3109359a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac0d9d7b9bc867b32cf6cbeb804c08e67df757e4b22914e2da69a376511cf75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac0d9d7b9bc867b32cf6cbeb804c08e67df757e4b22914e2da69a376511cf75c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b807d53d620e44d3196f68492b9f9665593adac804bc7b91569796e5594c4ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b807d53d620e44d3196f68492b9f9665593adac804bc7b91569796e5594c4ebb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebc044bbd5b0c1e725e6ed46ec2e0015b254070b7518d20815539dcf8c6e4e45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ebc044bbd5b0c1e725e6ed46ec2e0015b254070b7518d20815539dcf8c6e4e45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sgcq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:35Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.480947 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"302eb215-a4c2-4324-836d-f37debfd620f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a99dc228d1c4cb8aeab2140ec5f37358b65684739f833fbde45822782430a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14f794063db9d33b82b79acc7134b5a2a3435e09d7cffe49d49251d1f7d20402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5eb3d3728da247903bece56bc2481547317e386e9e97b6c69926a3b55dc89b8e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b42d13c1767c29d22e52537ff45561a0fbf8f2069b2c3612e903e8c127e12485\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e865e6a50c03519ce22958ac4fcbb911fc140c330b6d7ffe5c9d7dd381e96\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T09:04:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0218 09:04:15.143085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 09:04:15.144650 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2160144045/tls.crt::/tmp/serving-cert-2160144045/tls.key\\\\\\\"\\\\nI0218 09:04:20.680799 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 09:04:20.682555 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 09:04:20.682573 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 09:04:20.682590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 09:04:20.682594 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 09:04:20.686822 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 09:04:20.686847 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686852 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686856 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 09:04:20.686859 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 09:04:20.686861 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 09:04:20.686864 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 09:04:20.688257 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 09:04:20.690726 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://974e17ae7082631df2d1bcee60c8a5f142bae2659e9d200d537a4d650f401b4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:35Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.518479 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:35Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.534339 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.534380 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.534392 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.534409 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.534420 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:35Z","lastTransitionTime":"2026-02-18T09:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.537463 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qfksk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e226b189-cfa9-47fc-b94d-19b5cbe0859f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc2cb92a7735d16b8e5918797e52e1e488d9261a0418ed5ebbb94c033ca9328a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2fx4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qfksk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:35Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.546107 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5vcgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d50c39aa-f956-441a-8abe-1d7247a7fd86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ddc5d42163d1fa86f1a59e224adac434c990224e839ef3cb12c6d7689e2b439\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4llw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5vcgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:35Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.560646 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09971522-31bc-4e10-8831-4fb62675e220\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9170bdcbaf1065995d03e6448e40bdced0741c716c9e81f37a72b015132f1161\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c526c0613f19e1a1a584b7129bd56001faf08d0b97fd77feeb69f764ed01f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cd6e758ee396c9000b1bdb6ded2176ec803fa9da6201cdd9b8146940a2c00cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b0412f3330999c3348ac726ac5a0c2d541571f4b13786ca24cb3109a44aa653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://318cb69e50df4c14df9bf5ef57d5c381421a9bdb72135c9b9620f55c0547b410\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0f15d331af9bc2ca3082c1573dd167c66773eb390593844e681f8d0b455d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75e3e104e07b685e6805dd63a14e3d79a2023787668d9c547b694b22143cfbf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e3e104e07b685e6805dd63a14e3d79a2023787668d9c547b694b22143cfbf5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T09:04:34Z\\\",\\\"message\\\":\\\" 5883 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0218 09:04:34.900397 5883 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0218 09:04:34.900434 5883 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0218 09:04:34.900487 5883 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0218 09:04:34.900529 5883 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0218 09:04:34.900567 5883 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0218 09:04:34.900580 5883 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0218 09:04:34.900537 5883 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0218 09:04:34.900634 5883 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0218 09:04:34.900657 5883 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0218 09:04:34.900685 5883 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0218 09:04:34.900700 5883 handler.go:208] Removed *v1.Node event handler 2\\\\nI0218 09:04:34.900720 5883 handler.go:208] Removed *v1.Node event handler 7\\\\nI0218 09:04:34.900685 5883 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0218 09:04:34.900746 5883 factory.go:656] Stopping watch factory\\\\nI0218 09:04:34.900773 5883 ovnkube.go:599] Stopped ovnkube\\\\nI0218 09:04:3\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aec6030df0a59ad8902d5c694d5cd30d249ca1e895155c57a458e359fa531b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7fsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:35Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.573770 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47341f31f94eaf5dd86efbe78ff62873e733b1f86ea895f281b6797770519459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:35Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.583133 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jflp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689d3d82-4071-4fe5-b3e3-b5b4f511580a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d56d65964aaf12feaaa3e5fa3ca779bffdb13fcebfc9f027b563c7dd01f4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bm85s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jflp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:35Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.596178 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d653c21d3d7c875bed37ed6da700dc8c801b95c80d7e361121dfbbd3d05a922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5211d0cd5172d45b92b003e2bf97d7969a65ea448b0491f8f6ea67364925ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:35Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.606753 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6387f58543cf2601a53b0bbe9be3ec7816ba90d5ffdc4465594b37ccac0f6dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:35Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.617081 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:35Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.631726 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ae60f3-adab-4740-821a-1570059c7101\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f629262e27020512301a7b53ee836ee98e1e7cdee7a52b0ed506bbaaebd823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f59306c0d55e9d00714ce2ee2c1ed6f2ed33d8b21eecde490e86a93335a14f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c87bc671742a47de6c8e68aaf0c8645f9a8cae7b3d52264f36ae3827b3a8d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2678001ff826712da6716541210f018c74fd3d0218b388d2e7a6f16a1f4027\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6f37933281c927328ecfaf842d97ee8361726768f393726cb39517d2bdc9425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:35Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.636419 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.636447 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.636457 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.636475 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.636485 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:35Z","lastTransitionTime":"2026-02-18T09:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.641800 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8f6014-2544-4173-a94d-441cf42ba891\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77b0269625b7c6baab44a40373618f4ccfe9e1a028a1bc746b7d60c20ac4efaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac09c421b3101887bfb83107dbe26f2b5856d9b32a9cea92f6233fb6ae218cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce79cfe0242ecf403f989dc6b1c1b8be5ab54eee75ac413a8f9d03cef8bca0ed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://095ecf7e8763fa098b06eb9a8aac9f39451f7762bf20196e0803bbc841610e02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:35Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.651676 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:35Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.738785 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.738817 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.738826 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.738843 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.738854 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:35Z","lastTransitionTime":"2026-02-18T09:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.842959 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.843001 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.843014 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.843030 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.843039 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:35Z","lastTransitionTime":"2026-02-18T09:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.945350 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.945396 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.945406 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.945424 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:35 crc kubenswrapper[4556]: I0218 09:04:35.945436 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:35Z","lastTransitionTime":"2026-02-18T09:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.047606 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.047640 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.047651 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.047666 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.047681 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:36Z","lastTransitionTime":"2026-02-18T09:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.150867 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.150912 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.150923 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.150941 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.150955 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:36Z","lastTransitionTime":"2026-02-18T09:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.252934 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.252974 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.252983 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.252998 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.253009 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:36Z","lastTransitionTime":"2026-02-18T09:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.268296 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-17 06:37:44.354511328 +0000 UTC Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.354824 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.354859 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.354868 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.354881 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.354891 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:36Z","lastTransitionTime":"2026-02-18T09:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.443580 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-b7fsj_09971522-31bc-4e10-8831-4fb62675e220/ovnkube-controller/1.log" Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.444117 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-b7fsj_09971522-31bc-4e10-8831-4fb62675e220/ovnkube-controller/0.log" Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.447082 4556 generic.go:334] "Generic (PLEG): container finished" podID="09971522-31bc-4e10-8831-4fb62675e220" containerID="8fe96f3b6f1f0ce811435d163297af1becac58208736b4da93517aa8780700c8" exitCode=1 Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.447127 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" event={"ID":"09971522-31bc-4e10-8831-4fb62675e220","Type":"ContainerDied","Data":"8fe96f3b6f1f0ce811435d163297af1becac58208736b4da93517aa8780700c8"} Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.447195 4556 scope.go:117] "RemoveContainer" containerID="75e3e104e07b685e6805dd63a14e3d79a2023787668d9c547b694b22143cfbf5" Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.447779 4556 scope.go:117] "RemoveContainer" containerID="8fe96f3b6f1f0ce811435d163297af1becac58208736b4da93517aa8780700c8" Feb 18 09:04:36 crc kubenswrapper[4556]: E0218 09:04:36.447921 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-b7fsj_openshift-ovn-kubernetes(09971522-31bc-4e10-8831-4fb62675e220)\"" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" podUID="09971522-31bc-4e10-8831-4fb62675e220" Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.456530 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47341f31f94eaf5dd86efbe78ff62873e733b1f86ea895f281b6797770519459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:36Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.456681 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.456705 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.456714 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.456726 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.456735 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:36Z","lastTransitionTime":"2026-02-18T09:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.464423 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5vcgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d50c39aa-f956-441a-8abe-1d7247a7fd86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ddc5d42163d1fa86f1a59e224adac434c990224e839ef3cb12c6d7689e2b439\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4llw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5vcgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:36Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.484017 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09971522-31bc-4e10-8831-4fb62675e220\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9170bdcbaf1065995d03e6448e40bdced0741c716c9e81f37a72b015132f1161\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c526c0613f19e1a1a584b7129bd56001faf08d0b97fd77feeb69f764ed01f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cd6e758ee396c9000b1bdb6ded2176ec803fa9da6201cdd9b8146940a2c00cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b0412f3330999c3348ac726ac5a0c2d541571f4b13786ca24cb3109a44aa653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://318cb69e50df4c14df9bf5ef57d5c381421a9bdb72135c9b9620f55c0547b410\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0f15d331af9bc2ca3082c1573dd167c66773eb390593844e681f8d0b455d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fe96f3b6f1f0ce811435d163297af1becac58208736b4da93517aa8780700c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e3e104e07b685e6805dd63a14e3d79a2023787668d9c547b694b22143cfbf5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T09:04:34Z\\\",\\\"message\\\":\\\" 5883 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0218 09:04:34.900397 5883 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0218 09:04:34.900434 5883 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0218 09:04:34.900487 5883 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0218 09:04:34.900529 5883 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0218 09:04:34.900567 5883 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0218 09:04:34.900580 5883 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0218 09:04:34.900537 5883 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0218 09:04:34.900634 5883 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0218 09:04:34.900657 5883 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0218 09:04:34.900685 5883 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0218 09:04:34.900700 5883 handler.go:208] Removed *v1.Node event handler 2\\\\nI0218 09:04:34.900720 5883 handler.go:208] Removed *v1.Node event handler 7\\\\nI0218 09:04:34.900685 5883 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0218 09:04:34.900746 5883 factory.go:656] Stopping watch factory\\\\nI0218 09:04:34.900773 5883 ovnkube.go:599] Stopped ovnkube\\\\nI0218 09:04:3\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:33Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8fe96f3b6f1f0ce811435d163297af1becac58208736b4da93517aa8780700c8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T09:04:36Z\\\",\\\"message\\\":\\\":{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0218 09:04:36.174933 6015 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0218 09:04:36.174942 6015 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI0218 09:04:36.174757 6015 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-sgcq7 in node crc\\\\nI0218 09:04:36.174948 6015 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nF0218 09:04:36.174847 6015 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling w\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aec6030df0a59ad8902d5c694d5cd30d249ca1e895155c57a458e359fa531b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7fsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:36Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.490897 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jflp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689d3d82-4071-4fe5-b3e3-b5b4f511580a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d56d65964aaf12feaaa3e5fa3ca779bffdb13fcebfc9f027b563c7dd01f4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bm85s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jflp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:36Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.504132 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ae60f3-adab-4740-821a-1570059c7101\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f629262e27020512301a7b53ee836ee98e1e7cdee7a52b0ed506bbaaebd823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f59306c0d55e9d00714ce2ee2c1ed6f2ed33d8b21eecde490e86a93335a14f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c87bc671742a47de6c8e68aaf0c8645f9a8cae7b3d52264f36ae3827b3a8d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2678001ff826712da6716541210f018c74fd3d0218b388d2e7a6f16a1f4027\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6f37933281c927328ecfaf842d97ee8361726768f393726cb39517d2bdc9425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:36Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.513741 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8f6014-2544-4173-a94d-441cf42ba891\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77b0269625b7c6baab44a40373618f4ccfe9e1a028a1bc746b7d60c20ac4efaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac09c421b3101887bfb83107dbe26f2b5856d9b32a9cea92f6233fb6ae218cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce79cfe0242ecf403f989dc6b1c1b8be5ab54eee75ac413a8f9d03cef8bca0ed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://095ecf7e8763fa098b06eb9a8aac9f39451f7762bf20196e0803bbc841610e02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:36Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.523270 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:36Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.532832 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d653c21d3d7c875bed37ed6da700dc8c801b95c80d7e361121dfbbd3d05a922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5211d0cd5172d45b92b003e2bf97d7969a65ea448b0491f8f6ea67364925ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:36Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.545383 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6387f58543cf2601a53b0bbe9be3ec7816ba90d5ffdc4465594b37ccac0f6dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:36Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.555275 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:36Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.558578 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.558624 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.558634 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.558651 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.558663 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:36Z","lastTransitionTime":"2026-02-18T09:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.565289 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"302eb215-a4c2-4324-836d-f37debfd620f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a99dc228d1c4cb8aeab2140ec5f37358b65684739f833fbde45822782430a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14f794063db9d33b82b79acc7134b5a2a3435e09d7cffe49d49251d1f7d20402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5eb3d3728da247903bece56bc2481547317e386e9e97b6c69926a3b55dc89b8e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b42d13c1767c29d22e52537ff45561a0fbf8f2069b2c3612e903e8c127e12485\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e865e6a50c03519ce22958ac4fcbb911fc140c330b6d7ffe5c9d7dd381e96\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T09:04:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0218 09:04:15.143085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 09:04:15.144650 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2160144045/tls.crt::/tmp/serving-cert-2160144045/tls.key\\\\\\\"\\\\nI0218 09:04:20.680799 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 09:04:20.682555 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 09:04:20.682573 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 09:04:20.682590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 09:04:20.682594 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 09:04:20.686822 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 09:04:20.686847 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686852 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686856 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 09:04:20.686859 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 09:04:20.686861 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 09:04:20.686864 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 09:04:20.688257 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 09:04:20.690726 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://974e17ae7082631df2d1bcee60c8a5f142bae2659e9d200d537a4d650f401b4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:36Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.573781 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:36Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.582569 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qfksk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e226b189-cfa9-47fc-b94d-19b5cbe0859f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc2cb92a7735d16b8e5918797e52e1e488d9261a0418ed5ebbb94c033ca9328a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2fx4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qfksk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:36Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.591178 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dac7f27-d3d1-4778-9e54-f273035a1d37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18da54fe071964d8aae8a977f5ee209c31f1c75e1041c44f8d755f58f3e39d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d41b85076e695bbc51a41fce200e0e47c099d88b4d825b4d133d33ed96237e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f76hs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:36Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.601495 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sgcq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9945b792-abbd-4103-a7b4-9496e1cc1b56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5dd402827182bc79283f1ea1c355b60c4ced53ffa9ce6781b0a0db3109359a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac0d9d7b9bc867b32cf6cbeb804c08e67df757e4b22914e2da69a376511cf75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac0d9d7b9bc867b32cf6cbeb804c08e67df757e4b22914e2da69a376511cf75c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b807d53d620e44d3196f68492b9f9665593adac804bc7b91569796e5594c4ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b807d53d620e44d3196f68492b9f9665593adac804bc7b91569796e5594c4ebb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebc044bbd5b0c1e725e6ed46ec2e0015b254070b7518d20815539dcf8c6e4e45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ebc044bbd5b0c1e725e6ed46ec2e0015b254070b7518d20815539dcf8c6e4e45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sgcq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:36Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.660417 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.660455 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.660466 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.660482 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.660492 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:36Z","lastTransitionTime":"2026-02-18T09:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.763676 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.763744 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.763754 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.763769 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.763790 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:36Z","lastTransitionTime":"2026-02-18T09:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.866161 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.866210 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.866221 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.866238 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.866249 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:36Z","lastTransitionTime":"2026-02-18T09:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.968965 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.969026 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.969038 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.969053 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:36 crc kubenswrapper[4556]: I0218 09:04:36.969064 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:36Z","lastTransitionTime":"2026-02-18T09:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.043653 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.043819 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:04:37 crc kubenswrapper[4556]: E0218 09:04:37.043844 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 09:04:53.043820908 +0000 UTC m=+50.060781888 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.043898 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:04:37 crc kubenswrapper[4556]: E0218 09:04:37.043969 4556 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 18 09:04:37 crc kubenswrapper[4556]: E0218 09:04:37.043987 4556 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 18 09:04:37 crc kubenswrapper[4556]: E0218 09:04:37.044037 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-18 09:04:53.044028708 +0000 UTC m=+50.060989698 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 18 09:04:37 crc kubenswrapper[4556]: E0218 09:04:37.044057 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-18 09:04:53.044050539 +0000 UTC m=+50.061011529 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.071638 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.071677 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.071687 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.071700 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.071710 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:37Z","lastTransitionTime":"2026-02-18T09:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.144314 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.144382 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:04:37 crc kubenswrapper[4556]: E0218 09:04:37.144576 4556 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 18 09:04:37 crc kubenswrapper[4556]: E0218 09:04:37.144624 4556 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 18 09:04:37 crc kubenswrapper[4556]: E0218 09:04:37.144575 4556 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 18 09:04:37 crc kubenswrapper[4556]: E0218 09:04:37.144681 4556 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 18 09:04:37 crc kubenswrapper[4556]: E0218 09:04:37.144700 4556 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 09:04:37 crc kubenswrapper[4556]: E0218 09:04:37.144754 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-18 09:04:53.144737206 +0000 UTC m=+50.161698196 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 09:04:37 crc kubenswrapper[4556]: E0218 09:04:37.144643 4556 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 09:04:37 crc kubenswrapper[4556]: E0218 09:04:37.144845 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-18 09:04:53.144823557 +0000 UTC m=+50.161784547 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.173518 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.173555 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.173566 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.173589 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.173602 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:37Z","lastTransitionTime":"2026-02-18T09:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.268785 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-14 13:17:23.259292113 +0000 UTC Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.275956 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.275989 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.276000 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.276030 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.276041 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:37Z","lastTransitionTime":"2026-02-18T09:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.282418 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.282509 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.282428 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:04:37 crc kubenswrapper[4556]: E0218 09:04:37.282681 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 09:04:37 crc kubenswrapper[4556]: E0218 09:04:37.282775 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 09:04:37 crc kubenswrapper[4556]: E0218 09:04:37.282535 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.382128 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.382227 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.382315 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.382370 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.382390 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:37Z","lastTransitionTime":"2026-02-18T09:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.453615 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-b7fsj_09971522-31bc-4e10-8831-4fb62675e220/ovnkube-controller/1.log" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.459389 4556 scope.go:117] "RemoveContainer" containerID="8fe96f3b6f1f0ce811435d163297af1becac58208736b4da93517aa8780700c8" Feb 18 09:04:37 crc kubenswrapper[4556]: E0218 09:04:37.459594 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-b7fsj_openshift-ovn-kubernetes(09971522-31bc-4e10-8831-4fb62675e220)\"" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" podUID="09971522-31bc-4e10-8831-4fb62675e220" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.468197 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jflp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689d3d82-4071-4fe5-b3e3-b5b4f511580a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d56d65964aaf12feaaa3e5fa3ca779bffdb13fcebfc9f027b563c7dd01f4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bm85s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jflp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:37Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.478909 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:37Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.484974 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.485007 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.485018 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.485032 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.485042 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:37Z","lastTransitionTime":"2026-02-18T09:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.496740 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ae60f3-adab-4740-821a-1570059c7101\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f629262e27020512301a7b53ee836ee98e1e7cdee7a52b0ed506bbaaebd823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f59306c0d55e9d00714ce2ee2c1ed6f2ed33d8b21eecde490e86a93335a14f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c87bc671742a47de6c8e68aaf0c8645f9a8cae7b3d52264f36ae3827b3a8d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2678001ff826712da6716541210f018c74fd3d0218b388d2e7a6f16a1f4027\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6f37933281c927328ecfaf842d97ee8361726768f393726cb39517d2bdc9425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:37Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.506785 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8f6014-2544-4173-a94d-441cf42ba891\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77b0269625b7c6baab44a40373618f4ccfe9e1a028a1bc746b7d60c20ac4efaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac09c421b3101887bfb83107dbe26f2b5856d9b32a9cea92f6233fb6ae218cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce79cfe0242ecf403f989dc6b1c1b8be5ab54eee75ac413a8f9d03cef8bca0ed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://095ecf7e8763fa098b06eb9a8aac9f39451f7762bf20196e0803bbc841610e02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:37Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.516144 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:37Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.525554 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d653c21d3d7c875bed37ed6da700dc8c801b95c80d7e361121dfbbd3d05a922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5211d0cd5172d45b92b003e2bf97d7969a65ea448b0491f8f6ea67364925ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:37Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.535350 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6387f58543cf2601a53b0bbe9be3ec7816ba90d5ffdc4465594b37ccac0f6dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:37Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.548892 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"302eb215-a4c2-4324-836d-f37debfd620f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a99dc228d1c4cb8aeab2140ec5f37358b65684739f833fbde45822782430a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14f794063db9d33b82b79acc7134b5a2a3435e09d7cffe49d49251d1f7d20402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5eb3d3728da247903bece56bc2481547317e386e9e97b6c69926a3b55dc89b8e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b42d13c1767c29d22e52537ff45561a0fbf8f2069b2c3612e903e8c127e12485\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e865e6a50c03519ce22958ac4fcbb911fc140c330b6d7ffe5c9d7dd381e96\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T09:04:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0218 09:04:15.143085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 09:04:15.144650 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2160144045/tls.crt::/tmp/serving-cert-2160144045/tls.key\\\\\\\"\\\\nI0218 09:04:20.680799 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 09:04:20.682555 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 09:04:20.682573 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 09:04:20.682590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 09:04:20.682594 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 09:04:20.686822 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 09:04:20.686847 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686852 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686856 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 09:04:20.686859 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 09:04:20.686861 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 09:04:20.686864 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 09:04:20.688257 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 09:04:20.690726 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://974e17ae7082631df2d1bcee60c8a5f142bae2659e9d200d537a4d650f401b4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:37Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.558995 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:37Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.568763 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qfksk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e226b189-cfa9-47fc-b94d-19b5cbe0859f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc2cb92a7735d16b8e5918797e52e1e488d9261a0418ed5ebbb94c033ca9328a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2fx4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qfksk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:37Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.578116 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dac7f27-d3d1-4778-9e54-f273035a1d37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18da54fe071964d8aae8a977f5ee209c31f1c75e1041c44f8d755f58f3e39d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d41b85076e695bbc51a41fce200e0e47c099d88b4d825b4d133d33ed96237e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f76hs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:37Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.586770 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.586844 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.586859 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.586882 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.586894 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:37Z","lastTransitionTime":"2026-02-18T09:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.589589 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sgcq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9945b792-abbd-4103-a7b4-9496e1cc1b56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5dd402827182bc79283f1ea1c355b60c4ced53ffa9ce6781b0a0db3109359a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac0d9d7b9bc867b32cf6cbeb804c08e67df757e4b22914e2da69a376511cf75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac0d9d7b9bc867b32cf6cbeb804c08e67df757e4b22914e2da69a376511cf75c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b807d53d620e44d3196f68492b9f9665593adac804bc7b91569796e5594c4ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b807d53d620e44d3196f68492b9f9665593adac804bc7b91569796e5594c4ebb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebc044bbd5b0c1e725e6ed46ec2e0015b254070b7518d20815539dcf8c6e4e45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ebc044bbd5b0c1e725e6ed46ec2e0015b254070b7518d20815539dcf8c6e4e45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sgcq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:37Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.599475 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47341f31f94eaf5dd86efbe78ff62873e733b1f86ea895f281b6797770519459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:37Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.608312 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5vcgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d50c39aa-f956-441a-8abe-1d7247a7fd86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ddc5d42163d1fa86f1a59e224adac434c990224e839ef3cb12c6d7689e2b439\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4llw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5vcgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:37Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.621375 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09971522-31bc-4e10-8831-4fb62675e220\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9170bdcbaf1065995d03e6448e40bdced0741c716c9e81f37a72b015132f1161\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c526c0613f19e1a1a584b7129bd56001faf08d0b97fd77feeb69f764ed01f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cd6e758ee396c9000b1bdb6ded2176ec803fa9da6201cdd9b8146940a2c00cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b0412f3330999c3348ac726ac5a0c2d541571f4b13786ca24cb3109a44aa653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://318cb69e50df4c14df9bf5ef57d5c381421a9bdb72135c9b9620f55c0547b410\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0f15d331af9bc2ca3082c1573dd167c66773eb390593844e681f8d0b455d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fe96f3b6f1f0ce811435d163297af1becac58208736b4da93517aa8780700c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8fe96f3b6f1f0ce811435d163297af1becac58208736b4da93517aa8780700c8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T09:04:36Z\\\",\\\"message\\\":\\\":{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0218 09:04:36.174933 6015 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0218 09:04:36.174942 6015 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI0218 09:04:36.174757 6015 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-sgcq7 in node crc\\\\nI0218 09:04:36.174948 6015 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nF0218 09:04:36.174847 6015 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling w\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:35Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-b7fsj_openshift-ovn-kubernetes(09971522-31bc-4e10-8831-4fb62675e220)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aec6030df0a59ad8902d5c694d5cd30d249ca1e895155c57a458e359fa531b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7fsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:37Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.689611 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.689754 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.689822 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.689883 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.689951 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:37Z","lastTransitionTime":"2026-02-18T09:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.792804 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.792926 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.792988 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.793053 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.793116 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:37Z","lastTransitionTime":"2026-02-18T09:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.895940 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.895985 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.895996 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.896015 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.896026 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:37Z","lastTransitionTime":"2026-02-18T09:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.997928 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.997972 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.997982 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.997997 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:37 crc kubenswrapper[4556]: I0218 09:04:37.998009 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:37Z","lastTransitionTime":"2026-02-18T09:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.101282 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.101330 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.101341 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.101362 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.101371 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:38Z","lastTransitionTime":"2026-02-18T09:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.203935 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.203982 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.203991 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.204006 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.204017 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:38Z","lastTransitionTime":"2026-02-18T09:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.269087 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 07:38:09.523362643 +0000 UTC Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.306464 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.306504 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.306516 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.306531 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.306541 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:38Z","lastTransitionTime":"2026-02-18T09:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.408472 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.408514 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.408524 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.408541 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.408549 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:38Z","lastTransitionTime":"2026-02-18T09:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.509945 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.509977 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.509986 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.509998 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.510006 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:38Z","lastTransitionTime":"2026-02-18T09:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.545599 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bp8gg"] Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.546039 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bp8gg" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.548017 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.548683 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.556560 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jflp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689d3d82-4071-4fe5-b3e3-b5b4f511580a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d56d65964aaf12feaaa3e5fa3ca779bffdb13fcebfc9f027b563c7dd01f4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bm85s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jflp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:38Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.574176 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ae60f3-adab-4740-821a-1570059c7101\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f629262e27020512301a7b53ee836ee98e1e7cdee7a52b0ed506bbaaebd823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f59306c0d55e9d00714ce2ee2c1ed6f2ed33d8b21eecde490e86a93335a14f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c87bc671742a47de6c8e68aaf0c8645f9a8cae7b3d52264f36ae3827b3a8d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2678001ff826712da6716541210f018c74fd3d0218b388d2e7a6f16a1f4027\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6f37933281c927328ecfaf842d97ee8361726768f393726cb39517d2bdc9425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:38Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.585242 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8f6014-2544-4173-a94d-441cf42ba891\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77b0269625b7c6baab44a40373618f4ccfe9e1a028a1bc746b7d60c20ac4efaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac09c421b3101887bfb83107dbe26f2b5856d9b32a9cea92f6233fb6ae218cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce79cfe0242ecf403f989dc6b1c1b8be5ab54eee75ac413a8f9d03cef8bca0ed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://095ecf7e8763fa098b06eb9a8aac9f39451f7762bf20196e0803bbc841610e02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:38Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.595662 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:38Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.605240 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d653c21d3d7c875bed37ed6da700dc8c801b95c80d7e361121dfbbd3d05a922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5211d0cd5172d45b92b003e2bf97d7969a65ea448b0491f8f6ea67364925ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:38Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.612479 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.612518 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.612534 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.612599 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.612612 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:38Z","lastTransitionTime":"2026-02-18T09:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.614940 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6387f58543cf2601a53b0bbe9be3ec7816ba90d5ffdc4465594b37ccac0f6dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:38Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.624950 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:38Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.634316 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bp8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"804f8152-54ae-4486-92e6-f9595c035820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phr7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phr7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bp8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:38Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.644649 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"302eb215-a4c2-4324-836d-f37debfd620f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a99dc228d1c4cb8aeab2140ec5f37358b65684739f833fbde45822782430a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14f794063db9d33b82b79acc7134b5a2a3435e09d7cffe49d49251d1f7d20402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5eb3d3728da247903bece56bc2481547317e386e9e97b6c69926a3b55dc89b8e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b42d13c1767c29d22e52537ff45561a0fbf8f2069b2c3612e903e8c127e12485\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e865e6a50c03519ce22958ac4fcbb911fc140c330b6d7ffe5c9d7dd381e96\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T09:04:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0218 09:04:15.143085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 09:04:15.144650 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2160144045/tls.crt::/tmp/serving-cert-2160144045/tls.key\\\\\\\"\\\\nI0218 09:04:20.680799 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 09:04:20.682555 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 09:04:20.682573 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 09:04:20.682590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 09:04:20.682594 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 09:04:20.686822 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 09:04:20.686847 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686852 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686856 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 09:04:20.686859 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 09:04:20.686861 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 09:04:20.686864 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 09:04:20.688257 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 09:04:20.690726 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://974e17ae7082631df2d1bcee60c8a5f142bae2659e9d200d537a4d650f401b4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:38Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.653865 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:38Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.660026 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/804f8152-54ae-4486-92e6-f9595c035820-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-bp8gg\" (UID: \"804f8152-54ae-4486-92e6-f9595c035820\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bp8gg" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.660068 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/804f8152-54ae-4486-92e6-f9595c035820-env-overrides\") pod \"ovnkube-control-plane-749d76644c-bp8gg\" (UID: \"804f8152-54ae-4486-92e6-f9595c035820\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bp8gg" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.660126 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/804f8152-54ae-4486-92e6-f9595c035820-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-bp8gg\" (UID: \"804f8152-54ae-4486-92e6-f9595c035820\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bp8gg" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.660168 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phr7c\" (UniqueName: \"kubernetes.io/projected/804f8152-54ae-4486-92e6-f9595c035820-kube-api-access-phr7c\") pod \"ovnkube-control-plane-749d76644c-bp8gg\" (UID: \"804f8152-54ae-4486-92e6-f9595c035820\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bp8gg" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.663614 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qfksk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e226b189-cfa9-47fc-b94d-19b5cbe0859f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc2cb92a7735d16b8e5918797e52e1e488d9261a0418ed5ebbb94c033ca9328a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2fx4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qfksk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:38Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.671786 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dac7f27-d3d1-4778-9e54-f273035a1d37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18da54fe071964d8aae8a977f5ee209c31f1c75e1041c44f8d755f58f3e39d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d41b85076e695bbc51a41fce200e0e47c099d88b4d825b4d133d33ed96237e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f76hs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:38Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.684551 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sgcq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9945b792-abbd-4103-a7b4-9496e1cc1b56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5dd402827182bc79283f1ea1c355b60c4ced53ffa9ce6781b0a0db3109359a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac0d9d7b9bc867b32cf6cbeb804c08e67df757e4b22914e2da69a376511cf75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac0d9d7b9bc867b32cf6cbeb804c08e67df757e4b22914e2da69a376511cf75c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b807d53d620e44d3196f68492b9f9665593adac804bc7b91569796e5594c4ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b807d53d620e44d3196f68492b9f9665593adac804bc7b91569796e5594c4ebb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebc044bbd5b0c1e725e6ed46ec2e0015b254070b7518d20815539dcf8c6e4e45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ebc044bbd5b0c1e725e6ed46ec2e0015b254070b7518d20815539dcf8c6e4e45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sgcq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:38Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.692668 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47341f31f94eaf5dd86efbe78ff62873e733b1f86ea895f281b6797770519459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:38Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.701098 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5vcgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d50c39aa-f956-441a-8abe-1d7247a7fd86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ddc5d42163d1fa86f1a59e224adac434c990224e839ef3cb12c6d7689e2b439\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4llw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5vcgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:38Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.715479 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.715527 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.715541 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.715564 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.715590 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:38Z","lastTransitionTime":"2026-02-18T09:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.720665 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09971522-31bc-4e10-8831-4fb62675e220\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9170bdcbaf1065995d03e6448e40bdced0741c716c9e81f37a72b015132f1161\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c526c0613f19e1a1a584b7129bd56001faf08d0b97fd77feeb69f764ed01f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cd6e758ee396c9000b1bdb6ded2176ec803fa9da6201cdd9b8146940a2c00cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b0412f3330999c3348ac726ac5a0c2d541571f4b13786ca24cb3109a44aa653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://318cb69e50df4c14df9bf5ef57d5c381421a9bdb72135c9b9620f55c0547b410\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0f15d331af9bc2ca3082c1573dd167c66773eb390593844e681f8d0b455d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fe96f3b6f1f0ce811435d163297af1becac58208736b4da93517aa8780700c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8fe96f3b6f1f0ce811435d163297af1becac58208736b4da93517aa8780700c8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T09:04:36Z\\\",\\\"message\\\":\\\":{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0218 09:04:36.174933 6015 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0218 09:04:36.174942 6015 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI0218 09:04:36.174757 6015 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-sgcq7 in node crc\\\\nI0218 09:04:36.174948 6015 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nF0218 09:04:36.174847 6015 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling w\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:35Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-b7fsj_openshift-ovn-kubernetes(09971522-31bc-4e10-8831-4fb62675e220)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aec6030df0a59ad8902d5c694d5cd30d249ca1e895155c57a458e359fa531b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7fsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:38Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.761239 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/804f8152-54ae-4486-92e6-f9595c035820-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-bp8gg\" (UID: \"804f8152-54ae-4486-92e6-f9595c035820\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bp8gg" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.761284 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phr7c\" (UniqueName: \"kubernetes.io/projected/804f8152-54ae-4486-92e6-f9595c035820-kube-api-access-phr7c\") pod \"ovnkube-control-plane-749d76644c-bp8gg\" (UID: \"804f8152-54ae-4486-92e6-f9595c035820\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bp8gg" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.761346 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/804f8152-54ae-4486-92e6-f9595c035820-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-bp8gg\" (UID: \"804f8152-54ae-4486-92e6-f9595c035820\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bp8gg" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.761364 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/804f8152-54ae-4486-92e6-f9595c035820-env-overrides\") pod \"ovnkube-control-plane-749d76644c-bp8gg\" (UID: \"804f8152-54ae-4486-92e6-f9595c035820\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bp8gg" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.762027 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/804f8152-54ae-4486-92e6-f9595c035820-env-overrides\") pod \"ovnkube-control-plane-749d76644c-bp8gg\" (UID: \"804f8152-54ae-4486-92e6-f9595c035820\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bp8gg" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.762333 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/804f8152-54ae-4486-92e6-f9595c035820-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-bp8gg\" (UID: \"804f8152-54ae-4486-92e6-f9595c035820\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bp8gg" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.767732 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/804f8152-54ae-4486-92e6-f9595c035820-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-bp8gg\" (UID: \"804f8152-54ae-4486-92e6-f9595c035820\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bp8gg" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.776651 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phr7c\" (UniqueName: \"kubernetes.io/projected/804f8152-54ae-4486-92e6-f9595c035820-kube-api-access-phr7c\") pod \"ovnkube-control-plane-749d76644c-bp8gg\" (UID: \"804f8152-54ae-4486-92e6-f9595c035820\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bp8gg" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.818186 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.818219 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.818229 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.818245 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.818257 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:38Z","lastTransitionTime":"2026-02-18T09:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.857551 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bp8gg" Feb 18 09:04:38 crc kubenswrapper[4556]: W0218 09:04:38.870777 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod804f8152_54ae_4486_92e6_f9595c035820.slice/crio-99cc90f470fd1f2b8ed7eb8dff54806ed90f4d24b337e9023c2e370ba67bc998 WatchSource:0}: Error finding container 99cc90f470fd1f2b8ed7eb8dff54806ed90f4d24b337e9023c2e370ba67bc998: Status 404 returned error can't find the container with id 99cc90f470fd1f2b8ed7eb8dff54806ed90f4d24b337e9023c2e370ba67bc998 Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.920820 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.920853 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.920862 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.920879 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:38 crc kubenswrapper[4556]: I0218 09:04:38.920895 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:38Z","lastTransitionTime":"2026-02-18T09:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.023551 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.024094 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.024109 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.024165 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.024188 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:39Z","lastTransitionTime":"2026-02-18T09:04:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.128291 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.128335 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.128345 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.128363 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.128376 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:39Z","lastTransitionTime":"2026-02-18T09:04:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.230115 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.230176 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.230187 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.230202 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.230210 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:39Z","lastTransitionTime":"2026-02-18T09:04:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.269859 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-19 00:38:27.061419832 +0000 UTC Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.282336 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.282393 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.282442 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:04:39 crc kubenswrapper[4556]: E0218 09:04:39.282481 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 09:04:39 crc kubenswrapper[4556]: E0218 09:04:39.282582 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 09:04:39 crc kubenswrapper[4556]: E0218 09:04:39.282688 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.332453 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.332504 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.332519 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.332541 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.332558 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:39Z","lastTransitionTime":"2026-02-18T09:04:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.435005 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.435043 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.435055 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.435072 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.435083 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:39Z","lastTransitionTime":"2026-02-18T09:04:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.464846 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bp8gg" event={"ID":"804f8152-54ae-4486-92e6-f9595c035820","Type":"ContainerStarted","Data":"104b8ff3c41d4304d414e58303c23792a7f0ba580f73a2a8678044741af71380"} Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.464893 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bp8gg" event={"ID":"804f8152-54ae-4486-92e6-f9595c035820","Type":"ContainerStarted","Data":"cb617e46b8522d73dabef4bbc660e41b737f6e6e62f7ce42dad722490a1041dc"} Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.464905 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bp8gg" event={"ID":"804f8152-54ae-4486-92e6-f9595c035820","Type":"ContainerStarted","Data":"99cc90f470fd1f2b8ed7eb8dff54806ed90f4d24b337e9023c2e370ba67bc998"} Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.477525 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d653c21d3d7c875bed37ed6da700dc8c801b95c80d7e361121dfbbd3d05a922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5211d0cd5172d45b92b003e2bf97d7969a65ea448b0491f8f6ea67364925ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:39Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.488964 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6387f58543cf2601a53b0bbe9be3ec7816ba90d5ffdc4465594b37ccac0f6dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:39Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.498686 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:39Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.509312 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bp8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"804f8152-54ae-4486-92e6-f9595c035820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb617e46b8522d73dabef4bbc660e41b737f6e6e62f7ce42dad722490a1041dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phr7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://104b8ff3c41d4304d414e58303c23792a7f0ba580f73a2a8678044741af71380\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phr7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bp8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:39Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.524305 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ae60f3-adab-4740-821a-1570059c7101\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f629262e27020512301a7b53ee836ee98e1e7cdee7a52b0ed506bbaaebd823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f59306c0d55e9d00714ce2ee2c1ed6f2ed33d8b21eecde490e86a93335a14f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c87bc671742a47de6c8e68aaf0c8645f9a8cae7b3d52264f36ae3827b3a8d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2678001ff826712da6716541210f018c74fd3d0218b388d2e7a6f16a1f4027\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6f37933281c927328ecfaf842d97ee8361726768f393726cb39517d2bdc9425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:39Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.533595 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8f6014-2544-4173-a94d-441cf42ba891\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77b0269625b7c6baab44a40373618f4ccfe9e1a028a1bc746b7d60c20ac4efaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac09c421b3101887bfb83107dbe26f2b5856d9b32a9cea92f6233fb6ae218cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce79cfe0242ecf403f989dc6b1c1b8be5ab54eee75ac413a8f9d03cef8bca0ed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://095ecf7e8763fa098b06eb9a8aac9f39451f7762bf20196e0803bbc841610e02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:39Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.536991 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.537025 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.537036 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.537052 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.537066 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:39Z","lastTransitionTime":"2026-02-18T09:04:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.542258 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:39Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.550054 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dac7f27-d3d1-4778-9e54-f273035a1d37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18da54fe071964d8aae8a977f5ee209c31f1c75e1041c44f8d755f58f3e39d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d41b85076e695bbc51a41fce200e0e47c099d88b4d825b4d133d33ed96237e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f76hs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:39Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.559442 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sgcq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9945b792-abbd-4103-a7b4-9496e1cc1b56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5dd402827182bc79283f1ea1c355b60c4ced53ffa9ce6781b0a0db3109359a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac0d9d7b9bc867b32cf6cbeb804c08e67df757e4b22914e2da69a376511cf75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac0d9d7b9bc867b32cf6cbeb804c08e67df757e4b22914e2da69a376511cf75c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b807d53d620e44d3196f68492b9f9665593adac804bc7b91569796e5594c4ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b807d53d620e44d3196f68492b9f9665593adac804bc7b91569796e5594c4ebb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebc044bbd5b0c1e725e6ed46ec2e0015b254070b7518d20815539dcf8c6e4e45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ebc044bbd5b0c1e725e6ed46ec2e0015b254070b7518d20815539dcf8c6e4e45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sgcq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:39Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.568489 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"302eb215-a4c2-4324-836d-f37debfd620f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a99dc228d1c4cb8aeab2140ec5f37358b65684739f833fbde45822782430a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14f794063db9d33b82b79acc7134b5a2a3435e09d7cffe49d49251d1f7d20402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5eb3d3728da247903bece56bc2481547317e386e9e97b6c69926a3b55dc89b8e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b42d13c1767c29d22e52537ff45561a0fbf8f2069b2c3612e903e8c127e12485\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e865e6a50c03519ce22958ac4fcbb911fc140c330b6d7ffe5c9d7dd381e96\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T09:04:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0218 09:04:15.143085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 09:04:15.144650 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2160144045/tls.crt::/tmp/serving-cert-2160144045/tls.key\\\\\\\"\\\\nI0218 09:04:20.680799 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 09:04:20.682555 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 09:04:20.682573 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 09:04:20.682590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 09:04:20.682594 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 09:04:20.686822 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 09:04:20.686847 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686852 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686856 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 09:04:20.686859 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 09:04:20.686861 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 09:04:20.686864 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 09:04:20.688257 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 09:04:20.690726 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://974e17ae7082631df2d1bcee60c8a5f142bae2659e9d200d537a4d650f401b4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:39Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.576903 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:39Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.585170 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qfksk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e226b189-cfa9-47fc-b94d-19b5cbe0859f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc2cb92a7735d16b8e5918797e52e1e488d9261a0418ed5ebbb94c033ca9328a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2fx4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qfksk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:39Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.591907 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5vcgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d50c39aa-f956-441a-8abe-1d7247a7fd86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ddc5d42163d1fa86f1a59e224adac434c990224e839ef3cb12c6d7689e2b439\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4llw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5vcgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:39Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.604113 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09971522-31bc-4e10-8831-4fb62675e220\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9170bdcbaf1065995d03e6448e40bdced0741c716c9e81f37a72b015132f1161\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c526c0613f19e1a1a584b7129bd56001faf08d0b97fd77feeb69f764ed01f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cd6e758ee396c9000b1bdb6ded2176ec803fa9da6201cdd9b8146940a2c00cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b0412f3330999c3348ac726ac5a0c2d541571f4b13786ca24cb3109a44aa653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://318cb69e50df4c14df9bf5ef57d5c381421a9bdb72135c9b9620f55c0547b410\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0f15d331af9bc2ca3082c1573dd167c66773eb390593844e681f8d0b455d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fe96f3b6f1f0ce811435d163297af1becac58208736b4da93517aa8780700c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8fe96f3b6f1f0ce811435d163297af1becac58208736b4da93517aa8780700c8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T09:04:36Z\\\",\\\"message\\\":\\\":{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0218 09:04:36.174933 6015 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0218 09:04:36.174942 6015 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI0218 09:04:36.174757 6015 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-sgcq7 in node crc\\\\nI0218 09:04:36.174948 6015 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nF0218 09:04:36.174847 6015 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling w\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:35Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-b7fsj_openshift-ovn-kubernetes(09971522-31bc-4e10-8831-4fb62675e220)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aec6030df0a59ad8902d5c694d5cd30d249ca1e895155c57a458e359fa531b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7fsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:39Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.612213 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47341f31f94eaf5dd86efbe78ff62873e733b1f86ea895f281b6797770519459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:39Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.618787 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jflp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689d3d82-4071-4fe5-b3e3-b5b4f511580a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d56d65964aaf12feaaa3e5fa3ca779bffdb13fcebfc9f027b563c7dd01f4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bm85s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jflp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:39Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.639473 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.639507 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.639518 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.639534 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.639544 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:39Z","lastTransitionTime":"2026-02-18T09:04:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.741410 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.741453 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.741464 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.741482 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.741493 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:39Z","lastTransitionTime":"2026-02-18T09:04:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.843318 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.843361 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.843373 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.843390 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.843399 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:39Z","lastTransitionTime":"2026-02-18T09:04:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.945322 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.945364 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.945374 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.945390 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:39 crc kubenswrapper[4556]: I0218 09:04:39.945404 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:39Z","lastTransitionTime":"2026-02-18T09:04:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.047302 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.047343 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.047353 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.047369 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.047380 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:40Z","lastTransitionTime":"2026-02-18T09:04:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.149505 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.149547 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.149567 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.149583 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.149593 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:40Z","lastTransitionTime":"2026-02-18T09:04:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.251780 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.251817 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.251827 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.251841 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.251850 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:40Z","lastTransitionTime":"2026-02-18T09:04:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.270192 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-12 00:37:25.246268535 +0000 UTC Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.318673 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-cgwv9"] Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.319134 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cgwv9" Feb 18 09:04:40 crc kubenswrapper[4556]: E0218 09:04:40.319222 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cgwv9" podUID="d2f0e36c-cb29-4b06-bb22-6afd59466cab" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.327614 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5vcgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d50c39aa-f956-441a-8abe-1d7247a7fd86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ddc5d42163d1fa86f1a59e224adac434c990224e839ef3cb12c6d7689e2b439\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4llw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5vcgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:40Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.340608 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09971522-31bc-4e10-8831-4fb62675e220\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9170bdcbaf1065995d03e6448e40bdced0741c716c9e81f37a72b015132f1161\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c526c0613f19e1a1a584b7129bd56001faf08d0b97fd77feeb69f764ed01f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cd6e758ee396c9000b1bdb6ded2176ec803fa9da6201cdd9b8146940a2c00cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b0412f3330999c3348ac726ac5a0c2d541571f4b13786ca24cb3109a44aa653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://318cb69e50df4c14df9bf5ef57d5c381421a9bdb72135c9b9620f55c0547b410\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0f15d331af9bc2ca3082c1573dd167c66773eb390593844e681f8d0b455d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fe96f3b6f1f0ce811435d163297af1becac58208736b4da93517aa8780700c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8fe96f3b6f1f0ce811435d163297af1becac58208736b4da93517aa8780700c8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T09:04:36Z\\\",\\\"message\\\":\\\":{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0218 09:04:36.174933 6015 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0218 09:04:36.174942 6015 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI0218 09:04:36.174757 6015 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-sgcq7 in node crc\\\\nI0218 09:04:36.174948 6015 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nF0218 09:04:36.174847 6015 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling w\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:35Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-b7fsj_openshift-ovn-kubernetes(09971522-31bc-4e10-8831-4fb62675e220)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aec6030df0a59ad8902d5c694d5cd30d249ca1e895155c57a458e359fa531b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7fsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:40Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.349142 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47341f31f94eaf5dd86efbe78ff62873e733b1f86ea895f281b6797770519459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:40Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.353495 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.353521 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.353530 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.353543 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.353551 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:40Z","lastTransitionTime":"2026-02-18T09:04:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.356417 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jflp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689d3d82-4071-4fe5-b3e3-b5b4f511580a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d56d65964aaf12feaaa3e5fa3ca779bffdb13fcebfc9f027b563c7dd01f4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bm85s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jflp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:40Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.364917 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d653c21d3d7c875bed37ed6da700dc8c801b95c80d7e361121dfbbd3d05a922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5211d0cd5172d45b92b003e2bf97d7969a65ea448b0491f8f6ea67364925ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:40Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.372987 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6387f58543cf2601a53b0bbe9be3ec7816ba90d5ffdc4465594b37ccac0f6dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:40Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.388354 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:40Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.396522 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bp8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"804f8152-54ae-4486-92e6-f9595c035820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb617e46b8522d73dabef4bbc660e41b737f6e6e62f7ce42dad722490a1041dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phr7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://104b8ff3c41d4304d414e58303c23792a7f0ba580f73a2a8678044741af71380\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phr7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bp8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:40Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.408813 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ae60f3-adab-4740-821a-1570059c7101\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f629262e27020512301a7b53ee836ee98e1e7cdee7a52b0ed506bbaaebd823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f59306c0d55e9d00714ce2ee2c1ed6f2ed33d8b21eecde490e86a93335a14f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c87bc671742a47de6c8e68aaf0c8645f9a8cae7b3d52264f36ae3827b3a8d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2678001ff826712da6716541210f018c74fd3d0218b388d2e7a6f16a1f4027\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6f37933281c927328ecfaf842d97ee8361726768f393726cb39517d2bdc9425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:40Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.417124 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8f6014-2544-4173-a94d-441cf42ba891\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77b0269625b7c6baab44a40373618f4ccfe9e1a028a1bc746b7d60c20ac4efaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac09c421b3101887bfb83107dbe26f2b5856d9b32a9cea92f6233fb6ae218cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce79cfe0242ecf403f989dc6b1c1b8be5ab54eee75ac413a8f9d03cef8bca0ed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://095ecf7e8763fa098b06eb9a8aac9f39451f7762bf20196e0803bbc841610e02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:40Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.424927 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:40Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.432361 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dac7f27-d3d1-4778-9e54-f273035a1d37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18da54fe071964d8aae8a977f5ee209c31f1c75e1041c44f8d755f58f3e39d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d41b85076e695bbc51a41fce200e0e47c099d88b4d825b4d133d33ed96237e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f76hs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:40Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.454892 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sgcq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9945b792-abbd-4103-a7b4-9496e1cc1b56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5dd402827182bc79283f1ea1c355b60c4ced53ffa9ce6781b0a0db3109359a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac0d9d7b9bc867b32cf6cbeb804c08e67df757e4b22914e2da69a376511cf75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac0d9d7b9bc867b32cf6cbeb804c08e67df757e4b22914e2da69a376511cf75c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b807d53d620e44d3196f68492b9f9665593adac804bc7b91569796e5594c4ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b807d53d620e44d3196f68492b9f9665593adac804bc7b91569796e5594c4ebb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebc044bbd5b0c1e725e6ed46ec2e0015b254070b7518d20815539dcf8c6e4e45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ebc044bbd5b0c1e725e6ed46ec2e0015b254070b7518d20815539dcf8c6e4e45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sgcq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:40Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.455301 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.455337 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.455347 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.455360 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.455369 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:40Z","lastTransitionTime":"2026-02-18T09:04:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.470008 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cgwv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2f0e36c-cb29-4b06-bb22-6afd59466cab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpldv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpldv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:40Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cgwv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:40Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.478814 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d2f0e36c-cb29-4b06-bb22-6afd59466cab-metrics-certs\") pod \"network-metrics-daemon-cgwv9\" (UID: \"d2f0e36c-cb29-4b06-bb22-6afd59466cab\") " pod="openshift-multus/network-metrics-daemon-cgwv9" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.478867 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpldv\" (UniqueName: \"kubernetes.io/projected/d2f0e36c-cb29-4b06-bb22-6afd59466cab-kube-api-access-wpldv\") pod \"network-metrics-daemon-cgwv9\" (UID: \"d2f0e36c-cb29-4b06-bb22-6afd59466cab\") " pod="openshift-multus/network-metrics-daemon-cgwv9" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.486878 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"302eb215-a4c2-4324-836d-f37debfd620f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a99dc228d1c4cb8aeab2140ec5f37358b65684739f833fbde45822782430a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14f794063db9d33b82b79acc7134b5a2a3435e09d7cffe49d49251d1f7d20402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5eb3d3728da247903bece56bc2481547317e386e9e97b6c69926a3b55dc89b8e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b42d13c1767c29d22e52537ff45561a0fbf8f2069b2c3612e903e8c127e12485\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e865e6a50c03519ce22958ac4fcbb911fc140c330b6d7ffe5c9d7dd381e96\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T09:04:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0218 09:04:15.143085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 09:04:15.144650 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2160144045/tls.crt::/tmp/serving-cert-2160144045/tls.key\\\\\\\"\\\\nI0218 09:04:20.680799 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 09:04:20.682555 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 09:04:20.682573 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 09:04:20.682590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 09:04:20.682594 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 09:04:20.686822 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 09:04:20.686847 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686852 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686856 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 09:04:20.686859 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 09:04:20.686861 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 09:04:20.686864 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 09:04:20.688257 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 09:04:20.690726 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://974e17ae7082631df2d1bcee60c8a5f142bae2659e9d200d537a4d650f401b4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:40Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.499793 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:40Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.511142 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qfksk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e226b189-cfa9-47fc-b94d-19b5cbe0859f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc2cb92a7735d16b8e5918797e52e1e488d9261a0418ed5ebbb94c033ca9328a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2fx4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qfksk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:40Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.557765 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.557804 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.557813 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.557827 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.557836 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:40Z","lastTransitionTime":"2026-02-18T09:04:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.580442 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpldv\" (UniqueName: \"kubernetes.io/projected/d2f0e36c-cb29-4b06-bb22-6afd59466cab-kube-api-access-wpldv\") pod \"network-metrics-daemon-cgwv9\" (UID: \"d2f0e36c-cb29-4b06-bb22-6afd59466cab\") " pod="openshift-multus/network-metrics-daemon-cgwv9" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.580528 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d2f0e36c-cb29-4b06-bb22-6afd59466cab-metrics-certs\") pod \"network-metrics-daemon-cgwv9\" (UID: \"d2f0e36c-cb29-4b06-bb22-6afd59466cab\") " pod="openshift-multus/network-metrics-daemon-cgwv9" Feb 18 09:04:40 crc kubenswrapper[4556]: E0218 09:04:40.580675 4556 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 18 09:04:40 crc kubenswrapper[4556]: E0218 09:04:40.580745 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d2f0e36c-cb29-4b06-bb22-6afd59466cab-metrics-certs podName:d2f0e36c-cb29-4b06-bb22-6afd59466cab nodeName:}" failed. No retries permitted until 2026-02-18 09:04:41.080729019 +0000 UTC m=+38.097689999 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d2f0e36c-cb29-4b06-bb22-6afd59466cab-metrics-certs") pod "network-metrics-daemon-cgwv9" (UID: "d2f0e36c-cb29-4b06-bb22-6afd59466cab") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.595384 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpldv\" (UniqueName: \"kubernetes.io/projected/d2f0e36c-cb29-4b06-bb22-6afd59466cab-kube-api-access-wpldv\") pod \"network-metrics-daemon-cgwv9\" (UID: \"d2f0e36c-cb29-4b06-bb22-6afd59466cab\") " pod="openshift-multus/network-metrics-daemon-cgwv9" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.660599 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.660656 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.660666 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.660688 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.660705 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:40Z","lastTransitionTime":"2026-02-18T09:04:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.763315 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.763355 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.763365 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.763382 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.763398 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:40Z","lastTransitionTime":"2026-02-18T09:04:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.865436 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.865496 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.865508 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.865528 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.865541 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:40Z","lastTransitionTime":"2026-02-18T09:04:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.967640 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.967686 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.967697 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.967709 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:40 crc kubenswrapper[4556]: I0218 09:04:40.967719 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:40Z","lastTransitionTime":"2026-02-18T09:04:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:41 crc kubenswrapper[4556]: I0218 09:04:41.070020 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:41 crc kubenswrapper[4556]: I0218 09:04:41.070055 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:41 crc kubenswrapper[4556]: I0218 09:04:41.070064 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:41 crc kubenswrapper[4556]: I0218 09:04:41.070098 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:41 crc kubenswrapper[4556]: I0218 09:04:41.070109 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:41Z","lastTransitionTime":"2026-02-18T09:04:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:41 crc kubenswrapper[4556]: I0218 09:04:41.085595 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d2f0e36c-cb29-4b06-bb22-6afd59466cab-metrics-certs\") pod \"network-metrics-daemon-cgwv9\" (UID: \"d2f0e36c-cb29-4b06-bb22-6afd59466cab\") " pod="openshift-multus/network-metrics-daemon-cgwv9" Feb 18 09:04:41 crc kubenswrapper[4556]: E0218 09:04:41.085737 4556 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 18 09:04:41 crc kubenswrapper[4556]: E0218 09:04:41.085798 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d2f0e36c-cb29-4b06-bb22-6afd59466cab-metrics-certs podName:d2f0e36c-cb29-4b06-bb22-6afd59466cab nodeName:}" failed. No retries permitted until 2026-02-18 09:04:42.085783904 +0000 UTC m=+39.102744884 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d2f0e36c-cb29-4b06-bb22-6afd59466cab-metrics-certs") pod "network-metrics-daemon-cgwv9" (UID: "d2f0e36c-cb29-4b06-bb22-6afd59466cab") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 18 09:04:41 crc kubenswrapper[4556]: I0218 09:04:41.172905 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:41 crc kubenswrapper[4556]: I0218 09:04:41.173037 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:41 crc kubenswrapper[4556]: I0218 09:04:41.173103 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:41 crc kubenswrapper[4556]: I0218 09:04:41.173189 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:41 crc kubenswrapper[4556]: I0218 09:04:41.173258 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:41Z","lastTransitionTime":"2026-02-18T09:04:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:41 crc kubenswrapper[4556]: I0218 09:04:41.271264 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-09 23:41:08.729086265 +0000 UTC Feb 18 09:04:41 crc kubenswrapper[4556]: I0218 09:04:41.276043 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:41 crc kubenswrapper[4556]: I0218 09:04:41.276089 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:41 crc kubenswrapper[4556]: I0218 09:04:41.276099 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:41 crc kubenswrapper[4556]: I0218 09:04:41.276115 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:41 crc kubenswrapper[4556]: I0218 09:04:41.276129 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:41Z","lastTransitionTime":"2026-02-18T09:04:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:41 crc kubenswrapper[4556]: I0218 09:04:41.281786 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:04:41 crc kubenswrapper[4556]: I0218 09:04:41.281826 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:04:41 crc kubenswrapper[4556]: I0218 09:04:41.281864 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:04:41 crc kubenswrapper[4556]: E0218 09:04:41.282235 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 09:04:41 crc kubenswrapper[4556]: E0218 09:04:41.281960 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 09:04:41 crc kubenswrapper[4556]: E0218 09:04:41.282399 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 09:04:41 crc kubenswrapper[4556]: I0218 09:04:41.378518 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:41 crc kubenswrapper[4556]: I0218 09:04:41.378570 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:41 crc kubenswrapper[4556]: I0218 09:04:41.378586 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:41 crc kubenswrapper[4556]: I0218 09:04:41.378607 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:41 crc kubenswrapper[4556]: I0218 09:04:41.378621 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:41Z","lastTransitionTime":"2026-02-18T09:04:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:41 crc kubenswrapper[4556]: I0218 09:04:41.480952 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:41 crc kubenswrapper[4556]: I0218 09:04:41.480984 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:41 crc kubenswrapper[4556]: I0218 09:04:41.480993 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:41 crc kubenswrapper[4556]: I0218 09:04:41.481004 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:41 crc kubenswrapper[4556]: I0218 09:04:41.481015 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:41Z","lastTransitionTime":"2026-02-18T09:04:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:41 crc kubenswrapper[4556]: I0218 09:04:41.583219 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:41 crc kubenswrapper[4556]: I0218 09:04:41.583248 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:41 crc kubenswrapper[4556]: I0218 09:04:41.583262 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:41 crc kubenswrapper[4556]: I0218 09:04:41.583274 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:41 crc kubenswrapper[4556]: I0218 09:04:41.583283 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:41Z","lastTransitionTime":"2026-02-18T09:04:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:41 crc kubenswrapper[4556]: I0218 09:04:41.685388 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:41 crc kubenswrapper[4556]: I0218 09:04:41.685432 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:41 crc kubenswrapper[4556]: I0218 09:04:41.685445 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:41 crc kubenswrapper[4556]: I0218 09:04:41.685462 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:41 crc kubenswrapper[4556]: I0218 09:04:41.685474 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:41Z","lastTransitionTime":"2026-02-18T09:04:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:41 crc kubenswrapper[4556]: I0218 09:04:41.787699 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:41 crc kubenswrapper[4556]: I0218 09:04:41.787737 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:41 crc kubenswrapper[4556]: I0218 09:04:41.787747 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:41 crc kubenswrapper[4556]: I0218 09:04:41.787763 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:41 crc kubenswrapper[4556]: I0218 09:04:41.787775 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:41Z","lastTransitionTime":"2026-02-18T09:04:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:41 crc kubenswrapper[4556]: I0218 09:04:41.890483 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:41 crc kubenswrapper[4556]: I0218 09:04:41.890559 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:41 crc kubenswrapper[4556]: I0218 09:04:41.890577 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:41 crc kubenswrapper[4556]: I0218 09:04:41.890601 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:41 crc kubenswrapper[4556]: I0218 09:04:41.890616 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:41Z","lastTransitionTime":"2026-02-18T09:04:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:41 crc kubenswrapper[4556]: I0218 09:04:41.993142 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:41 crc kubenswrapper[4556]: I0218 09:04:41.993209 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:41 crc kubenswrapper[4556]: I0218 09:04:41.993218 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:41 crc kubenswrapper[4556]: I0218 09:04:41.993238 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:41 crc kubenswrapper[4556]: I0218 09:04:41.993250 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:41Z","lastTransitionTime":"2026-02-18T09:04:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:42 crc kubenswrapper[4556]: I0218 09:04:42.092588 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d2f0e36c-cb29-4b06-bb22-6afd59466cab-metrics-certs\") pod \"network-metrics-daemon-cgwv9\" (UID: \"d2f0e36c-cb29-4b06-bb22-6afd59466cab\") " pod="openshift-multus/network-metrics-daemon-cgwv9" Feb 18 09:04:42 crc kubenswrapper[4556]: E0218 09:04:42.092874 4556 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 18 09:04:42 crc kubenswrapper[4556]: E0218 09:04:42.093036 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d2f0e36c-cb29-4b06-bb22-6afd59466cab-metrics-certs podName:d2f0e36c-cb29-4b06-bb22-6afd59466cab nodeName:}" failed. No retries permitted until 2026-02-18 09:04:44.093006932 +0000 UTC m=+41.109967911 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d2f0e36c-cb29-4b06-bb22-6afd59466cab-metrics-certs") pod "network-metrics-daemon-cgwv9" (UID: "d2f0e36c-cb29-4b06-bb22-6afd59466cab") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 18 09:04:42 crc kubenswrapper[4556]: I0218 09:04:42.094801 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:42 crc kubenswrapper[4556]: I0218 09:04:42.094841 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:42 crc kubenswrapper[4556]: I0218 09:04:42.094853 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:42 crc kubenswrapper[4556]: I0218 09:04:42.094875 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:42 crc kubenswrapper[4556]: I0218 09:04:42.094898 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:42Z","lastTransitionTime":"2026-02-18T09:04:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:42 crc kubenswrapper[4556]: I0218 09:04:42.197400 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:42 crc kubenswrapper[4556]: I0218 09:04:42.197444 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:42 crc kubenswrapper[4556]: I0218 09:04:42.197455 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:42 crc kubenswrapper[4556]: I0218 09:04:42.197476 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:42 crc kubenswrapper[4556]: I0218 09:04:42.197487 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:42Z","lastTransitionTime":"2026-02-18T09:04:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:42 crc kubenswrapper[4556]: I0218 09:04:42.272142 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 22:59:42.624808582 +0000 UTC Feb 18 09:04:42 crc kubenswrapper[4556]: I0218 09:04:42.281583 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cgwv9" Feb 18 09:04:42 crc kubenswrapper[4556]: E0218 09:04:42.281816 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cgwv9" podUID="d2f0e36c-cb29-4b06-bb22-6afd59466cab" Feb 18 09:04:42 crc kubenswrapper[4556]: I0218 09:04:42.299238 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:42 crc kubenswrapper[4556]: I0218 09:04:42.299288 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:42 crc kubenswrapper[4556]: I0218 09:04:42.299300 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:42 crc kubenswrapper[4556]: I0218 09:04:42.299320 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:42 crc kubenswrapper[4556]: I0218 09:04:42.299337 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:42Z","lastTransitionTime":"2026-02-18T09:04:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:42 crc kubenswrapper[4556]: I0218 09:04:42.401848 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:42 crc kubenswrapper[4556]: I0218 09:04:42.401897 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:42 crc kubenswrapper[4556]: I0218 09:04:42.401908 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:42 crc kubenswrapper[4556]: I0218 09:04:42.401924 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:42 crc kubenswrapper[4556]: I0218 09:04:42.401938 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:42Z","lastTransitionTime":"2026-02-18T09:04:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:42 crc kubenswrapper[4556]: I0218 09:04:42.504257 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:42 crc kubenswrapper[4556]: I0218 09:04:42.504292 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:42 crc kubenswrapper[4556]: I0218 09:04:42.504300 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:42 crc kubenswrapper[4556]: I0218 09:04:42.504315 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:42 crc kubenswrapper[4556]: I0218 09:04:42.504327 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:42Z","lastTransitionTime":"2026-02-18T09:04:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:42 crc kubenswrapper[4556]: I0218 09:04:42.606276 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:42 crc kubenswrapper[4556]: I0218 09:04:42.606332 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:42 crc kubenswrapper[4556]: I0218 09:04:42.606345 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:42 crc kubenswrapper[4556]: I0218 09:04:42.606370 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:42 crc kubenswrapper[4556]: I0218 09:04:42.606385 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:42Z","lastTransitionTime":"2026-02-18T09:04:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:42 crc kubenswrapper[4556]: I0218 09:04:42.708641 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:42 crc kubenswrapper[4556]: I0218 09:04:42.708677 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:42 crc kubenswrapper[4556]: I0218 09:04:42.708720 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:42 crc kubenswrapper[4556]: I0218 09:04:42.708734 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:42 crc kubenswrapper[4556]: I0218 09:04:42.708747 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:42Z","lastTransitionTime":"2026-02-18T09:04:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:42 crc kubenswrapper[4556]: I0218 09:04:42.810507 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:42 crc kubenswrapper[4556]: I0218 09:04:42.810563 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:42 crc kubenswrapper[4556]: I0218 09:04:42.810574 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:42 crc kubenswrapper[4556]: I0218 09:04:42.810590 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:42 crc kubenswrapper[4556]: I0218 09:04:42.810604 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:42Z","lastTransitionTime":"2026-02-18T09:04:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:42 crc kubenswrapper[4556]: I0218 09:04:42.915132 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:42 crc kubenswrapper[4556]: I0218 09:04:42.915199 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:42 crc kubenswrapper[4556]: I0218 09:04:42.915212 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:42 crc kubenswrapper[4556]: I0218 09:04:42.915228 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:42 crc kubenswrapper[4556]: I0218 09:04:42.915241 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:42Z","lastTransitionTime":"2026-02-18T09:04:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.017461 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.017511 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.017523 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.017545 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.017557 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:43Z","lastTransitionTime":"2026-02-18T09:04:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.119660 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.119693 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.119703 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.119732 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.119930 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:43Z","lastTransitionTime":"2026-02-18T09:04:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.222566 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.222588 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.222596 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.222608 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.222616 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:43Z","lastTransitionTime":"2026-02-18T09:04:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.272237 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-17 19:22:33.880470307 +0000 UTC Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.281685 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.281785 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.281746 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:04:43 crc kubenswrapper[4556]: E0218 09:04:43.281971 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 09:04:43 crc kubenswrapper[4556]: E0218 09:04:43.282048 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 09:04:43 crc kubenswrapper[4556]: E0218 09:04:43.282109 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.293267 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47341f31f94eaf5dd86efbe78ff62873e733b1f86ea895f281b6797770519459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:43Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.301184 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5vcgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d50c39aa-f956-441a-8abe-1d7247a7fd86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ddc5d42163d1fa86f1a59e224adac434c990224e839ef3cb12c6d7689e2b439\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4llw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5vcgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:43Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.314571 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09971522-31bc-4e10-8831-4fb62675e220\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9170bdcbaf1065995d03e6448e40bdced0741c716c9e81f37a72b015132f1161\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c526c0613f19e1a1a584b7129bd56001faf08d0b97fd77feeb69f764ed01f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cd6e758ee396c9000b1bdb6ded2176ec803fa9da6201cdd9b8146940a2c00cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b0412f3330999c3348ac726ac5a0c2d541571f4b13786ca24cb3109a44aa653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://318cb69e50df4c14df9bf5ef57d5c381421a9bdb72135c9b9620f55c0547b410\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0f15d331af9bc2ca3082c1573dd167c66773eb390593844e681f8d0b455d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fe96f3b6f1f0ce811435d163297af1becac58208736b4da93517aa8780700c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8fe96f3b6f1f0ce811435d163297af1becac58208736b4da93517aa8780700c8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T09:04:36Z\\\",\\\"message\\\":\\\":{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0218 09:04:36.174933 6015 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0218 09:04:36.174942 6015 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI0218 09:04:36.174757 6015 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-sgcq7 in node crc\\\\nI0218 09:04:36.174948 6015 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nF0218 09:04:36.174847 6015 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling w\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:35Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-b7fsj_openshift-ovn-kubernetes(09971522-31bc-4e10-8831-4fb62675e220)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aec6030df0a59ad8902d5c694d5cd30d249ca1e895155c57a458e359fa531b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7fsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:43Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.323492 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jflp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689d3d82-4071-4fe5-b3e3-b5b4f511580a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d56d65964aaf12feaaa3e5fa3ca779bffdb13fcebfc9f027b563c7dd01f4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bm85s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jflp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:43Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.324686 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.324723 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.324743 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.324758 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.324769 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:43Z","lastTransitionTime":"2026-02-18T09:04:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.341888 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ae60f3-adab-4740-821a-1570059c7101\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f629262e27020512301a7b53ee836ee98e1e7cdee7a52b0ed506bbaaebd823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f59306c0d55e9d00714ce2ee2c1ed6f2ed33d8b21eecde490e86a93335a14f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c87bc671742a47de6c8e68aaf0c8645f9a8cae7b3d52264f36ae3827b3a8d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2678001ff826712da6716541210f018c74fd3d0218b388d2e7a6f16a1f4027\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6f37933281c927328ecfaf842d97ee8361726768f393726cb39517d2bdc9425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:43Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.352245 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8f6014-2544-4173-a94d-441cf42ba891\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77b0269625b7c6baab44a40373618f4ccfe9e1a028a1bc746b7d60c20ac4efaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac09c421b3101887bfb83107dbe26f2b5856d9b32a9cea92f6233fb6ae218cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce79cfe0242ecf403f989dc6b1c1b8be5ab54eee75ac413a8f9d03cef8bca0ed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://095ecf7e8763fa098b06eb9a8aac9f39451f7762bf20196e0803bbc841610e02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:43Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.361922 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:43Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.370969 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d653c21d3d7c875bed37ed6da700dc8c801b95c80d7e361121dfbbd3d05a922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5211d0cd5172d45b92b003e2bf97d7969a65ea448b0491f8f6ea67364925ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:43Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.381076 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6387f58543cf2601a53b0bbe9be3ec7816ba90d5ffdc4465594b37ccac0f6dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:43Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.390238 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:43Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.403683 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bp8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"804f8152-54ae-4486-92e6-f9595c035820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb617e46b8522d73dabef4bbc660e41b737f6e6e62f7ce42dad722490a1041dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phr7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://104b8ff3c41d4304d414e58303c23792a7f0ba580f73a2a8678044741af71380\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phr7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bp8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:43Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.416370 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"302eb215-a4c2-4324-836d-f37debfd620f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a99dc228d1c4cb8aeab2140ec5f37358b65684739f833fbde45822782430a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14f794063db9d33b82b79acc7134b5a2a3435e09d7cffe49d49251d1f7d20402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5eb3d3728da247903bece56bc2481547317e386e9e97b6c69926a3b55dc89b8e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b42d13c1767c29d22e52537ff45561a0fbf8f2069b2c3612e903e8c127e12485\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e865e6a50c03519ce22958ac4fcbb911fc140c330b6d7ffe5c9d7dd381e96\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T09:04:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0218 09:04:15.143085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 09:04:15.144650 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2160144045/tls.crt::/tmp/serving-cert-2160144045/tls.key\\\\\\\"\\\\nI0218 09:04:20.680799 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 09:04:20.682555 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 09:04:20.682573 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 09:04:20.682590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 09:04:20.682594 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 09:04:20.686822 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 09:04:20.686847 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686852 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686856 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 09:04:20.686859 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 09:04:20.686861 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 09:04:20.686864 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 09:04:20.688257 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 09:04:20.690726 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://974e17ae7082631df2d1bcee60c8a5f142bae2659e9d200d537a4d650f401b4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:43Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.425969 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:43Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.426396 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.426501 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.426596 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.426660 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.426728 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:43Z","lastTransitionTime":"2026-02-18T09:04:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.437405 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qfksk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e226b189-cfa9-47fc-b94d-19b5cbe0859f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc2cb92a7735d16b8e5918797e52e1e488d9261a0418ed5ebbb94c033ca9328a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2fx4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qfksk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:43Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.446755 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dac7f27-d3d1-4778-9e54-f273035a1d37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18da54fe071964d8aae8a977f5ee209c31f1c75e1041c44f8d755f58f3e39d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d41b85076e695bbc51a41fce200e0e47c099d88b4d825b4d133d33ed96237e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f76hs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:43Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.457413 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sgcq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9945b792-abbd-4103-a7b4-9496e1cc1b56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5dd402827182bc79283f1ea1c355b60c4ced53ffa9ce6781b0a0db3109359a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac0d9d7b9bc867b32cf6cbeb804c08e67df757e4b22914e2da69a376511cf75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac0d9d7b9bc867b32cf6cbeb804c08e67df757e4b22914e2da69a376511cf75c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b807d53d620e44d3196f68492b9f9665593adac804bc7b91569796e5594c4ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b807d53d620e44d3196f68492b9f9665593adac804bc7b91569796e5594c4ebb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebc044bbd5b0c1e725e6ed46ec2e0015b254070b7518d20815539dcf8c6e4e45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ebc044bbd5b0c1e725e6ed46ec2e0015b254070b7518d20815539dcf8c6e4e45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sgcq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:43Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.464731 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cgwv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2f0e36c-cb29-4b06-bb22-6afd59466cab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpldv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpldv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:40Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cgwv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:43Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.528838 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.528875 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.528885 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.528899 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.528911 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:43Z","lastTransitionTime":"2026-02-18T09:04:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.630723 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.630751 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.630762 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.630776 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.630786 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:43Z","lastTransitionTime":"2026-02-18T09:04:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.732838 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.732868 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.732878 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.732892 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.732902 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:43Z","lastTransitionTime":"2026-02-18T09:04:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.834282 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.834325 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.834335 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.834349 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.834362 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:43Z","lastTransitionTime":"2026-02-18T09:04:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.935995 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.936099 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.936176 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.936257 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:43 crc kubenswrapper[4556]: I0218 09:04:43.936329 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:43Z","lastTransitionTime":"2026-02-18T09:04:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.038227 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.038247 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.038255 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.038264 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.038271 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:44Z","lastTransitionTime":"2026-02-18T09:04:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.108773 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d2f0e36c-cb29-4b06-bb22-6afd59466cab-metrics-certs\") pod \"network-metrics-daemon-cgwv9\" (UID: \"d2f0e36c-cb29-4b06-bb22-6afd59466cab\") " pod="openshift-multus/network-metrics-daemon-cgwv9" Feb 18 09:04:44 crc kubenswrapper[4556]: E0218 09:04:44.108881 4556 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 18 09:04:44 crc kubenswrapper[4556]: E0218 09:04:44.108927 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d2f0e36c-cb29-4b06-bb22-6afd59466cab-metrics-certs podName:d2f0e36c-cb29-4b06-bb22-6afd59466cab nodeName:}" failed. No retries permitted until 2026-02-18 09:04:48.108913279 +0000 UTC m=+45.125874260 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d2f0e36c-cb29-4b06-bb22-6afd59466cab-metrics-certs") pod "network-metrics-daemon-cgwv9" (UID: "d2f0e36c-cb29-4b06-bb22-6afd59466cab") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.140138 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.140211 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.140223 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.140242 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.140258 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:44Z","lastTransitionTime":"2026-02-18T09:04:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.184126 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.184179 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.184191 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.184204 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.184215 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:44Z","lastTransitionTime":"2026-02-18T09:04:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:44 crc kubenswrapper[4556]: E0218 09:04:44.194408 4556 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8758c02e-ef59-46a7-908c-d5c97feb8ceb\\\",\\\"systemUUID\\\":\\\"0f904183-79bf-4790-8551-43a6eb3adbe4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:44Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.197373 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.197403 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.197412 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.197423 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.197437 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:44Z","lastTransitionTime":"2026-02-18T09:04:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:44 crc kubenswrapper[4556]: E0218 09:04:44.205818 4556 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8758c02e-ef59-46a7-908c-d5c97feb8ceb\\\",\\\"systemUUID\\\":\\\"0f904183-79bf-4790-8551-43a6eb3adbe4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:44Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.209326 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.209351 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.209359 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.209368 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.209375 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:44Z","lastTransitionTime":"2026-02-18T09:04:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:44 crc kubenswrapper[4556]: E0218 09:04:44.219577 4556 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8758c02e-ef59-46a7-908c-d5c97feb8ceb\\\",\\\"systemUUID\\\":\\\"0f904183-79bf-4790-8551-43a6eb3adbe4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:44Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.224579 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.224604 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.224615 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.224625 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.224633 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:44Z","lastTransitionTime":"2026-02-18T09:04:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:44 crc kubenswrapper[4556]: E0218 09:04:44.233779 4556 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8758c02e-ef59-46a7-908c-d5c97feb8ceb\\\",\\\"systemUUID\\\":\\\"0f904183-79bf-4790-8551-43a6eb3adbe4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:44Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.239598 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.239636 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.239646 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.239660 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.239670 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:44Z","lastTransitionTime":"2026-02-18T09:04:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:44 crc kubenswrapper[4556]: E0218 09:04:44.249364 4556 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8758c02e-ef59-46a7-908c-d5c97feb8ceb\\\",\\\"systemUUID\\\":\\\"0f904183-79bf-4790-8551-43a6eb3adbe4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:44Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:44 crc kubenswrapper[4556]: E0218 09:04:44.249474 4556 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.251006 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.251029 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.251039 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.251051 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.251060 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:44Z","lastTransitionTime":"2026-02-18T09:04:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.272522 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-25 05:12:46.322981374 +0000 UTC Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.281924 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cgwv9" Feb 18 09:04:44 crc kubenswrapper[4556]: E0218 09:04:44.282054 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cgwv9" podUID="d2f0e36c-cb29-4b06-bb22-6afd59466cab" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.352999 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.353034 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.353044 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.353062 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.353074 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:44Z","lastTransitionTime":"2026-02-18T09:04:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.455496 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.455546 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.455556 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.455568 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.455577 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:44Z","lastTransitionTime":"2026-02-18T09:04:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.557905 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.557944 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.557953 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.557967 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.557978 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:44Z","lastTransitionTime":"2026-02-18T09:04:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.659822 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.659853 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.659862 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.659872 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.659881 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:44Z","lastTransitionTime":"2026-02-18T09:04:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.761958 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.762004 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.762013 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.762027 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.762037 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:44Z","lastTransitionTime":"2026-02-18T09:04:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.863471 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.863517 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.863538 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.863555 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.863569 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:44Z","lastTransitionTime":"2026-02-18T09:04:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.965888 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.965914 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.965922 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.965933 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:44 crc kubenswrapper[4556]: I0218 09:04:44.965942 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:44Z","lastTransitionTime":"2026-02-18T09:04:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:45 crc kubenswrapper[4556]: I0218 09:04:45.067364 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:45 crc kubenswrapper[4556]: I0218 09:04:45.067411 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:45 crc kubenswrapper[4556]: I0218 09:04:45.067421 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:45 crc kubenswrapper[4556]: I0218 09:04:45.067436 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:45 crc kubenswrapper[4556]: I0218 09:04:45.067448 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:45Z","lastTransitionTime":"2026-02-18T09:04:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:45 crc kubenswrapper[4556]: I0218 09:04:45.169107 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:45 crc kubenswrapper[4556]: I0218 09:04:45.169165 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:45 crc kubenswrapper[4556]: I0218 09:04:45.169177 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:45 crc kubenswrapper[4556]: I0218 09:04:45.169195 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:45 crc kubenswrapper[4556]: I0218 09:04:45.169215 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:45Z","lastTransitionTime":"2026-02-18T09:04:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:45 crc kubenswrapper[4556]: I0218 09:04:45.271527 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:45 crc kubenswrapper[4556]: I0218 09:04:45.271573 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:45 crc kubenswrapper[4556]: I0218 09:04:45.271587 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:45 crc kubenswrapper[4556]: I0218 09:04:45.271597 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:45 crc kubenswrapper[4556]: I0218 09:04:45.271606 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:45Z","lastTransitionTime":"2026-02-18T09:04:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:45 crc kubenswrapper[4556]: I0218 09:04:45.272750 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-15 21:47:43.674452264 +0000 UTC Feb 18 09:04:45 crc kubenswrapper[4556]: I0218 09:04:45.282004 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:04:45 crc kubenswrapper[4556]: I0218 09:04:45.282133 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:04:45 crc kubenswrapper[4556]: E0218 09:04:45.282246 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 09:04:45 crc kubenswrapper[4556]: I0218 09:04:45.282346 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:04:45 crc kubenswrapper[4556]: E0218 09:04:45.282488 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 09:04:45 crc kubenswrapper[4556]: E0218 09:04:45.282587 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 09:04:45 crc kubenswrapper[4556]: I0218 09:04:45.373426 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:45 crc kubenswrapper[4556]: I0218 09:04:45.373459 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:45 crc kubenswrapper[4556]: I0218 09:04:45.373470 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:45 crc kubenswrapper[4556]: I0218 09:04:45.373509 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:45 crc kubenswrapper[4556]: I0218 09:04:45.373535 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:45Z","lastTransitionTime":"2026-02-18T09:04:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:45 crc kubenswrapper[4556]: I0218 09:04:45.476360 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:45 crc kubenswrapper[4556]: I0218 09:04:45.476392 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:45 crc kubenswrapper[4556]: I0218 09:04:45.476404 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:45 crc kubenswrapper[4556]: I0218 09:04:45.476418 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:45 crc kubenswrapper[4556]: I0218 09:04:45.476427 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:45Z","lastTransitionTime":"2026-02-18T09:04:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:45 crc kubenswrapper[4556]: I0218 09:04:45.578229 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:45 crc kubenswrapper[4556]: I0218 09:04:45.578271 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:45 crc kubenswrapper[4556]: I0218 09:04:45.578284 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:45 crc kubenswrapper[4556]: I0218 09:04:45.578301 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:45 crc kubenswrapper[4556]: I0218 09:04:45.578316 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:45Z","lastTransitionTime":"2026-02-18T09:04:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:45 crc kubenswrapper[4556]: I0218 09:04:45.680222 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:45 crc kubenswrapper[4556]: I0218 09:04:45.680266 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:45 crc kubenswrapper[4556]: I0218 09:04:45.680276 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:45 crc kubenswrapper[4556]: I0218 09:04:45.680290 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:45 crc kubenswrapper[4556]: I0218 09:04:45.680301 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:45Z","lastTransitionTime":"2026-02-18T09:04:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:45 crc kubenswrapper[4556]: I0218 09:04:45.782118 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:45 crc kubenswrapper[4556]: I0218 09:04:45.782208 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:45 crc kubenswrapper[4556]: I0218 09:04:45.782221 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:45 crc kubenswrapper[4556]: I0218 09:04:45.782251 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:45 crc kubenswrapper[4556]: I0218 09:04:45.782266 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:45Z","lastTransitionTime":"2026-02-18T09:04:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:45 crc kubenswrapper[4556]: I0218 09:04:45.884143 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:45 crc kubenswrapper[4556]: I0218 09:04:45.884200 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:45 crc kubenswrapper[4556]: I0218 09:04:45.884212 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:45 crc kubenswrapper[4556]: I0218 09:04:45.884227 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:45 crc kubenswrapper[4556]: I0218 09:04:45.884238 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:45Z","lastTransitionTime":"2026-02-18T09:04:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:45 crc kubenswrapper[4556]: I0218 09:04:45.987149 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:45 crc kubenswrapper[4556]: I0218 09:04:45.987241 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:45 crc kubenswrapper[4556]: I0218 09:04:45.987250 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:45 crc kubenswrapper[4556]: I0218 09:04:45.987264 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:45 crc kubenswrapper[4556]: I0218 09:04:45.987273 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:45Z","lastTransitionTime":"2026-02-18T09:04:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:46 crc kubenswrapper[4556]: I0218 09:04:46.089110 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:46 crc kubenswrapper[4556]: I0218 09:04:46.089176 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:46 crc kubenswrapper[4556]: I0218 09:04:46.089187 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:46 crc kubenswrapper[4556]: I0218 09:04:46.089202 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:46 crc kubenswrapper[4556]: I0218 09:04:46.089213 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:46Z","lastTransitionTime":"2026-02-18T09:04:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:46 crc kubenswrapper[4556]: I0218 09:04:46.191402 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:46 crc kubenswrapper[4556]: I0218 09:04:46.191440 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:46 crc kubenswrapper[4556]: I0218 09:04:46.191450 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:46 crc kubenswrapper[4556]: I0218 09:04:46.191463 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:46 crc kubenswrapper[4556]: I0218 09:04:46.191473 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:46Z","lastTransitionTime":"2026-02-18T09:04:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:46 crc kubenswrapper[4556]: I0218 09:04:46.273209 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-17 15:10:20.938191982 +0000 UTC Feb 18 09:04:46 crc kubenswrapper[4556]: I0218 09:04:46.281415 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cgwv9" Feb 18 09:04:46 crc kubenswrapper[4556]: E0218 09:04:46.281545 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cgwv9" podUID="d2f0e36c-cb29-4b06-bb22-6afd59466cab" Feb 18 09:04:46 crc kubenswrapper[4556]: I0218 09:04:46.294006 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:46 crc kubenswrapper[4556]: I0218 09:04:46.294039 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:46 crc kubenswrapper[4556]: I0218 09:04:46.294048 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:46 crc kubenswrapper[4556]: I0218 09:04:46.294060 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:46 crc kubenswrapper[4556]: I0218 09:04:46.294071 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:46Z","lastTransitionTime":"2026-02-18T09:04:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:46 crc kubenswrapper[4556]: I0218 09:04:46.395895 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:46 crc kubenswrapper[4556]: I0218 09:04:46.395930 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:46 crc kubenswrapper[4556]: I0218 09:04:46.395939 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:46 crc kubenswrapper[4556]: I0218 09:04:46.395951 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:46 crc kubenswrapper[4556]: I0218 09:04:46.395961 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:46Z","lastTransitionTime":"2026-02-18T09:04:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:46 crc kubenswrapper[4556]: I0218 09:04:46.497652 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:46 crc kubenswrapper[4556]: I0218 09:04:46.497683 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:46 crc kubenswrapper[4556]: I0218 09:04:46.497691 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:46 crc kubenswrapper[4556]: I0218 09:04:46.497702 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:46 crc kubenswrapper[4556]: I0218 09:04:46.497710 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:46Z","lastTransitionTime":"2026-02-18T09:04:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:46 crc kubenswrapper[4556]: I0218 09:04:46.601551 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:46 crc kubenswrapper[4556]: I0218 09:04:46.601596 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:46 crc kubenswrapper[4556]: I0218 09:04:46.601613 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:46 crc kubenswrapper[4556]: I0218 09:04:46.601631 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:46 crc kubenswrapper[4556]: I0218 09:04:46.601640 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:46Z","lastTransitionTime":"2026-02-18T09:04:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:46 crc kubenswrapper[4556]: I0218 09:04:46.704128 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:46 crc kubenswrapper[4556]: I0218 09:04:46.704224 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:46 crc kubenswrapper[4556]: I0218 09:04:46.704240 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:46 crc kubenswrapper[4556]: I0218 09:04:46.704265 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:46 crc kubenswrapper[4556]: I0218 09:04:46.704280 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:46Z","lastTransitionTime":"2026-02-18T09:04:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:46 crc kubenswrapper[4556]: I0218 09:04:46.806298 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:46 crc kubenswrapper[4556]: I0218 09:04:46.806340 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:46 crc kubenswrapper[4556]: I0218 09:04:46.806352 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:46 crc kubenswrapper[4556]: I0218 09:04:46.806368 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:46 crc kubenswrapper[4556]: I0218 09:04:46.806381 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:46Z","lastTransitionTime":"2026-02-18T09:04:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:46 crc kubenswrapper[4556]: I0218 09:04:46.908026 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:46 crc kubenswrapper[4556]: I0218 09:04:46.908070 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:46 crc kubenswrapper[4556]: I0218 09:04:46.908078 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:46 crc kubenswrapper[4556]: I0218 09:04:46.908093 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:46 crc kubenswrapper[4556]: I0218 09:04:46.908103 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:46Z","lastTransitionTime":"2026-02-18T09:04:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:47 crc kubenswrapper[4556]: I0218 09:04:47.010213 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:47 crc kubenswrapper[4556]: I0218 09:04:47.010254 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:47 crc kubenswrapper[4556]: I0218 09:04:47.010264 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:47 crc kubenswrapper[4556]: I0218 09:04:47.010284 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:47 crc kubenswrapper[4556]: I0218 09:04:47.010294 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:47Z","lastTransitionTime":"2026-02-18T09:04:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:47 crc kubenswrapper[4556]: I0218 09:04:47.112697 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:47 crc kubenswrapper[4556]: I0218 09:04:47.112747 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:47 crc kubenswrapper[4556]: I0218 09:04:47.112758 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:47 crc kubenswrapper[4556]: I0218 09:04:47.112777 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:47 crc kubenswrapper[4556]: I0218 09:04:47.112789 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:47Z","lastTransitionTime":"2026-02-18T09:04:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:47 crc kubenswrapper[4556]: I0218 09:04:47.214716 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:47 crc kubenswrapper[4556]: I0218 09:04:47.214768 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:47 crc kubenswrapper[4556]: I0218 09:04:47.214777 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:47 crc kubenswrapper[4556]: I0218 09:04:47.214790 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:47 crc kubenswrapper[4556]: I0218 09:04:47.214801 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:47Z","lastTransitionTime":"2026-02-18T09:04:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:47 crc kubenswrapper[4556]: I0218 09:04:47.273799 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-27 18:57:16.755567342 +0000 UTC Feb 18 09:04:47 crc kubenswrapper[4556]: I0218 09:04:47.282207 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:04:47 crc kubenswrapper[4556]: I0218 09:04:47.282284 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:04:47 crc kubenswrapper[4556]: E0218 09:04:47.282380 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 09:04:47 crc kubenswrapper[4556]: I0218 09:04:47.282430 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:04:47 crc kubenswrapper[4556]: E0218 09:04:47.282575 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 09:04:47 crc kubenswrapper[4556]: E0218 09:04:47.282656 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 09:04:47 crc kubenswrapper[4556]: I0218 09:04:47.316706 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:47 crc kubenswrapper[4556]: I0218 09:04:47.316735 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:47 crc kubenswrapper[4556]: I0218 09:04:47.316743 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:47 crc kubenswrapper[4556]: I0218 09:04:47.316754 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:47 crc kubenswrapper[4556]: I0218 09:04:47.316764 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:47Z","lastTransitionTime":"2026-02-18T09:04:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:47 crc kubenswrapper[4556]: I0218 09:04:47.419120 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:47 crc kubenswrapper[4556]: I0218 09:04:47.419188 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:47 crc kubenswrapper[4556]: I0218 09:04:47.419198 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:47 crc kubenswrapper[4556]: I0218 09:04:47.419212 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:47 crc kubenswrapper[4556]: I0218 09:04:47.419222 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:47Z","lastTransitionTime":"2026-02-18T09:04:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:47 crc kubenswrapper[4556]: I0218 09:04:47.521346 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:47 crc kubenswrapper[4556]: I0218 09:04:47.521382 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:47 crc kubenswrapper[4556]: I0218 09:04:47.521390 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:47 crc kubenswrapper[4556]: I0218 09:04:47.521402 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:47 crc kubenswrapper[4556]: I0218 09:04:47.521414 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:47Z","lastTransitionTime":"2026-02-18T09:04:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:47 crc kubenswrapper[4556]: I0218 09:04:47.623303 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:47 crc kubenswrapper[4556]: I0218 09:04:47.623346 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:47 crc kubenswrapper[4556]: I0218 09:04:47.623356 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:47 crc kubenswrapper[4556]: I0218 09:04:47.623369 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:47 crc kubenswrapper[4556]: I0218 09:04:47.623379 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:47Z","lastTransitionTime":"2026-02-18T09:04:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:47 crc kubenswrapper[4556]: I0218 09:04:47.725108 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:47 crc kubenswrapper[4556]: I0218 09:04:47.725128 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:47 crc kubenswrapper[4556]: I0218 09:04:47.725138 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:47 crc kubenswrapper[4556]: I0218 09:04:47.725175 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:47 crc kubenswrapper[4556]: I0218 09:04:47.725185 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:47Z","lastTransitionTime":"2026-02-18T09:04:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:47 crc kubenswrapper[4556]: I0218 09:04:47.827477 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:47 crc kubenswrapper[4556]: I0218 09:04:47.827519 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:47 crc kubenswrapper[4556]: I0218 09:04:47.827530 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:47 crc kubenswrapper[4556]: I0218 09:04:47.827541 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:47 crc kubenswrapper[4556]: I0218 09:04:47.827551 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:47Z","lastTransitionTime":"2026-02-18T09:04:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:47 crc kubenswrapper[4556]: I0218 09:04:47.929403 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:47 crc kubenswrapper[4556]: I0218 09:04:47.929435 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:47 crc kubenswrapper[4556]: I0218 09:04:47.929443 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:47 crc kubenswrapper[4556]: I0218 09:04:47.929454 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:47 crc kubenswrapper[4556]: I0218 09:04:47.929464 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:47Z","lastTransitionTime":"2026-02-18T09:04:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.033049 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.033108 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.033119 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.033132 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.033140 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:48Z","lastTransitionTime":"2026-02-18T09:04:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.134987 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.135022 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.135032 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.135047 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.135056 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:48Z","lastTransitionTime":"2026-02-18T09:04:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.147010 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d2f0e36c-cb29-4b06-bb22-6afd59466cab-metrics-certs\") pod \"network-metrics-daemon-cgwv9\" (UID: \"d2f0e36c-cb29-4b06-bb22-6afd59466cab\") " pod="openshift-multus/network-metrics-daemon-cgwv9" Feb 18 09:04:48 crc kubenswrapper[4556]: E0218 09:04:48.147130 4556 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 18 09:04:48 crc kubenswrapper[4556]: E0218 09:04:48.147197 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d2f0e36c-cb29-4b06-bb22-6afd59466cab-metrics-certs podName:d2f0e36c-cb29-4b06-bb22-6afd59466cab nodeName:}" failed. No retries permitted until 2026-02-18 09:04:56.147181857 +0000 UTC m=+53.164142838 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d2f0e36c-cb29-4b06-bb22-6afd59466cab-metrics-certs") pod "network-metrics-daemon-cgwv9" (UID: "d2f0e36c-cb29-4b06-bb22-6afd59466cab") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.237017 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.237051 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.237063 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.237080 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.237092 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:48Z","lastTransitionTime":"2026-02-18T09:04:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.274188 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-11 07:28:44.937477835 +0000 UTC Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.281964 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cgwv9" Feb 18 09:04:48 crc kubenswrapper[4556]: E0218 09:04:48.282114 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cgwv9" podUID="d2f0e36c-cb29-4b06-bb22-6afd59466cab" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.282190 4556 scope.go:117] "RemoveContainer" containerID="8fe96f3b6f1f0ce811435d163297af1becac58208736b4da93517aa8780700c8" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.340305 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.340512 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.340521 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.340542 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.340554 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:48Z","lastTransitionTime":"2026-02-18T09:04:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.442655 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.442703 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.442711 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.442723 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.442732 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:48Z","lastTransitionTime":"2026-02-18T09:04:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.489370 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-b7fsj_09971522-31bc-4e10-8831-4fb62675e220/ovnkube-controller/1.log" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.491133 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" event={"ID":"09971522-31bc-4e10-8831-4fb62675e220","Type":"ContainerStarted","Data":"b03acfee4df70e62846f679407e6f9960e44946d5fd4b71d7ea1011729eb8c05"} Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.492111 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.506892 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sgcq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9945b792-abbd-4103-a7b4-9496e1cc1b56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5dd402827182bc79283f1ea1c355b60c4ced53ffa9ce6781b0a0db3109359a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac0d9d7b9bc867b32cf6cbeb804c08e67df757e4b22914e2da69a376511cf75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac0d9d7b9bc867b32cf6cbeb804c08e67df757e4b22914e2da69a376511cf75c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b807d53d620e44d3196f68492b9f9665593adac804bc7b91569796e5594c4ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b807d53d620e44d3196f68492b9f9665593adac804bc7b91569796e5594c4ebb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebc044bbd5b0c1e725e6ed46ec2e0015b254070b7518d20815539dcf8c6e4e45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ebc044bbd5b0c1e725e6ed46ec2e0015b254070b7518d20815539dcf8c6e4e45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sgcq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:48Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.515670 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cgwv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2f0e36c-cb29-4b06-bb22-6afd59466cab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpldv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpldv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:40Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cgwv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:48Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.525118 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"302eb215-a4c2-4324-836d-f37debfd620f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a99dc228d1c4cb8aeab2140ec5f37358b65684739f833fbde45822782430a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14f794063db9d33b82b79acc7134b5a2a3435e09d7cffe49d49251d1f7d20402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5eb3d3728da247903bece56bc2481547317e386e9e97b6c69926a3b55dc89b8e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b42d13c1767c29d22e52537ff45561a0fbf8f2069b2c3612e903e8c127e12485\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e865e6a50c03519ce22958ac4fcbb911fc140c330b6d7ffe5c9d7dd381e96\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T09:04:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0218 09:04:15.143085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 09:04:15.144650 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2160144045/tls.crt::/tmp/serving-cert-2160144045/tls.key\\\\\\\"\\\\nI0218 09:04:20.680799 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 09:04:20.682555 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 09:04:20.682573 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 09:04:20.682590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 09:04:20.682594 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 09:04:20.686822 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 09:04:20.686847 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686852 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686856 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 09:04:20.686859 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 09:04:20.686861 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 09:04:20.686864 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 09:04:20.688257 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 09:04:20.690726 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://974e17ae7082631df2d1bcee60c8a5f142bae2659e9d200d537a4d650f401b4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:48Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.536930 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:48Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.545166 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.545213 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.545223 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.545241 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.545252 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:48Z","lastTransitionTime":"2026-02-18T09:04:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.548955 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qfksk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e226b189-cfa9-47fc-b94d-19b5cbe0859f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc2cb92a7735d16b8e5918797e52e1e488d9261a0418ed5ebbb94c033ca9328a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2fx4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qfksk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:48Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.567841 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dac7f27-d3d1-4778-9e54-f273035a1d37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18da54fe071964d8aae8a977f5ee209c31f1c75e1041c44f8d755f58f3e39d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d41b85076e695bbc51a41fce200e0e47c099d88b4d825b4d133d33ed96237e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f76hs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:48Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.583831 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09971522-31bc-4e10-8831-4fb62675e220\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9170bdcbaf1065995d03e6448e40bdced0741c716c9e81f37a72b015132f1161\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c526c0613f19e1a1a584b7129bd56001faf08d0b97fd77feeb69f764ed01f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cd6e758ee396c9000b1bdb6ded2176ec803fa9da6201cdd9b8146940a2c00cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b0412f3330999c3348ac726ac5a0c2d541571f4b13786ca24cb3109a44aa653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://318cb69e50df4c14df9bf5ef57d5c381421a9bdb72135c9b9620f55c0547b410\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0f15d331af9bc2ca3082c1573dd167c66773eb390593844e681f8d0b455d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03acfee4df70e62846f679407e6f9960e44946d5fd4b71d7ea1011729eb8c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8fe96f3b6f1f0ce811435d163297af1becac58208736b4da93517aa8780700c8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T09:04:36Z\\\",\\\"message\\\":\\\":{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0218 09:04:36.174933 6015 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0218 09:04:36.174942 6015 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI0218 09:04:36.174757 6015 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-sgcq7 in node crc\\\\nI0218 09:04:36.174948 6015 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nF0218 09:04:36.174847 6015 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling w\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:35Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aec6030df0a59ad8902d5c694d5cd30d249ca1e895155c57a458e359fa531b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7fsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:48Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.592710 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47341f31f94eaf5dd86efbe78ff62873e733b1f86ea895f281b6797770519459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:48Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.600336 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5vcgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d50c39aa-f956-441a-8abe-1d7247a7fd86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ddc5d42163d1fa86f1a59e224adac434c990224e839ef3cb12c6d7689e2b439\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4llw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5vcgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:48Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.608441 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jflp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689d3d82-4071-4fe5-b3e3-b5b4f511580a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d56d65964aaf12feaaa3e5fa3ca779bffdb13fcebfc9f027b563c7dd01f4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bm85s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jflp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:48Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.618331 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6387f58543cf2601a53b0bbe9be3ec7816ba90d5ffdc4465594b37ccac0f6dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:48Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.630804 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:48Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.641580 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bp8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"804f8152-54ae-4486-92e6-f9595c035820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb617e46b8522d73dabef4bbc660e41b737f6e6e62f7ce42dad722490a1041dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phr7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://104b8ff3c41d4304d414e58303c23792a7f0ba580f73a2a8678044741af71380\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phr7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bp8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:48Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.647692 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.647730 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.647743 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.647766 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.647782 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:48Z","lastTransitionTime":"2026-02-18T09:04:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.656030 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ae60f3-adab-4740-821a-1570059c7101\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f629262e27020512301a7b53ee836ee98e1e7cdee7a52b0ed506bbaaebd823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f59306c0d55e9d00714ce2ee2c1ed6f2ed33d8b21eecde490e86a93335a14f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c87bc671742a47de6c8e68aaf0c8645f9a8cae7b3d52264f36ae3827b3a8d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2678001ff826712da6716541210f018c74fd3d0218b388d2e7a6f16a1f4027\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6f37933281c927328ecfaf842d97ee8361726768f393726cb39517d2bdc9425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:48Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.666075 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8f6014-2544-4173-a94d-441cf42ba891\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77b0269625b7c6baab44a40373618f4ccfe9e1a028a1bc746b7d60c20ac4efaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac09c421b3101887bfb83107dbe26f2b5856d9b32a9cea92f6233fb6ae218cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce79cfe0242ecf403f989dc6b1c1b8be5ab54eee75ac413a8f9d03cef8bca0ed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://095ecf7e8763fa098b06eb9a8aac9f39451f7762bf20196e0803bbc841610e02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:48Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.676702 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:48Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.685525 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d653c21d3d7c875bed37ed6da700dc8c801b95c80d7e361121dfbbd3d05a922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5211d0cd5172d45b92b003e2bf97d7969a65ea448b0491f8f6ea67364925ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:48Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.750022 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.750061 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.750077 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.750096 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.750109 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:48Z","lastTransitionTime":"2026-02-18T09:04:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.851898 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.851935 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.851945 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.851960 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.851971 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:48Z","lastTransitionTime":"2026-02-18T09:04:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.953873 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.953922 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.953934 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.953951 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:48 crc kubenswrapper[4556]: I0218 09:04:48.953961 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:48Z","lastTransitionTime":"2026-02-18T09:04:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.056260 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.056301 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.056311 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.056327 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.056339 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:49Z","lastTransitionTime":"2026-02-18T09:04:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.158608 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.158648 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.158660 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.158684 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.158697 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:49Z","lastTransitionTime":"2026-02-18T09:04:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.260088 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.260119 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.260129 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.260140 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.260170 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:49Z","lastTransitionTime":"2026-02-18T09:04:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.274656 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-17 11:45:53.471267376 +0000 UTC Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.282099 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:04:49 crc kubenswrapper[4556]: E0218 09:04:49.282253 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.282317 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.282105 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:04:49 crc kubenswrapper[4556]: E0218 09:04:49.282424 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 09:04:49 crc kubenswrapper[4556]: E0218 09:04:49.282460 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.362024 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.362049 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.362060 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.362074 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.362083 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:49Z","lastTransitionTime":"2026-02-18T09:04:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.464177 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.464206 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.464216 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.464231 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.464245 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:49Z","lastTransitionTime":"2026-02-18T09:04:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.495232 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-b7fsj_09971522-31bc-4e10-8831-4fb62675e220/ovnkube-controller/2.log" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.495744 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-b7fsj_09971522-31bc-4e10-8831-4fb62675e220/ovnkube-controller/1.log" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.498006 4556 generic.go:334] "Generic (PLEG): container finished" podID="09971522-31bc-4e10-8831-4fb62675e220" containerID="b03acfee4df70e62846f679407e6f9960e44946d5fd4b71d7ea1011729eb8c05" exitCode=1 Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.498051 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" event={"ID":"09971522-31bc-4e10-8831-4fb62675e220","Type":"ContainerDied","Data":"b03acfee4df70e62846f679407e6f9960e44946d5fd4b71d7ea1011729eb8c05"} Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.498105 4556 scope.go:117] "RemoveContainer" containerID="8fe96f3b6f1f0ce811435d163297af1becac58208736b4da93517aa8780700c8" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.498631 4556 scope.go:117] "RemoveContainer" containerID="b03acfee4df70e62846f679407e6f9960e44946d5fd4b71d7ea1011729eb8c05" Feb 18 09:04:49 crc kubenswrapper[4556]: E0218 09:04:49.498788 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-b7fsj_openshift-ovn-kubernetes(09971522-31bc-4e10-8831-4fb62675e220)\"" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" podUID="09971522-31bc-4e10-8831-4fb62675e220" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.509405 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bp8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"804f8152-54ae-4486-92e6-f9595c035820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb617e46b8522d73dabef4bbc660e41b737f6e6e62f7ce42dad722490a1041dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phr7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://104b8ff3c41d4304d414e58303c23792a7f0ba580f73a2a8678044741af71380\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phr7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bp8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:49Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.523971 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ae60f3-adab-4740-821a-1570059c7101\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f629262e27020512301a7b53ee836ee98e1e7cdee7a52b0ed506bbaaebd823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f59306c0d55e9d00714ce2ee2c1ed6f2ed33d8b21eecde490e86a93335a14f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c87bc671742a47de6c8e68aaf0c8645f9a8cae7b3d52264f36ae3827b3a8d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2678001ff826712da6716541210f018c74fd3d0218b388d2e7a6f16a1f4027\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6f37933281c927328ecfaf842d97ee8361726768f393726cb39517d2bdc9425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:49Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.532832 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8f6014-2544-4173-a94d-441cf42ba891\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77b0269625b7c6baab44a40373618f4ccfe9e1a028a1bc746b7d60c20ac4efaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac09c421b3101887bfb83107dbe26f2b5856d9b32a9cea92f6233fb6ae218cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce79cfe0242ecf403f989dc6b1c1b8be5ab54eee75ac413a8f9d03cef8bca0ed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://095ecf7e8763fa098b06eb9a8aac9f39451f7762bf20196e0803bbc841610e02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:49Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.541633 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:49Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.552845 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d653c21d3d7c875bed37ed6da700dc8c801b95c80d7e361121dfbbd3d05a922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5211d0cd5172d45b92b003e2bf97d7969a65ea448b0491f8f6ea67364925ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:49Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.561804 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6387f58543cf2601a53b0bbe9be3ec7816ba90d5ffdc4465594b37ccac0f6dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:49Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.571325 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.571416 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.571546 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.571612 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.571679 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:49Z","lastTransitionTime":"2026-02-18T09:04:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.573779 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:49Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.583824 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"302eb215-a4c2-4324-836d-f37debfd620f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a99dc228d1c4cb8aeab2140ec5f37358b65684739f833fbde45822782430a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14f794063db9d33b82b79acc7134b5a2a3435e09d7cffe49d49251d1f7d20402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5eb3d3728da247903bece56bc2481547317e386e9e97b6c69926a3b55dc89b8e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b42d13c1767c29d22e52537ff45561a0fbf8f2069b2c3612e903e8c127e12485\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e865e6a50c03519ce22958ac4fcbb911fc140c330b6d7ffe5c9d7dd381e96\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T09:04:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0218 09:04:15.143085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 09:04:15.144650 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2160144045/tls.crt::/tmp/serving-cert-2160144045/tls.key\\\\\\\"\\\\nI0218 09:04:20.680799 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 09:04:20.682555 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 09:04:20.682573 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 09:04:20.682590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 09:04:20.682594 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 09:04:20.686822 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 09:04:20.686847 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686852 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686856 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 09:04:20.686859 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 09:04:20.686861 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 09:04:20.686864 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 09:04:20.688257 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 09:04:20.690726 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://974e17ae7082631df2d1bcee60c8a5f142bae2659e9d200d537a4d650f401b4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:49Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.594913 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:49Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.605566 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qfksk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e226b189-cfa9-47fc-b94d-19b5cbe0859f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc2cb92a7735d16b8e5918797e52e1e488d9261a0418ed5ebbb94c033ca9328a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2fx4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qfksk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:49Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.614725 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dac7f27-d3d1-4778-9e54-f273035a1d37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18da54fe071964d8aae8a977f5ee209c31f1c75e1041c44f8d755f58f3e39d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d41b85076e695bbc51a41fce200e0e47c099d88b4d825b4d133d33ed96237e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f76hs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:49Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.625620 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sgcq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9945b792-abbd-4103-a7b4-9496e1cc1b56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5dd402827182bc79283f1ea1c355b60c4ced53ffa9ce6781b0a0db3109359a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac0d9d7b9bc867b32cf6cbeb804c08e67df757e4b22914e2da69a376511cf75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac0d9d7b9bc867b32cf6cbeb804c08e67df757e4b22914e2da69a376511cf75c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b807d53d620e44d3196f68492b9f9665593adac804bc7b91569796e5594c4ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b807d53d620e44d3196f68492b9f9665593adac804bc7b91569796e5594c4ebb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebc044bbd5b0c1e725e6ed46ec2e0015b254070b7518d20815539dcf8c6e4e45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ebc044bbd5b0c1e725e6ed46ec2e0015b254070b7518d20815539dcf8c6e4e45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sgcq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:49Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.632971 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cgwv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2f0e36c-cb29-4b06-bb22-6afd59466cab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpldv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpldv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:40Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cgwv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:49Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.641194 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47341f31f94eaf5dd86efbe78ff62873e733b1f86ea895f281b6797770519459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:49Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.648334 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5vcgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d50c39aa-f956-441a-8abe-1d7247a7fd86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ddc5d42163d1fa86f1a59e224adac434c990224e839ef3cb12c6d7689e2b439\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4llw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5vcgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:49Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.661182 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09971522-31bc-4e10-8831-4fb62675e220\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9170bdcbaf1065995d03e6448e40bdced0741c716c9e81f37a72b015132f1161\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c526c0613f19e1a1a584b7129bd56001faf08d0b97fd77feeb69f764ed01f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cd6e758ee396c9000b1bdb6ded2176ec803fa9da6201cdd9b8146940a2c00cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b0412f3330999c3348ac726ac5a0c2d541571f4b13786ca24cb3109a44aa653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://318cb69e50df4c14df9bf5ef57d5c381421a9bdb72135c9b9620f55c0547b410\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0f15d331af9bc2ca3082c1573dd167c66773eb390593844e681f8d0b455d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03acfee4df70e62846f679407e6f9960e44946d5fd4b71d7ea1011729eb8c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8fe96f3b6f1f0ce811435d163297af1becac58208736b4da93517aa8780700c8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T09:04:36Z\\\",\\\"message\\\":\\\":{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0218 09:04:36.174933 6015 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0218 09:04:36.174942 6015 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI0218 09:04:36.174757 6015 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-sgcq7 in node crc\\\\nI0218 09:04:36.174948 6015 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nF0218 09:04:36.174847 6015 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling w\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:35Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b03acfee4df70e62846f679407e6f9960e44946d5fd4b71d7ea1011729eb8c05\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T09:04:48Z\\\",\\\"message\\\":\\\"ddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-controller-manager/kube-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager/kube-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.36\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0218 09:04:48.895203 6239 lb_config.go:1031] Cluster endpoints for openshift-machine-api/machine-api-controllers for network=default are: map[]\\\\nI0218 09:04:48.895220 6239 services_controller.go:443] Built service openshift-machine-api/machine-api-controllers LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.167\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8441, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]s\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aec6030df0a59ad8902d5c694d5cd30d249ca1e895155c57a458e359fa531b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7fsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:49Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.668343 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jflp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689d3d82-4071-4fe5-b3e3-b5b4f511580a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d56d65964aaf12feaaa3e5fa3ca779bffdb13fcebfc9f027b563c7dd01f4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bm85s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jflp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:49Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.673988 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.674026 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.674040 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.674056 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.674068 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:49Z","lastTransitionTime":"2026-02-18T09:04:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.776605 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.776648 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.776659 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.776676 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.776687 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:49Z","lastTransitionTime":"2026-02-18T09:04:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.878866 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.878900 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.878929 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.878942 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.878949 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:49Z","lastTransitionTime":"2026-02-18T09:04:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.981004 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.981055 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.981065 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.981077 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:49 crc kubenswrapper[4556]: I0218 09:04:49.981087 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:49Z","lastTransitionTime":"2026-02-18T09:04:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.083182 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.083210 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.083221 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.083231 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.083249 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:50Z","lastTransitionTime":"2026-02-18T09:04:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.185356 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.185398 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.185407 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.185425 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.185437 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:50Z","lastTransitionTime":"2026-02-18T09:04:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.275180 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-17 02:39:10.898098967 +0000 UTC Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.281461 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cgwv9" Feb 18 09:04:50 crc kubenswrapper[4556]: E0218 09:04:50.281603 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cgwv9" podUID="d2f0e36c-cb29-4b06-bb22-6afd59466cab" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.287037 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.287078 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.287090 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.287103 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.287114 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:50Z","lastTransitionTime":"2026-02-18T09:04:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.389700 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.389741 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.389752 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.389765 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.389777 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:50Z","lastTransitionTime":"2026-02-18T09:04:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.428364 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.435182 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.440246 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6387f58543cf2601a53b0bbe9be3ec7816ba90d5ffdc4465594b37ccac0f6dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:50Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.449074 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:50Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.456441 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bp8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"804f8152-54ae-4486-92e6-f9595c035820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb617e46b8522d73dabef4bbc660e41b737f6e6e62f7ce42dad722490a1041dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phr7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://104b8ff3c41d4304d414e58303c23792a7f0ba580f73a2a8678044741af71380\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phr7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bp8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:50Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.470755 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ae60f3-adab-4740-821a-1570059c7101\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f629262e27020512301a7b53ee836ee98e1e7cdee7a52b0ed506bbaaebd823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f59306c0d55e9d00714ce2ee2c1ed6f2ed33d8b21eecde490e86a93335a14f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c87bc671742a47de6c8e68aaf0c8645f9a8cae7b3d52264f36ae3827b3a8d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2678001ff826712da6716541210f018c74fd3d0218b388d2e7a6f16a1f4027\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6f37933281c927328ecfaf842d97ee8361726768f393726cb39517d2bdc9425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:50Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.480664 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8f6014-2544-4173-a94d-441cf42ba891\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77b0269625b7c6baab44a40373618f4ccfe9e1a028a1bc746b7d60c20ac4efaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac09c421b3101887bfb83107dbe26f2b5856d9b32a9cea92f6233fb6ae218cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce79cfe0242ecf403f989dc6b1c1b8be5ab54eee75ac413a8f9d03cef8bca0ed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://095ecf7e8763fa098b06eb9a8aac9f39451f7762bf20196e0803bbc841610e02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:50Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.489577 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:50Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.492283 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.492313 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.492324 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.492340 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.492352 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:50Z","lastTransitionTime":"2026-02-18T09:04:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.499583 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d653c21d3d7c875bed37ed6da700dc8c801b95c80d7e361121dfbbd3d05a922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5211d0cd5172d45b92b003e2bf97d7969a65ea448b0491f8f6ea67364925ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:50Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.504695 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-b7fsj_09971522-31bc-4e10-8831-4fb62675e220/ovnkube-controller/2.log" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.508387 4556 scope.go:117] "RemoveContainer" containerID="b03acfee4df70e62846f679407e6f9960e44946d5fd4b71d7ea1011729eb8c05" Feb 18 09:04:50 crc kubenswrapper[4556]: E0218 09:04:50.508637 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-b7fsj_openshift-ovn-kubernetes(09971522-31bc-4e10-8831-4fb62675e220)\"" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" podUID="09971522-31bc-4e10-8831-4fb62675e220" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.513497 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sgcq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9945b792-abbd-4103-a7b4-9496e1cc1b56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5dd402827182bc79283f1ea1c355b60c4ced53ffa9ce6781b0a0db3109359a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac0d9d7b9bc867b32cf6cbeb804c08e67df757e4b22914e2da69a376511cf75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac0d9d7b9bc867b32cf6cbeb804c08e67df757e4b22914e2da69a376511cf75c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b807d53d620e44d3196f68492b9f9665593adac804bc7b91569796e5594c4ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b807d53d620e44d3196f68492b9f9665593adac804bc7b91569796e5594c4ebb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebc044bbd5b0c1e725e6ed46ec2e0015b254070b7518d20815539dcf8c6e4e45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ebc044bbd5b0c1e725e6ed46ec2e0015b254070b7518d20815539dcf8c6e4e45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sgcq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:50Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.521967 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cgwv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2f0e36c-cb29-4b06-bb22-6afd59466cab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpldv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpldv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:40Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cgwv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:50Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.532058 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"302eb215-a4c2-4324-836d-f37debfd620f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a99dc228d1c4cb8aeab2140ec5f37358b65684739f833fbde45822782430a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14f794063db9d33b82b79acc7134b5a2a3435e09d7cffe49d49251d1f7d20402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5eb3d3728da247903bece56bc2481547317e386e9e97b6c69926a3b55dc89b8e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b42d13c1767c29d22e52537ff45561a0fbf8f2069b2c3612e903e8c127e12485\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e865e6a50c03519ce22958ac4fcbb911fc140c330b6d7ffe5c9d7dd381e96\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T09:04:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0218 09:04:15.143085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 09:04:15.144650 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2160144045/tls.crt::/tmp/serving-cert-2160144045/tls.key\\\\\\\"\\\\nI0218 09:04:20.680799 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 09:04:20.682555 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 09:04:20.682573 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 09:04:20.682590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 09:04:20.682594 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 09:04:20.686822 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 09:04:20.686847 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686852 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686856 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 09:04:20.686859 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 09:04:20.686861 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 09:04:20.686864 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 09:04:20.688257 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 09:04:20.690726 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://974e17ae7082631df2d1bcee60c8a5f142bae2659e9d200d537a4d650f401b4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:50Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.541776 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:50Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.551007 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qfksk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e226b189-cfa9-47fc-b94d-19b5cbe0859f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc2cb92a7735d16b8e5918797e52e1e488d9261a0418ed5ebbb94c033ca9328a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2fx4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qfksk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:50Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.559422 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dac7f27-d3d1-4778-9e54-f273035a1d37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18da54fe071964d8aae8a977f5ee209c31f1c75e1041c44f8d755f58f3e39d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d41b85076e695bbc51a41fce200e0e47c099d88b4d825b4d133d33ed96237e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f76hs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:50Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.572981 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09971522-31bc-4e10-8831-4fb62675e220\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9170bdcbaf1065995d03e6448e40bdced0741c716c9e81f37a72b015132f1161\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c526c0613f19e1a1a584b7129bd56001faf08d0b97fd77feeb69f764ed01f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cd6e758ee396c9000b1bdb6ded2176ec803fa9da6201cdd9b8146940a2c00cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b0412f3330999c3348ac726ac5a0c2d541571f4b13786ca24cb3109a44aa653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://318cb69e50df4c14df9bf5ef57d5c381421a9bdb72135c9b9620f55c0547b410\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0f15d331af9bc2ca3082c1573dd167c66773eb390593844e681f8d0b455d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03acfee4df70e62846f679407e6f9960e44946d5fd4b71d7ea1011729eb8c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8fe96f3b6f1f0ce811435d163297af1becac58208736b4da93517aa8780700c8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T09:04:36Z\\\",\\\"message\\\":\\\":{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0218 09:04:36.174933 6015 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0218 09:04:36.174942 6015 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI0218 09:04:36.174757 6015 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-sgcq7 in node crc\\\\nI0218 09:04:36.174948 6015 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nF0218 09:04:36.174847 6015 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling w\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:35Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b03acfee4df70e62846f679407e6f9960e44946d5fd4b71d7ea1011729eb8c05\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T09:04:48Z\\\",\\\"message\\\":\\\"ddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-controller-manager/kube-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager/kube-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.36\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0218 09:04:48.895203 6239 lb_config.go:1031] Cluster endpoints for openshift-machine-api/machine-api-controllers for network=default are: map[]\\\\nI0218 09:04:48.895220 6239 services_controller.go:443] Built service openshift-machine-api/machine-api-controllers LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.167\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8441, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]s\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aec6030df0a59ad8902d5c694d5cd30d249ca1e895155c57a458e359fa531b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7fsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:50Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.584929 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47341f31f94eaf5dd86efbe78ff62873e733b1f86ea895f281b6797770519459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:50Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.593091 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5vcgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d50c39aa-f956-441a-8abe-1d7247a7fd86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ddc5d42163d1fa86f1a59e224adac434c990224e839ef3cb12c6d7689e2b439\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4llw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5vcgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:50Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.594365 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.594402 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.594415 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.594431 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.594444 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:50Z","lastTransitionTime":"2026-02-18T09:04:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.602101 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jflp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689d3d82-4071-4fe5-b3e3-b5b4f511580a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d56d65964aaf12feaaa3e5fa3ca779bffdb13fcebfc9f027b563c7dd01f4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bm85s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jflp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:50Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.610495 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cgwv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2f0e36c-cb29-4b06-bb22-6afd59466cab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpldv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpldv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:40Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cgwv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:50Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.620898 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"302eb215-a4c2-4324-836d-f37debfd620f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a99dc228d1c4cb8aeab2140ec5f37358b65684739f833fbde45822782430a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14f794063db9d33b82b79acc7134b5a2a3435e09d7cffe49d49251d1f7d20402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5eb3d3728da247903bece56bc2481547317e386e9e97b6c69926a3b55dc89b8e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b42d13c1767c29d22e52537ff45561a0fbf8f2069b2c3612e903e8c127e12485\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e865e6a50c03519ce22958ac4fcbb911fc140c330b6d7ffe5c9d7dd381e96\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T09:04:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0218 09:04:15.143085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 09:04:15.144650 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2160144045/tls.crt::/tmp/serving-cert-2160144045/tls.key\\\\\\\"\\\\nI0218 09:04:20.680799 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 09:04:20.682555 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 09:04:20.682573 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 09:04:20.682590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 09:04:20.682594 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 09:04:20.686822 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 09:04:20.686847 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686852 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686856 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 09:04:20.686859 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 09:04:20.686861 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 09:04:20.686864 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 09:04:20.688257 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 09:04:20.690726 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://974e17ae7082631df2d1bcee60c8a5f142bae2659e9d200d537a4d650f401b4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:50Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.630299 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:50Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.640140 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qfksk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e226b189-cfa9-47fc-b94d-19b5cbe0859f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc2cb92a7735d16b8e5918797e52e1e488d9261a0418ed5ebbb94c033ca9328a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2fx4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qfksk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:50Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.648374 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dac7f27-d3d1-4778-9e54-f273035a1d37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18da54fe071964d8aae8a977f5ee209c31f1c75e1041c44f8d755f58f3e39d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d41b85076e695bbc51a41fce200e0e47c099d88b4d825b4d133d33ed96237e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f76hs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:50Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.659033 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sgcq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9945b792-abbd-4103-a7b4-9496e1cc1b56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5dd402827182bc79283f1ea1c355b60c4ced53ffa9ce6781b0a0db3109359a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac0d9d7b9bc867b32cf6cbeb804c08e67df757e4b22914e2da69a376511cf75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac0d9d7b9bc867b32cf6cbeb804c08e67df757e4b22914e2da69a376511cf75c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b807d53d620e44d3196f68492b9f9665593adac804bc7b91569796e5594c4ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b807d53d620e44d3196f68492b9f9665593adac804bc7b91569796e5594c4ebb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebc044bbd5b0c1e725e6ed46ec2e0015b254070b7518d20815539dcf8c6e4e45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ebc044bbd5b0c1e725e6ed46ec2e0015b254070b7518d20815539dcf8c6e4e45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sgcq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:50Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.667542 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4e7c95a-56e9-4e22-83dd-54962e1233b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52ea6a334e8cad154e46eb2f917cdac914557556bae88f922433a9545cb29eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28a600e8daaa2796b2f0cc6ce4112f8d41c611b5b6ca5a7eb0f260dab615caa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1fa733477ee6330835d52333d6e9588626b6fc7ae08ae450dd2e1821447e995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfbda9799e2d2bc16964e79b90fb7087f9e6893d867711612c63aaaa3dd2b141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfbda9799e2d2bc16964e79b90fb7087f9e6893d867711612c63aaaa3dd2b141\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:50Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.682344 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47341f31f94eaf5dd86efbe78ff62873e733b1f86ea895f281b6797770519459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:50Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.689350 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5vcgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d50c39aa-f956-441a-8abe-1d7247a7fd86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ddc5d42163d1fa86f1a59e224adac434c990224e839ef3cb12c6d7689e2b439\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4llw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5vcgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:50Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.696537 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.696570 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.696579 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.696593 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.696603 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:50Z","lastTransitionTime":"2026-02-18T09:04:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.702294 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09971522-31bc-4e10-8831-4fb62675e220\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9170bdcbaf1065995d03e6448e40bdced0741c716c9e81f37a72b015132f1161\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c526c0613f19e1a1a584b7129bd56001faf08d0b97fd77feeb69f764ed01f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cd6e758ee396c9000b1bdb6ded2176ec803fa9da6201cdd9b8146940a2c00cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b0412f3330999c3348ac726ac5a0c2d541571f4b13786ca24cb3109a44aa653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://318cb69e50df4c14df9bf5ef57d5c381421a9bdb72135c9b9620f55c0547b410\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0f15d331af9bc2ca3082c1573dd167c66773eb390593844e681f8d0b455d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03acfee4df70e62846f679407e6f9960e44946d5fd4b71d7ea1011729eb8c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b03acfee4df70e62846f679407e6f9960e44946d5fd4b71d7ea1011729eb8c05\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T09:04:48Z\\\",\\\"message\\\":\\\"ddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-controller-manager/kube-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager/kube-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.36\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0218 09:04:48.895203 6239 lb_config.go:1031] Cluster endpoints for openshift-machine-api/machine-api-controllers for network=default are: map[]\\\\nI0218 09:04:48.895220 6239 services_controller.go:443] Built service openshift-machine-api/machine-api-controllers LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.167\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8441, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]s\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-b7fsj_openshift-ovn-kubernetes(09971522-31bc-4e10-8831-4fb62675e220)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aec6030df0a59ad8902d5c694d5cd30d249ca1e895155c57a458e359fa531b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7fsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:50Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.709121 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jflp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689d3d82-4071-4fe5-b3e3-b5b4f511580a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d56d65964aaf12feaaa3e5fa3ca779bffdb13fcebfc9f027b563c7dd01f4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bm85s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jflp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:50Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.718544 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:50Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.726636 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bp8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"804f8152-54ae-4486-92e6-f9595c035820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb617e46b8522d73dabef4bbc660e41b737f6e6e62f7ce42dad722490a1041dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phr7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://104b8ff3c41d4304d414e58303c23792a7f0ba580f73a2a8678044741af71380\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phr7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bp8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:50Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.739501 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ae60f3-adab-4740-821a-1570059c7101\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f629262e27020512301a7b53ee836ee98e1e7cdee7a52b0ed506bbaaebd823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f59306c0d55e9d00714ce2ee2c1ed6f2ed33d8b21eecde490e86a93335a14f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c87bc671742a47de6c8e68aaf0c8645f9a8cae7b3d52264f36ae3827b3a8d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2678001ff826712da6716541210f018c74fd3d0218b388d2e7a6f16a1f4027\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6f37933281c927328ecfaf842d97ee8361726768f393726cb39517d2bdc9425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:50Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.748329 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8f6014-2544-4173-a94d-441cf42ba891\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77b0269625b7c6baab44a40373618f4ccfe9e1a028a1bc746b7d60c20ac4efaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac09c421b3101887bfb83107dbe26f2b5856d9b32a9cea92f6233fb6ae218cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce79cfe0242ecf403f989dc6b1c1b8be5ab54eee75ac413a8f9d03cef8bca0ed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://095ecf7e8763fa098b06eb9a8aac9f39451f7762bf20196e0803bbc841610e02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:50Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.757724 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:50Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.766373 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d653c21d3d7c875bed37ed6da700dc8c801b95c80d7e361121dfbbd3d05a922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5211d0cd5172d45b92b003e2bf97d7969a65ea448b0491f8f6ea67364925ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:50Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.775808 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6387f58543cf2601a53b0bbe9be3ec7816ba90d5ffdc4465594b37ccac0f6dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:50Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.798463 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.798526 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.798538 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.798555 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.798566 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:50Z","lastTransitionTime":"2026-02-18T09:04:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.900597 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.900618 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.900665 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.900679 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:50 crc kubenswrapper[4556]: I0218 09:04:50.900689 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:50Z","lastTransitionTime":"2026-02-18T09:04:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:51 crc kubenswrapper[4556]: I0218 09:04:51.002888 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:51 crc kubenswrapper[4556]: I0218 09:04:51.002936 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:51 crc kubenswrapper[4556]: I0218 09:04:51.002947 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:51 crc kubenswrapper[4556]: I0218 09:04:51.002959 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:51 crc kubenswrapper[4556]: I0218 09:04:51.002966 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:51Z","lastTransitionTime":"2026-02-18T09:04:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:51 crc kubenswrapper[4556]: I0218 09:04:51.104795 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:51 crc kubenswrapper[4556]: I0218 09:04:51.104826 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:51 crc kubenswrapper[4556]: I0218 09:04:51.104835 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:51 crc kubenswrapper[4556]: I0218 09:04:51.104881 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:51 crc kubenswrapper[4556]: I0218 09:04:51.104891 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:51Z","lastTransitionTime":"2026-02-18T09:04:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:51 crc kubenswrapper[4556]: I0218 09:04:51.207195 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:51 crc kubenswrapper[4556]: I0218 09:04:51.207230 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:51 crc kubenswrapper[4556]: I0218 09:04:51.207239 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:51 crc kubenswrapper[4556]: I0218 09:04:51.207254 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:51 crc kubenswrapper[4556]: I0218 09:04:51.207264 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:51Z","lastTransitionTime":"2026-02-18T09:04:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:51 crc kubenswrapper[4556]: I0218 09:04:51.276230 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-14 03:09:54.340803357 +0000 UTC Feb 18 09:04:51 crc kubenswrapper[4556]: I0218 09:04:51.281662 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:04:51 crc kubenswrapper[4556]: I0218 09:04:51.281790 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:04:51 crc kubenswrapper[4556]: E0218 09:04:51.281892 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 09:04:51 crc kubenswrapper[4556]: E0218 09:04:51.281902 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 09:04:51 crc kubenswrapper[4556]: I0218 09:04:51.281689 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:04:51 crc kubenswrapper[4556]: E0218 09:04:51.282105 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 09:04:51 crc kubenswrapper[4556]: I0218 09:04:51.309582 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:51 crc kubenswrapper[4556]: I0218 09:04:51.309615 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:51 crc kubenswrapper[4556]: I0218 09:04:51.309624 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:51 crc kubenswrapper[4556]: I0218 09:04:51.309636 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:51 crc kubenswrapper[4556]: I0218 09:04:51.309645 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:51Z","lastTransitionTime":"2026-02-18T09:04:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:51 crc kubenswrapper[4556]: I0218 09:04:51.411909 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:51 crc kubenswrapper[4556]: I0218 09:04:51.411941 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:51 crc kubenswrapper[4556]: I0218 09:04:51.411952 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:51 crc kubenswrapper[4556]: I0218 09:04:51.411967 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:51 crc kubenswrapper[4556]: I0218 09:04:51.411979 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:51Z","lastTransitionTime":"2026-02-18T09:04:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:51 crc kubenswrapper[4556]: I0218 09:04:51.514191 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:51 crc kubenswrapper[4556]: I0218 09:04:51.514233 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:51 crc kubenswrapper[4556]: I0218 09:04:51.514242 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:51 crc kubenswrapper[4556]: I0218 09:04:51.514253 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:51 crc kubenswrapper[4556]: I0218 09:04:51.514261 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:51Z","lastTransitionTime":"2026-02-18T09:04:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:51 crc kubenswrapper[4556]: I0218 09:04:51.615917 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:51 crc kubenswrapper[4556]: I0218 09:04:51.615965 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:51 crc kubenswrapper[4556]: I0218 09:04:51.615975 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:51 crc kubenswrapper[4556]: I0218 09:04:51.615986 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:51 crc kubenswrapper[4556]: I0218 09:04:51.615996 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:51Z","lastTransitionTime":"2026-02-18T09:04:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:51 crc kubenswrapper[4556]: I0218 09:04:51.718483 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:51 crc kubenswrapper[4556]: I0218 09:04:51.718803 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:51 crc kubenswrapper[4556]: I0218 09:04:51.718818 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:51 crc kubenswrapper[4556]: I0218 09:04:51.719046 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:51 crc kubenswrapper[4556]: I0218 09:04:51.719066 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:51Z","lastTransitionTime":"2026-02-18T09:04:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:51 crc kubenswrapper[4556]: I0218 09:04:51.822406 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:51 crc kubenswrapper[4556]: I0218 09:04:51.822547 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:51 crc kubenswrapper[4556]: I0218 09:04:51.822629 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:51 crc kubenswrapper[4556]: I0218 09:04:51.822696 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:51 crc kubenswrapper[4556]: I0218 09:04:51.822766 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:51Z","lastTransitionTime":"2026-02-18T09:04:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:51 crc kubenswrapper[4556]: I0218 09:04:51.925282 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:51 crc kubenswrapper[4556]: I0218 09:04:51.925319 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:51 crc kubenswrapper[4556]: I0218 09:04:51.925331 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:51 crc kubenswrapper[4556]: I0218 09:04:51.925345 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:51 crc kubenswrapper[4556]: I0218 09:04:51.925356 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:51Z","lastTransitionTime":"2026-02-18T09:04:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:52 crc kubenswrapper[4556]: I0218 09:04:52.027060 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:52 crc kubenswrapper[4556]: I0218 09:04:52.027109 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:52 crc kubenswrapper[4556]: I0218 09:04:52.027121 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:52 crc kubenswrapper[4556]: I0218 09:04:52.027140 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:52 crc kubenswrapper[4556]: I0218 09:04:52.027180 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:52Z","lastTransitionTime":"2026-02-18T09:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:52 crc kubenswrapper[4556]: I0218 09:04:52.129425 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:52 crc kubenswrapper[4556]: I0218 09:04:52.129494 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:52 crc kubenswrapper[4556]: I0218 09:04:52.129505 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:52 crc kubenswrapper[4556]: I0218 09:04:52.129524 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:52 crc kubenswrapper[4556]: I0218 09:04:52.129536 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:52Z","lastTransitionTime":"2026-02-18T09:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:52 crc kubenswrapper[4556]: I0218 09:04:52.232279 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:52 crc kubenswrapper[4556]: I0218 09:04:52.232331 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:52 crc kubenswrapper[4556]: I0218 09:04:52.232341 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:52 crc kubenswrapper[4556]: I0218 09:04:52.232361 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:52 crc kubenswrapper[4556]: I0218 09:04:52.232372 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:52Z","lastTransitionTime":"2026-02-18T09:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:52 crc kubenswrapper[4556]: I0218 09:04:52.277284 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-12 14:06:26.046076268 +0000 UTC Feb 18 09:04:52 crc kubenswrapper[4556]: I0218 09:04:52.281641 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cgwv9" Feb 18 09:04:52 crc kubenswrapper[4556]: E0218 09:04:52.281759 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cgwv9" podUID="d2f0e36c-cb29-4b06-bb22-6afd59466cab" Feb 18 09:04:52 crc kubenswrapper[4556]: I0218 09:04:52.333786 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:52 crc kubenswrapper[4556]: I0218 09:04:52.333815 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:52 crc kubenswrapper[4556]: I0218 09:04:52.333825 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:52 crc kubenswrapper[4556]: I0218 09:04:52.333838 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:52 crc kubenswrapper[4556]: I0218 09:04:52.333848 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:52Z","lastTransitionTime":"2026-02-18T09:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:52 crc kubenswrapper[4556]: I0218 09:04:52.436765 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:52 crc kubenswrapper[4556]: I0218 09:04:52.436801 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:52 crc kubenswrapper[4556]: I0218 09:04:52.436812 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:52 crc kubenswrapper[4556]: I0218 09:04:52.436824 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:52 crc kubenswrapper[4556]: I0218 09:04:52.436833 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:52Z","lastTransitionTime":"2026-02-18T09:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:52 crc kubenswrapper[4556]: I0218 09:04:52.539342 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:52 crc kubenswrapper[4556]: I0218 09:04:52.539392 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:52 crc kubenswrapper[4556]: I0218 09:04:52.539404 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:52 crc kubenswrapper[4556]: I0218 09:04:52.539426 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:52 crc kubenswrapper[4556]: I0218 09:04:52.539440 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:52Z","lastTransitionTime":"2026-02-18T09:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:52 crc kubenswrapper[4556]: I0218 09:04:52.641226 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:52 crc kubenswrapper[4556]: I0218 09:04:52.641267 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:52 crc kubenswrapper[4556]: I0218 09:04:52.641278 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:52 crc kubenswrapper[4556]: I0218 09:04:52.641299 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:52 crc kubenswrapper[4556]: I0218 09:04:52.641310 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:52Z","lastTransitionTime":"2026-02-18T09:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:52 crc kubenswrapper[4556]: I0218 09:04:52.743175 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:52 crc kubenswrapper[4556]: I0218 09:04:52.743218 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:52 crc kubenswrapper[4556]: I0218 09:04:52.743231 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:52 crc kubenswrapper[4556]: I0218 09:04:52.743248 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:52 crc kubenswrapper[4556]: I0218 09:04:52.743259 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:52Z","lastTransitionTime":"2026-02-18T09:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:52 crc kubenswrapper[4556]: I0218 09:04:52.845883 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:52 crc kubenswrapper[4556]: I0218 09:04:52.845949 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:52 crc kubenswrapper[4556]: I0218 09:04:52.845962 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:52 crc kubenswrapper[4556]: I0218 09:04:52.845983 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:52 crc kubenswrapper[4556]: I0218 09:04:52.846005 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:52Z","lastTransitionTime":"2026-02-18T09:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:52 crc kubenswrapper[4556]: I0218 09:04:52.948135 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:52 crc kubenswrapper[4556]: I0218 09:04:52.948214 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:52 crc kubenswrapper[4556]: I0218 09:04:52.948224 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:52 crc kubenswrapper[4556]: I0218 09:04:52.948246 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:52 crc kubenswrapper[4556]: I0218 09:04:52.948259 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:52Z","lastTransitionTime":"2026-02-18T09:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.050204 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.050251 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.050264 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.050284 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.050301 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:53Z","lastTransitionTime":"2026-02-18T09:04:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.095856 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.095948 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:04:53 crc kubenswrapper[4556]: E0218 09:04:53.096027 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 09:05:25.096002325 +0000 UTC m=+82.112963305 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.096080 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:04:53 crc kubenswrapper[4556]: E0218 09:04:53.096093 4556 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 18 09:04:53 crc kubenswrapper[4556]: E0218 09:04:53.096187 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-18 09:05:25.096144924 +0000 UTC m=+82.113105904 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 18 09:04:53 crc kubenswrapper[4556]: E0218 09:04:53.096250 4556 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 18 09:04:53 crc kubenswrapper[4556]: E0218 09:04:53.096300 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-18 09:05:25.096290589 +0000 UTC m=+82.113251569 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.151862 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.151896 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.151906 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.151920 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.151931 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:53Z","lastTransitionTime":"2026-02-18T09:04:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.197386 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.197440 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:04:53 crc kubenswrapper[4556]: E0218 09:04:53.197581 4556 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 18 09:04:53 crc kubenswrapper[4556]: E0218 09:04:53.197606 4556 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 18 09:04:53 crc kubenswrapper[4556]: E0218 09:04:53.197618 4556 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 09:04:53 crc kubenswrapper[4556]: E0218 09:04:53.197665 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-18 09:05:25.19764909 +0000 UTC m=+82.214610069 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 09:04:53 crc kubenswrapper[4556]: E0218 09:04:53.197581 4556 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 18 09:04:53 crc kubenswrapper[4556]: E0218 09:04:53.197716 4556 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 18 09:04:53 crc kubenswrapper[4556]: E0218 09:04:53.197730 4556 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 09:04:53 crc kubenswrapper[4556]: E0218 09:04:53.197770 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-18 09:05:25.197757814 +0000 UTC m=+82.214718794 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.253621 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.253652 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.253661 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.253673 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.253684 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:53Z","lastTransitionTime":"2026-02-18T09:04:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.277916 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-16 00:47:45.459608826 +0000 UTC Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.282380 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.282395 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:04:53 crc kubenswrapper[4556]: E0218 09:04:53.282880 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.282927 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:04:53 crc kubenswrapper[4556]: E0218 09:04:53.283020 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 09:04:53 crc kubenswrapper[4556]: E0218 09:04:53.283186 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.292555 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jflp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689d3d82-4071-4fe5-b3e3-b5b4f511580a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d56d65964aaf12feaaa3e5fa3ca779bffdb13fcebfc9f027b563c7dd01f4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bm85s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jflp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:53Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.301118 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:53Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.309037 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bp8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"804f8152-54ae-4486-92e6-f9595c035820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb617e46b8522d73dabef4bbc660e41b737f6e6e62f7ce42dad722490a1041dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phr7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://104b8ff3c41d4304d414e58303c23792a7f0ba580f73a2a8678044741af71380\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phr7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bp8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:53Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.321791 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ae60f3-adab-4740-821a-1570059c7101\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f629262e27020512301a7b53ee836ee98e1e7cdee7a52b0ed506bbaaebd823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f59306c0d55e9d00714ce2ee2c1ed6f2ed33d8b21eecde490e86a93335a14f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c87bc671742a47de6c8e68aaf0c8645f9a8cae7b3d52264f36ae3827b3a8d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2678001ff826712da6716541210f018c74fd3d0218b388d2e7a6f16a1f4027\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6f37933281c927328ecfaf842d97ee8361726768f393726cb39517d2bdc9425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:53Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.329799 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8f6014-2544-4173-a94d-441cf42ba891\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77b0269625b7c6baab44a40373618f4ccfe9e1a028a1bc746b7d60c20ac4efaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac09c421b3101887bfb83107dbe26f2b5856d9b32a9cea92f6233fb6ae218cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce79cfe0242ecf403f989dc6b1c1b8be5ab54eee75ac413a8f9d03cef8bca0ed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://095ecf7e8763fa098b06eb9a8aac9f39451f7762bf20196e0803bbc841610e02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:53Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.338794 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:53Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.351699 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d653c21d3d7c875bed37ed6da700dc8c801b95c80d7e361121dfbbd3d05a922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5211d0cd5172d45b92b003e2bf97d7969a65ea448b0491f8f6ea67364925ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:53Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.355327 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.355444 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.355551 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.355649 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.355726 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:53Z","lastTransitionTime":"2026-02-18T09:04:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.360773 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6387f58543cf2601a53b0bbe9be3ec7816ba90d5ffdc4465594b37ccac0f6dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:53Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.368262 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cgwv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2f0e36c-cb29-4b06-bb22-6afd59466cab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpldv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpldv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:40Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cgwv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:53Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.377338 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"302eb215-a4c2-4324-836d-f37debfd620f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a99dc228d1c4cb8aeab2140ec5f37358b65684739f833fbde45822782430a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14f794063db9d33b82b79acc7134b5a2a3435e09d7cffe49d49251d1f7d20402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5eb3d3728da247903bece56bc2481547317e386e9e97b6c69926a3b55dc89b8e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b42d13c1767c29d22e52537ff45561a0fbf8f2069b2c3612e903e8c127e12485\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e865e6a50c03519ce22958ac4fcbb911fc140c330b6d7ffe5c9d7dd381e96\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T09:04:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0218 09:04:15.143085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 09:04:15.144650 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2160144045/tls.crt::/tmp/serving-cert-2160144045/tls.key\\\\\\\"\\\\nI0218 09:04:20.680799 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 09:04:20.682555 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 09:04:20.682573 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 09:04:20.682590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 09:04:20.682594 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 09:04:20.686822 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 09:04:20.686847 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686852 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686856 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 09:04:20.686859 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 09:04:20.686861 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 09:04:20.686864 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 09:04:20.688257 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 09:04:20.690726 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://974e17ae7082631df2d1bcee60c8a5f142bae2659e9d200d537a4d650f401b4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:53Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.385072 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:53Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.392943 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qfksk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e226b189-cfa9-47fc-b94d-19b5cbe0859f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc2cb92a7735d16b8e5918797e52e1e488d9261a0418ed5ebbb94c033ca9328a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2fx4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qfksk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:53Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.399930 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dac7f27-d3d1-4778-9e54-f273035a1d37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18da54fe071964d8aae8a977f5ee209c31f1c75e1041c44f8d755f58f3e39d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d41b85076e695bbc51a41fce200e0e47c099d88b4d825b4d133d33ed96237e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f76hs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:53Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.409863 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sgcq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9945b792-abbd-4103-a7b4-9496e1cc1b56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5dd402827182bc79283f1ea1c355b60c4ced53ffa9ce6781b0a0db3109359a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac0d9d7b9bc867b32cf6cbeb804c08e67df757e4b22914e2da69a376511cf75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac0d9d7b9bc867b32cf6cbeb804c08e67df757e4b22914e2da69a376511cf75c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b807d53d620e44d3196f68492b9f9665593adac804bc7b91569796e5594c4ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b807d53d620e44d3196f68492b9f9665593adac804bc7b91569796e5594c4ebb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebc044bbd5b0c1e725e6ed46ec2e0015b254070b7518d20815539dcf8c6e4e45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ebc044bbd5b0c1e725e6ed46ec2e0015b254070b7518d20815539dcf8c6e4e45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sgcq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:53Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.417868 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4e7c95a-56e9-4e22-83dd-54962e1233b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52ea6a334e8cad154e46eb2f917cdac914557556bae88f922433a9545cb29eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28a600e8daaa2796b2f0cc6ce4112f8d41c611b5b6ca5a7eb0f260dab615caa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1fa733477ee6330835d52333d6e9588626b6fc7ae08ae450dd2e1821447e995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfbda9799e2d2bc16964e79b90fb7087f9e6893d867711612c63aaaa3dd2b141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfbda9799e2d2bc16964e79b90fb7087f9e6893d867711612c63aaaa3dd2b141\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:53Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.426361 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47341f31f94eaf5dd86efbe78ff62873e733b1f86ea895f281b6797770519459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:53Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.433118 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5vcgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d50c39aa-f956-441a-8abe-1d7247a7fd86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ddc5d42163d1fa86f1a59e224adac434c990224e839ef3cb12c6d7689e2b439\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4llw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5vcgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:53Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.445302 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09971522-31bc-4e10-8831-4fb62675e220\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9170bdcbaf1065995d03e6448e40bdced0741c716c9e81f37a72b015132f1161\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c526c0613f19e1a1a584b7129bd56001faf08d0b97fd77feeb69f764ed01f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cd6e758ee396c9000b1bdb6ded2176ec803fa9da6201cdd9b8146940a2c00cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b0412f3330999c3348ac726ac5a0c2d541571f4b13786ca24cb3109a44aa653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://318cb69e50df4c14df9bf5ef57d5c381421a9bdb72135c9b9620f55c0547b410\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0f15d331af9bc2ca3082c1573dd167c66773eb390593844e681f8d0b455d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03acfee4df70e62846f679407e6f9960e44946d5fd4b71d7ea1011729eb8c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b03acfee4df70e62846f679407e6f9960e44946d5fd4b71d7ea1011729eb8c05\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T09:04:48Z\\\",\\\"message\\\":\\\"ddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-controller-manager/kube-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager/kube-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.36\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0218 09:04:48.895203 6239 lb_config.go:1031] Cluster endpoints for openshift-machine-api/machine-api-controllers for network=default are: map[]\\\\nI0218 09:04:48.895220 6239 services_controller.go:443] Built service openshift-machine-api/machine-api-controllers LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.167\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8441, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]s\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-b7fsj_openshift-ovn-kubernetes(09971522-31bc-4e10-8831-4fb62675e220)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aec6030df0a59ad8902d5c694d5cd30d249ca1e895155c57a458e359fa531b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7fsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:53Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.457722 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.457758 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.457767 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.457784 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.457795 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:53Z","lastTransitionTime":"2026-02-18T09:04:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.560840 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.560885 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.560894 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.560909 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.560920 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:53Z","lastTransitionTime":"2026-02-18T09:04:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.663454 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.663517 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.663529 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.663552 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.663569 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:53Z","lastTransitionTime":"2026-02-18T09:04:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.766867 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.766910 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.766920 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.766938 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.766946 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:53Z","lastTransitionTime":"2026-02-18T09:04:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.868811 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.868848 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.868857 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.868870 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.868882 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:53Z","lastTransitionTime":"2026-02-18T09:04:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.971128 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.971181 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.971192 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.971210 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:53 crc kubenswrapper[4556]: I0218 09:04:53.971221 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:53Z","lastTransitionTime":"2026-02-18T09:04:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.073210 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.073281 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.073293 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.073314 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.073328 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:54Z","lastTransitionTime":"2026-02-18T09:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.175173 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.175214 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.175223 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.175238 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.175249 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:54Z","lastTransitionTime":"2026-02-18T09:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.277989 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-17 01:54:10.813386113 +0000 UTC Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.278287 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.278330 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.278344 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.278361 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.278378 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:54Z","lastTransitionTime":"2026-02-18T09:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.281556 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cgwv9" Feb 18 09:04:54 crc kubenswrapper[4556]: E0218 09:04:54.281689 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cgwv9" podUID="d2f0e36c-cb29-4b06-bb22-6afd59466cab" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.380794 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.380837 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.380848 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.380866 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.380877 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:54Z","lastTransitionTime":"2026-02-18T09:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.414992 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.415027 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.415037 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.415050 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.415060 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:54Z","lastTransitionTime":"2026-02-18T09:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:54 crc kubenswrapper[4556]: E0218 09:04:54.424960 4556 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8758c02e-ef59-46a7-908c-d5c97feb8ceb\\\",\\\"systemUUID\\\":\\\"0f904183-79bf-4790-8551-43a6eb3adbe4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:54Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.427923 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.427955 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.427983 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.427996 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.428005 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:54Z","lastTransitionTime":"2026-02-18T09:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:54 crc kubenswrapper[4556]: E0218 09:04:54.439377 4556 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8758c02e-ef59-46a7-908c-d5c97feb8ceb\\\",\\\"systemUUID\\\":\\\"0f904183-79bf-4790-8551-43a6eb3adbe4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:54Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.442858 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.442891 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.442902 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.442941 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.442950 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:54Z","lastTransitionTime":"2026-02-18T09:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:54 crc kubenswrapper[4556]: E0218 09:04:54.452517 4556 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8758c02e-ef59-46a7-908c-d5c97feb8ceb\\\",\\\"systemUUID\\\":\\\"0f904183-79bf-4790-8551-43a6eb3adbe4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:54Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.455054 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.455083 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.455092 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.455105 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.455131 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:54Z","lastTransitionTime":"2026-02-18T09:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:54 crc kubenswrapper[4556]: E0218 09:04:54.464970 4556 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8758c02e-ef59-46a7-908c-d5c97feb8ceb\\\",\\\"systemUUID\\\":\\\"0f904183-79bf-4790-8551-43a6eb3adbe4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:54Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.467621 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.467677 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.467690 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.467703 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.467714 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:54Z","lastTransitionTime":"2026-02-18T09:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:54 crc kubenswrapper[4556]: E0218 09:04:54.477595 4556 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:04:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8758c02e-ef59-46a7-908c-d5c97feb8ceb\\\",\\\"systemUUID\\\":\\\"0f904183-79bf-4790-8551-43a6eb3adbe4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:04:54Z is after 2025-08-24T17:21:41Z" Feb 18 09:04:54 crc kubenswrapper[4556]: E0218 09:04:54.477699 4556 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.482936 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.482984 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.482994 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.483004 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.483013 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:54Z","lastTransitionTime":"2026-02-18T09:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.584500 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.584533 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.584543 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.584558 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.584570 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:54Z","lastTransitionTime":"2026-02-18T09:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.687014 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.687042 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.687052 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.687065 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.687074 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:54Z","lastTransitionTime":"2026-02-18T09:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.789409 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.789443 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.789455 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.789479 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.789490 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:54Z","lastTransitionTime":"2026-02-18T09:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.891862 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.891888 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.891898 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.891910 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.891920 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:54Z","lastTransitionTime":"2026-02-18T09:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.994140 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.994183 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.994193 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.994204 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:54 crc kubenswrapper[4556]: I0218 09:04:54.994392 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:54Z","lastTransitionTime":"2026-02-18T09:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:55 crc kubenswrapper[4556]: I0218 09:04:55.095943 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:55 crc kubenswrapper[4556]: I0218 09:04:55.095972 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:55 crc kubenswrapper[4556]: I0218 09:04:55.095983 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:55 crc kubenswrapper[4556]: I0218 09:04:55.095996 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:55 crc kubenswrapper[4556]: I0218 09:04:55.096005 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:55Z","lastTransitionTime":"2026-02-18T09:04:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:55 crc kubenswrapper[4556]: I0218 09:04:55.197850 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:55 crc kubenswrapper[4556]: I0218 09:04:55.197897 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:55 crc kubenswrapper[4556]: I0218 09:04:55.197906 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:55 crc kubenswrapper[4556]: I0218 09:04:55.197924 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:55 crc kubenswrapper[4556]: I0218 09:04:55.197938 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:55Z","lastTransitionTime":"2026-02-18T09:04:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:55 crc kubenswrapper[4556]: I0218 09:04:55.278685 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-01 18:59:08.550853682 +0000 UTC Feb 18 09:04:55 crc kubenswrapper[4556]: I0218 09:04:55.282120 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:04:55 crc kubenswrapper[4556]: I0218 09:04:55.282139 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:04:55 crc kubenswrapper[4556]: I0218 09:04:55.282179 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:04:55 crc kubenswrapper[4556]: E0218 09:04:55.282294 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 09:04:55 crc kubenswrapper[4556]: E0218 09:04:55.282406 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 09:04:55 crc kubenswrapper[4556]: E0218 09:04:55.282492 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 09:04:55 crc kubenswrapper[4556]: I0218 09:04:55.299737 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:55 crc kubenswrapper[4556]: I0218 09:04:55.299767 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:55 crc kubenswrapper[4556]: I0218 09:04:55.299776 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:55 crc kubenswrapper[4556]: I0218 09:04:55.299791 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:55 crc kubenswrapper[4556]: I0218 09:04:55.299800 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:55Z","lastTransitionTime":"2026-02-18T09:04:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:55 crc kubenswrapper[4556]: I0218 09:04:55.402113 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:55 crc kubenswrapper[4556]: I0218 09:04:55.402278 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:55 crc kubenswrapper[4556]: I0218 09:04:55.402376 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:55 crc kubenswrapper[4556]: I0218 09:04:55.402449 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:55 crc kubenswrapper[4556]: I0218 09:04:55.402527 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:55Z","lastTransitionTime":"2026-02-18T09:04:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:55 crc kubenswrapper[4556]: I0218 09:04:55.506441 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:55 crc kubenswrapper[4556]: I0218 09:04:55.506505 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:55 crc kubenswrapper[4556]: I0218 09:04:55.506517 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:55 crc kubenswrapper[4556]: I0218 09:04:55.506531 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:55 crc kubenswrapper[4556]: I0218 09:04:55.506541 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:55Z","lastTransitionTime":"2026-02-18T09:04:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:55 crc kubenswrapper[4556]: I0218 09:04:55.607935 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:55 crc kubenswrapper[4556]: I0218 09:04:55.607971 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:55 crc kubenswrapper[4556]: I0218 09:04:55.607982 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:55 crc kubenswrapper[4556]: I0218 09:04:55.607996 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:55 crc kubenswrapper[4556]: I0218 09:04:55.608008 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:55Z","lastTransitionTime":"2026-02-18T09:04:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:55 crc kubenswrapper[4556]: I0218 09:04:55.710006 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:55 crc kubenswrapper[4556]: I0218 09:04:55.710036 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:55 crc kubenswrapper[4556]: I0218 09:04:55.710046 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:55 crc kubenswrapper[4556]: I0218 09:04:55.710059 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:55 crc kubenswrapper[4556]: I0218 09:04:55.710068 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:55Z","lastTransitionTime":"2026-02-18T09:04:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:55 crc kubenswrapper[4556]: I0218 09:04:55.812280 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:55 crc kubenswrapper[4556]: I0218 09:04:55.812343 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:55 crc kubenswrapper[4556]: I0218 09:04:55.812360 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:55 crc kubenswrapper[4556]: I0218 09:04:55.812381 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:55 crc kubenswrapper[4556]: I0218 09:04:55.812398 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:55Z","lastTransitionTime":"2026-02-18T09:04:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:55 crc kubenswrapper[4556]: I0218 09:04:55.913978 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:55 crc kubenswrapper[4556]: I0218 09:04:55.914020 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:55 crc kubenswrapper[4556]: I0218 09:04:55.914038 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:55 crc kubenswrapper[4556]: I0218 09:04:55.914055 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:55 crc kubenswrapper[4556]: I0218 09:04:55.914069 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:55Z","lastTransitionTime":"2026-02-18T09:04:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:56 crc kubenswrapper[4556]: I0218 09:04:56.016403 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:56 crc kubenswrapper[4556]: I0218 09:04:56.016435 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:56 crc kubenswrapper[4556]: I0218 09:04:56.016445 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:56 crc kubenswrapper[4556]: I0218 09:04:56.016458 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:56 crc kubenswrapper[4556]: I0218 09:04:56.016478 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:56Z","lastTransitionTime":"2026-02-18T09:04:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:56 crc kubenswrapper[4556]: I0218 09:04:56.118369 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:56 crc kubenswrapper[4556]: I0218 09:04:56.118400 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:56 crc kubenswrapper[4556]: I0218 09:04:56.118410 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:56 crc kubenswrapper[4556]: I0218 09:04:56.118422 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:56 crc kubenswrapper[4556]: I0218 09:04:56.118431 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:56Z","lastTransitionTime":"2026-02-18T09:04:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:56 crc kubenswrapper[4556]: I0218 09:04:56.222431 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d2f0e36c-cb29-4b06-bb22-6afd59466cab-metrics-certs\") pod \"network-metrics-daemon-cgwv9\" (UID: \"d2f0e36c-cb29-4b06-bb22-6afd59466cab\") " pod="openshift-multus/network-metrics-daemon-cgwv9" Feb 18 09:04:56 crc kubenswrapper[4556]: E0218 09:04:56.222573 4556 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 18 09:04:56 crc kubenswrapper[4556]: E0218 09:04:56.222632 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d2f0e36c-cb29-4b06-bb22-6afd59466cab-metrics-certs podName:d2f0e36c-cb29-4b06-bb22-6afd59466cab nodeName:}" failed. No retries permitted until 2026-02-18 09:05:12.222614787 +0000 UTC m=+69.239575767 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d2f0e36c-cb29-4b06-bb22-6afd59466cab-metrics-certs") pod "network-metrics-daemon-cgwv9" (UID: "d2f0e36c-cb29-4b06-bb22-6afd59466cab") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 18 09:04:56 crc kubenswrapper[4556]: I0218 09:04:56.223760 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:56 crc kubenswrapper[4556]: I0218 09:04:56.223804 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:56 crc kubenswrapper[4556]: I0218 09:04:56.223816 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:56 crc kubenswrapper[4556]: I0218 09:04:56.223833 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:56 crc kubenswrapper[4556]: I0218 09:04:56.223843 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:56Z","lastTransitionTime":"2026-02-18T09:04:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:56 crc kubenswrapper[4556]: I0218 09:04:56.279765 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-07 02:23:36.896831758 +0000 UTC Feb 18 09:04:56 crc kubenswrapper[4556]: I0218 09:04:56.282189 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cgwv9" Feb 18 09:04:56 crc kubenswrapper[4556]: E0218 09:04:56.282315 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cgwv9" podUID="d2f0e36c-cb29-4b06-bb22-6afd59466cab" Feb 18 09:04:56 crc kubenswrapper[4556]: I0218 09:04:56.325502 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:56 crc kubenswrapper[4556]: I0218 09:04:56.325535 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:56 crc kubenswrapper[4556]: I0218 09:04:56.325544 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:56 crc kubenswrapper[4556]: I0218 09:04:56.325560 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:56 crc kubenswrapper[4556]: I0218 09:04:56.325570 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:56Z","lastTransitionTime":"2026-02-18T09:04:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:56 crc kubenswrapper[4556]: I0218 09:04:56.427950 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:56 crc kubenswrapper[4556]: I0218 09:04:56.428020 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:56 crc kubenswrapper[4556]: I0218 09:04:56.428031 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:56 crc kubenswrapper[4556]: I0218 09:04:56.428047 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:56 crc kubenswrapper[4556]: I0218 09:04:56.428060 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:56Z","lastTransitionTime":"2026-02-18T09:04:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:56 crc kubenswrapper[4556]: I0218 09:04:56.529268 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:56 crc kubenswrapper[4556]: I0218 09:04:56.529297 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:56 crc kubenswrapper[4556]: I0218 09:04:56.529307 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:56 crc kubenswrapper[4556]: I0218 09:04:56.529319 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:56 crc kubenswrapper[4556]: I0218 09:04:56.529328 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:56Z","lastTransitionTime":"2026-02-18T09:04:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:56 crc kubenswrapper[4556]: I0218 09:04:56.631426 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:56 crc kubenswrapper[4556]: I0218 09:04:56.631474 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:56 crc kubenswrapper[4556]: I0218 09:04:56.631483 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:56 crc kubenswrapper[4556]: I0218 09:04:56.631498 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:56 crc kubenswrapper[4556]: I0218 09:04:56.631510 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:56Z","lastTransitionTime":"2026-02-18T09:04:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:56 crc kubenswrapper[4556]: I0218 09:04:56.733566 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:56 crc kubenswrapper[4556]: I0218 09:04:56.733596 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:56 crc kubenswrapper[4556]: I0218 09:04:56.733606 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:56 crc kubenswrapper[4556]: I0218 09:04:56.733617 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:56 crc kubenswrapper[4556]: I0218 09:04:56.733627 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:56Z","lastTransitionTime":"2026-02-18T09:04:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:56 crc kubenswrapper[4556]: I0218 09:04:56.835530 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:56 crc kubenswrapper[4556]: I0218 09:04:56.835556 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:56 crc kubenswrapper[4556]: I0218 09:04:56.835565 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:56 crc kubenswrapper[4556]: I0218 09:04:56.835590 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:56 crc kubenswrapper[4556]: I0218 09:04:56.835605 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:56Z","lastTransitionTime":"2026-02-18T09:04:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:56 crc kubenswrapper[4556]: I0218 09:04:56.937036 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:56 crc kubenswrapper[4556]: I0218 09:04:56.937068 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:56 crc kubenswrapper[4556]: I0218 09:04:56.937077 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:56 crc kubenswrapper[4556]: I0218 09:04:56.937088 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:56 crc kubenswrapper[4556]: I0218 09:04:56.937098 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:56Z","lastTransitionTime":"2026-02-18T09:04:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:57 crc kubenswrapper[4556]: I0218 09:04:57.039365 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:57 crc kubenswrapper[4556]: I0218 09:04:57.039391 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:57 crc kubenswrapper[4556]: I0218 09:04:57.039400 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:57 crc kubenswrapper[4556]: I0218 09:04:57.039410 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:57 crc kubenswrapper[4556]: I0218 09:04:57.039418 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:57Z","lastTransitionTime":"2026-02-18T09:04:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:57 crc kubenswrapper[4556]: I0218 09:04:57.141069 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:57 crc kubenswrapper[4556]: I0218 09:04:57.141109 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:57 crc kubenswrapper[4556]: I0218 09:04:57.141118 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:57 crc kubenswrapper[4556]: I0218 09:04:57.141135 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:57 crc kubenswrapper[4556]: I0218 09:04:57.141147 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:57Z","lastTransitionTime":"2026-02-18T09:04:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:57 crc kubenswrapper[4556]: I0218 09:04:57.243001 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:57 crc kubenswrapper[4556]: I0218 09:04:57.243035 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:57 crc kubenswrapper[4556]: I0218 09:04:57.243046 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:57 crc kubenswrapper[4556]: I0218 09:04:57.243080 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:57 crc kubenswrapper[4556]: I0218 09:04:57.243091 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:57Z","lastTransitionTime":"2026-02-18T09:04:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:57 crc kubenswrapper[4556]: I0218 09:04:57.280484 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-14 21:31:06.134134591 +0000 UTC Feb 18 09:04:57 crc kubenswrapper[4556]: I0218 09:04:57.281701 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:04:57 crc kubenswrapper[4556]: I0218 09:04:57.281731 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:04:57 crc kubenswrapper[4556]: I0218 09:04:57.281714 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:04:57 crc kubenswrapper[4556]: E0218 09:04:57.281803 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 09:04:57 crc kubenswrapper[4556]: E0218 09:04:57.281848 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 09:04:57 crc kubenswrapper[4556]: E0218 09:04:57.281889 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 09:04:57 crc kubenswrapper[4556]: I0218 09:04:57.345138 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:57 crc kubenswrapper[4556]: I0218 09:04:57.345191 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:57 crc kubenswrapper[4556]: I0218 09:04:57.345203 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:57 crc kubenswrapper[4556]: I0218 09:04:57.345223 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:57 crc kubenswrapper[4556]: I0218 09:04:57.345234 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:57Z","lastTransitionTime":"2026-02-18T09:04:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:57 crc kubenswrapper[4556]: I0218 09:04:57.451624 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:57 crc kubenswrapper[4556]: I0218 09:04:57.451654 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:57 crc kubenswrapper[4556]: I0218 09:04:57.451662 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:57 crc kubenswrapper[4556]: I0218 09:04:57.451677 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:57 crc kubenswrapper[4556]: I0218 09:04:57.451686 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:57Z","lastTransitionTime":"2026-02-18T09:04:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:57 crc kubenswrapper[4556]: I0218 09:04:57.552799 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:57 crc kubenswrapper[4556]: I0218 09:04:57.552831 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:57 crc kubenswrapper[4556]: I0218 09:04:57.552839 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:57 crc kubenswrapper[4556]: I0218 09:04:57.552852 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:57 crc kubenswrapper[4556]: I0218 09:04:57.552862 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:57Z","lastTransitionTime":"2026-02-18T09:04:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:57 crc kubenswrapper[4556]: I0218 09:04:57.654635 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:57 crc kubenswrapper[4556]: I0218 09:04:57.654663 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:57 crc kubenswrapper[4556]: I0218 09:04:57.654671 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:57 crc kubenswrapper[4556]: I0218 09:04:57.654680 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:57 crc kubenswrapper[4556]: I0218 09:04:57.654690 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:57Z","lastTransitionTime":"2026-02-18T09:04:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:57 crc kubenswrapper[4556]: I0218 09:04:57.756082 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:57 crc kubenswrapper[4556]: I0218 09:04:57.756108 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:57 crc kubenswrapper[4556]: I0218 09:04:57.756117 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:57 crc kubenswrapper[4556]: I0218 09:04:57.756128 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:57 crc kubenswrapper[4556]: I0218 09:04:57.756136 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:57Z","lastTransitionTime":"2026-02-18T09:04:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:57 crc kubenswrapper[4556]: I0218 09:04:57.857850 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:57 crc kubenswrapper[4556]: I0218 09:04:57.857874 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:57 crc kubenswrapper[4556]: I0218 09:04:57.857883 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:57 crc kubenswrapper[4556]: I0218 09:04:57.857893 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:57 crc kubenswrapper[4556]: I0218 09:04:57.857900 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:57Z","lastTransitionTime":"2026-02-18T09:04:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:57 crc kubenswrapper[4556]: I0218 09:04:57.959123 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:57 crc kubenswrapper[4556]: I0218 09:04:57.959203 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:57 crc kubenswrapper[4556]: I0218 09:04:57.959215 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:57 crc kubenswrapper[4556]: I0218 09:04:57.959224 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:57 crc kubenswrapper[4556]: I0218 09:04:57.959232 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:57Z","lastTransitionTime":"2026-02-18T09:04:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:58 crc kubenswrapper[4556]: I0218 09:04:58.061014 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:58 crc kubenswrapper[4556]: I0218 09:04:58.061046 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:58 crc kubenswrapper[4556]: I0218 09:04:58.061054 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:58 crc kubenswrapper[4556]: I0218 09:04:58.061062 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:58 crc kubenswrapper[4556]: I0218 09:04:58.061073 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:58Z","lastTransitionTime":"2026-02-18T09:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:58 crc kubenswrapper[4556]: I0218 09:04:58.162954 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:58 crc kubenswrapper[4556]: I0218 09:04:58.162986 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:58 crc kubenswrapper[4556]: I0218 09:04:58.162996 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:58 crc kubenswrapper[4556]: I0218 09:04:58.163012 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:58 crc kubenswrapper[4556]: I0218 09:04:58.163040 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:58Z","lastTransitionTime":"2026-02-18T09:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:58 crc kubenswrapper[4556]: I0218 09:04:58.264992 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:58 crc kubenswrapper[4556]: I0218 09:04:58.265138 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:58 crc kubenswrapper[4556]: I0218 09:04:58.265145 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:58 crc kubenswrapper[4556]: I0218 09:04:58.265181 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:58 crc kubenswrapper[4556]: I0218 09:04:58.265192 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:58Z","lastTransitionTime":"2026-02-18T09:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:58 crc kubenswrapper[4556]: I0218 09:04:58.281334 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-02 03:57:32.974209862 +0000 UTC Feb 18 09:04:58 crc kubenswrapper[4556]: I0218 09:04:58.281450 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cgwv9" Feb 18 09:04:58 crc kubenswrapper[4556]: E0218 09:04:58.281561 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cgwv9" podUID="d2f0e36c-cb29-4b06-bb22-6afd59466cab" Feb 18 09:04:58 crc kubenswrapper[4556]: I0218 09:04:58.366901 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:58 crc kubenswrapper[4556]: I0218 09:04:58.366932 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:58 crc kubenswrapper[4556]: I0218 09:04:58.366942 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:58 crc kubenswrapper[4556]: I0218 09:04:58.366956 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:58 crc kubenswrapper[4556]: I0218 09:04:58.366967 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:58Z","lastTransitionTime":"2026-02-18T09:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:58 crc kubenswrapper[4556]: I0218 09:04:58.468549 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:58 crc kubenswrapper[4556]: I0218 09:04:58.468589 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:58 crc kubenswrapper[4556]: I0218 09:04:58.468601 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:58 crc kubenswrapper[4556]: I0218 09:04:58.468619 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:58 crc kubenswrapper[4556]: I0218 09:04:58.468633 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:58Z","lastTransitionTime":"2026-02-18T09:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:58 crc kubenswrapper[4556]: I0218 09:04:58.571337 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:58 crc kubenswrapper[4556]: I0218 09:04:58.571378 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:58 crc kubenswrapper[4556]: I0218 09:04:58.571388 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:58 crc kubenswrapper[4556]: I0218 09:04:58.571406 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:58 crc kubenswrapper[4556]: I0218 09:04:58.571416 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:58Z","lastTransitionTime":"2026-02-18T09:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:58 crc kubenswrapper[4556]: I0218 09:04:58.673335 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:58 crc kubenswrapper[4556]: I0218 09:04:58.673374 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:58 crc kubenswrapper[4556]: I0218 09:04:58.673383 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:58 crc kubenswrapper[4556]: I0218 09:04:58.673397 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:58 crc kubenswrapper[4556]: I0218 09:04:58.673406 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:58Z","lastTransitionTime":"2026-02-18T09:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:58 crc kubenswrapper[4556]: I0218 09:04:58.775683 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:58 crc kubenswrapper[4556]: I0218 09:04:58.775747 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:58 crc kubenswrapper[4556]: I0218 09:04:58.775759 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:58 crc kubenswrapper[4556]: I0218 09:04:58.775817 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:58 crc kubenswrapper[4556]: I0218 09:04:58.775832 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:58Z","lastTransitionTime":"2026-02-18T09:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:58 crc kubenswrapper[4556]: I0218 09:04:58.878020 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:58 crc kubenswrapper[4556]: I0218 09:04:58.878059 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:58 crc kubenswrapper[4556]: I0218 09:04:58.878070 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:58 crc kubenswrapper[4556]: I0218 09:04:58.878084 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:58 crc kubenswrapper[4556]: I0218 09:04:58.878094 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:58Z","lastTransitionTime":"2026-02-18T09:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:58 crc kubenswrapper[4556]: I0218 09:04:58.979809 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:58 crc kubenswrapper[4556]: I0218 09:04:58.979847 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:58 crc kubenswrapper[4556]: I0218 09:04:58.979857 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:58 crc kubenswrapper[4556]: I0218 09:04:58.979871 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:58 crc kubenswrapper[4556]: I0218 09:04:58.979880 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:58Z","lastTransitionTime":"2026-02-18T09:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:59 crc kubenswrapper[4556]: I0218 09:04:59.081288 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:59 crc kubenswrapper[4556]: I0218 09:04:59.081322 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:59 crc kubenswrapper[4556]: I0218 09:04:59.081333 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:59 crc kubenswrapper[4556]: I0218 09:04:59.081343 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:59 crc kubenswrapper[4556]: I0218 09:04:59.081351 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:59Z","lastTransitionTime":"2026-02-18T09:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:59 crc kubenswrapper[4556]: I0218 09:04:59.183110 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:59 crc kubenswrapper[4556]: I0218 09:04:59.183134 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:59 crc kubenswrapper[4556]: I0218 09:04:59.183142 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:59 crc kubenswrapper[4556]: I0218 09:04:59.183173 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:59 crc kubenswrapper[4556]: I0218 09:04:59.183182 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:59Z","lastTransitionTime":"2026-02-18T09:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:59 crc kubenswrapper[4556]: I0218 09:04:59.281490 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:04:59 crc kubenswrapper[4556]: I0218 09:04:59.281521 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-05 04:09:36.395190972 +0000 UTC Feb 18 09:04:59 crc kubenswrapper[4556]: I0218 09:04:59.281498 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:04:59 crc kubenswrapper[4556]: E0218 09:04:59.281590 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 09:04:59 crc kubenswrapper[4556]: I0218 09:04:59.281605 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:04:59 crc kubenswrapper[4556]: E0218 09:04:59.281654 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 09:04:59 crc kubenswrapper[4556]: E0218 09:04:59.281707 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 09:04:59 crc kubenswrapper[4556]: I0218 09:04:59.284686 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:59 crc kubenswrapper[4556]: I0218 09:04:59.284708 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:59 crc kubenswrapper[4556]: I0218 09:04:59.284717 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:59 crc kubenswrapper[4556]: I0218 09:04:59.284728 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:59 crc kubenswrapper[4556]: I0218 09:04:59.284737 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:59Z","lastTransitionTime":"2026-02-18T09:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:59 crc kubenswrapper[4556]: I0218 09:04:59.386842 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:59 crc kubenswrapper[4556]: I0218 09:04:59.386880 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:59 crc kubenswrapper[4556]: I0218 09:04:59.386890 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:59 crc kubenswrapper[4556]: I0218 09:04:59.386903 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:59 crc kubenswrapper[4556]: I0218 09:04:59.386915 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:59Z","lastTransitionTime":"2026-02-18T09:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:59 crc kubenswrapper[4556]: I0218 09:04:59.488514 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:59 crc kubenswrapper[4556]: I0218 09:04:59.488542 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:59 crc kubenswrapper[4556]: I0218 09:04:59.488551 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:59 crc kubenswrapper[4556]: I0218 09:04:59.488563 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:59 crc kubenswrapper[4556]: I0218 09:04:59.488571 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:59Z","lastTransitionTime":"2026-02-18T09:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:59 crc kubenswrapper[4556]: I0218 09:04:59.589844 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:59 crc kubenswrapper[4556]: I0218 09:04:59.589880 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:59 crc kubenswrapper[4556]: I0218 09:04:59.589888 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:59 crc kubenswrapper[4556]: I0218 09:04:59.589903 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:59 crc kubenswrapper[4556]: I0218 09:04:59.589913 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:59Z","lastTransitionTime":"2026-02-18T09:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:59 crc kubenswrapper[4556]: I0218 09:04:59.692077 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:59 crc kubenswrapper[4556]: I0218 09:04:59.692115 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:59 crc kubenswrapper[4556]: I0218 09:04:59.692127 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:59 crc kubenswrapper[4556]: I0218 09:04:59.692141 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:59 crc kubenswrapper[4556]: I0218 09:04:59.692168 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:59Z","lastTransitionTime":"2026-02-18T09:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:59 crc kubenswrapper[4556]: I0218 09:04:59.793963 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:59 crc kubenswrapper[4556]: I0218 09:04:59.794002 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:59 crc kubenswrapper[4556]: I0218 09:04:59.794014 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:59 crc kubenswrapper[4556]: I0218 09:04:59.794027 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:59 crc kubenswrapper[4556]: I0218 09:04:59.794038 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:59Z","lastTransitionTime":"2026-02-18T09:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:59 crc kubenswrapper[4556]: I0218 09:04:59.896146 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:59 crc kubenswrapper[4556]: I0218 09:04:59.896191 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:59 crc kubenswrapper[4556]: I0218 09:04:59.896200 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:59 crc kubenswrapper[4556]: I0218 09:04:59.896211 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:59 crc kubenswrapper[4556]: I0218 09:04:59.896221 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:59Z","lastTransitionTime":"2026-02-18T09:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:04:59 crc kubenswrapper[4556]: I0218 09:04:59.998385 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:04:59 crc kubenswrapper[4556]: I0218 09:04:59.998416 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:04:59 crc kubenswrapper[4556]: I0218 09:04:59.998425 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:04:59 crc kubenswrapper[4556]: I0218 09:04:59.998435 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:04:59 crc kubenswrapper[4556]: I0218 09:04:59.998446 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:04:59Z","lastTransitionTime":"2026-02-18T09:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:00 crc kubenswrapper[4556]: I0218 09:05:00.100135 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:00 crc kubenswrapper[4556]: I0218 09:05:00.100187 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:00 crc kubenswrapper[4556]: I0218 09:05:00.100199 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:00 crc kubenswrapper[4556]: I0218 09:05:00.100213 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:00 crc kubenswrapper[4556]: I0218 09:05:00.100222 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:00Z","lastTransitionTime":"2026-02-18T09:05:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:00 crc kubenswrapper[4556]: I0218 09:05:00.202187 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:00 crc kubenswrapper[4556]: I0218 09:05:00.202342 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:00 crc kubenswrapper[4556]: I0218 09:05:00.202399 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:00 crc kubenswrapper[4556]: I0218 09:05:00.202479 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:00 crc kubenswrapper[4556]: I0218 09:05:00.202552 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:00Z","lastTransitionTime":"2026-02-18T09:05:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:00 crc kubenswrapper[4556]: I0218 09:05:00.282202 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-13 19:29:46.599334252 +0000 UTC Feb 18 09:05:00 crc kubenswrapper[4556]: I0218 09:05:00.282276 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cgwv9" Feb 18 09:05:00 crc kubenswrapper[4556]: E0218 09:05:00.282744 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cgwv9" podUID="d2f0e36c-cb29-4b06-bb22-6afd59466cab" Feb 18 09:05:00 crc kubenswrapper[4556]: I0218 09:05:00.304895 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:00 crc kubenswrapper[4556]: I0218 09:05:00.304943 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:00 crc kubenswrapper[4556]: I0218 09:05:00.304954 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:00 crc kubenswrapper[4556]: I0218 09:05:00.304971 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:00 crc kubenswrapper[4556]: I0218 09:05:00.304981 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:00Z","lastTransitionTime":"2026-02-18T09:05:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:00 crc kubenswrapper[4556]: I0218 09:05:00.406972 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:00 crc kubenswrapper[4556]: I0218 09:05:00.407005 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:00 crc kubenswrapper[4556]: I0218 09:05:00.407013 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:00 crc kubenswrapper[4556]: I0218 09:05:00.407027 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:00 crc kubenswrapper[4556]: I0218 09:05:00.407037 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:00Z","lastTransitionTime":"2026-02-18T09:05:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:00 crc kubenswrapper[4556]: I0218 09:05:00.508684 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:00 crc kubenswrapper[4556]: I0218 09:05:00.508830 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:00 crc kubenswrapper[4556]: I0218 09:05:00.508889 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:00 crc kubenswrapper[4556]: I0218 09:05:00.508951 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:00 crc kubenswrapper[4556]: I0218 09:05:00.509017 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:00Z","lastTransitionTime":"2026-02-18T09:05:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:00 crc kubenswrapper[4556]: I0218 09:05:00.611469 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:00 crc kubenswrapper[4556]: I0218 09:05:00.611507 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:00 crc kubenswrapper[4556]: I0218 09:05:00.611518 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:00 crc kubenswrapper[4556]: I0218 09:05:00.611534 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:00 crc kubenswrapper[4556]: I0218 09:05:00.611544 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:00Z","lastTransitionTime":"2026-02-18T09:05:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:00 crc kubenswrapper[4556]: I0218 09:05:00.714231 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:00 crc kubenswrapper[4556]: I0218 09:05:00.714267 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:00 crc kubenswrapper[4556]: I0218 09:05:00.714277 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:00 crc kubenswrapper[4556]: I0218 09:05:00.714291 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:00 crc kubenswrapper[4556]: I0218 09:05:00.714306 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:00Z","lastTransitionTime":"2026-02-18T09:05:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:00 crc kubenswrapper[4556]: I0218 09:05:00.815989 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:00 crc kubenswrapper[4556]: I0218 09:05:00.816020 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:00 crc kubenswrapper[4556]: I0218 09:05:00.816028 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:00 crc kubenswrapper[4556]: I0218 09:05:00.816041 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:00 crc kubenswrapper[4556]: I0218 09:05:00.816050 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:00Z","lastTransitionTime":"2026-02-18T09:05:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:00 crc kubenswrapper[4556]: I0218 09:05:00.918235 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:00 crc kubenswrapper[4556]: I0218 09:05:00.918262 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:00 crc kubenswrapper[4556]: I0218 09:05:00.918270 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:00 crc kubenswrapper[4556]: I0218 09:05:00.918280 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:00 crc kubenswrapper[4556]: I0218 09:05:00.918288 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:00Z","lastTransitionTime":"2026-02-18T09:05:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:01 crc kubenswrapper[4556]: I0218 09:05:01.020317 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:01 crc kubenswrapper[4556]: I0218 09:05:01.020363 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:01 crc kubenswrapper[4556]: I0218 09:05:01.020372 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:01 crc kubenswrapper[4556]: I0218 09:05:01.020387 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:01 crc kubenswrapper[4556]: I0218 09:05:01.020404 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:01Z","lastTransitionTime":"2026-02-18T09:05:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:01 crc kubenswrapper[4556]: I0218 09:05:01.122586 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:01 crc kubenswrapper[4556]: I0218 09:05:01.122622 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:01 crc kubenswrapper[4556]: I0218 09:05:01.122630 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:01 crc kubenswrapper[4556]: I0218 09:05:01.122645 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:01 crc kubenswrapper[4556]: I0218 09:05:01.122654 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:01Z","lastTransitionTime":"2026-02-18T09:05:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:01 crc kubenswrapper[4556]: I0218 09:05:01.224762 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:01 crc kubenswrapper[4556]: I0218 09:05:01.224793 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:01 crc kubenswrapper[4556]: I0218 09:05:01.224801 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:01 crc kubenswrapper[4556]: I0218 09:05:01.224812 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:01 crc kubenswrapper[4556]: I0218 09:05:01.224821 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:01Z","lastTransitionTime":"2026-02-18T09:05:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:01 crc kubenswrapper[4556]: I0218 09:05:01.281915 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:05:01 crc kubenswrapper[4556]: I0218 09:05:01.281978 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:05:01 crc kubenswrapper[4556]: E0218 09:05:01.282030 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 09:05:01 crc kubenswrapper[4556]: E0218 09:05:01.282100 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 09:05:01 crc kubenswrapper[4556]: I0218 09:05:01.282173 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:05:01 crc kubenswrapper[4556]: E0218 09:05:01.282241 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 09:05:01 crc kubenswrapper[4556]: I0218 09:05:01.283256 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-18 16:24:38.056924748 +0000 UTC Feb 18 09:05:01 crc kubenswrapper[4556]: I0218 09:05:01.326615 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:01 crc kubenswrapper[4556]: I0218 09:05:01.326644 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:01 crc kubenswrapper[4556]: I0218 09:05:01.326654 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:01 crc kubenswrapper[4556]: I0218 09:05:01.326667 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:01 crc kubenswrapper[4556]: I0218 09:05:01.326678 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:01Z","lastTransitionTime":"2026-02-18T09:05:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:01 crc kubenswrapper[4556]: I0218 09:05:01.428605 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:01 crc kubenswrapper[4556]: I0218 09:05:01.428638 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:01 crc kubenswrapper[4556]: I0218 09:05:01.428656 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:01 crc kubenswrapper[4556]: I0218 09:05:01.428665 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:01 crc kubenswrapper[4556]: I0218 09:05:01.428673 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:01Z","lastTransitionTime":"2026-02-18T09:05:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:01 crc kubenswrapper[4556]: I0218 09:05:01.530474 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:01 crc kubenswrapper[4556]: I0218 09:05:01.530499 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:01 crc kubenswrapper[4556]: I0218 09:05:01.530507 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:01 crc kubenswrapper[4556]: I0218 09:05:01.532268 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:01 crc kubenswrapper[4556]: I0218 09:05:01.532283 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:01Z","lastTransitionTime":"2026-02-18T09:05:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:01 crc kubenswrapper[4556]: I0218 09:05:01.633997 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:01 crc kubenswrapper[4556]: I0218 09:05:01.634026 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:01 crc kubenswrapper[4556]: I0218 09:05:01.634035 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:01 crc kubenswrapper[4556]: I0218 09:05:01.634046 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:01 crc kubenswrapper[4556]: I0218 09:05:01.634053 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:01Z","lastTransitionTime":"2026-02-18T09:05:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:01 crc kubenswrapper[4556]: I0218 09:05:01.738581 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:01 crc kubenswrapper[4556]: I0218 09:05:01.738620 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:01 crc kubenswrapper[4556]: I0218 09:05:01.738630 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:01 crc kubenswrapper[4556]: I0218 09:05:01.738646 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:01 crc kubenswrapper[4556]: I0218 09:05:01.738656 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:01Z","lastTransitionTime":"2026-02-18T09:05:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:01 crc kubenswrapper[4556]: I0218 09:05:01.840810 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:01 crc kubenswrapper[4556]: I0218 09:05:01.840845 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:01 crc kubenswrapper[4556]: I0218 09:05:01.840857 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:01 crc kubenswrapper[4556]: I0218 09:05:01.840869 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:01 crc kubenswrapper[4556]: I0218 09:05:01.840879 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:01Z","lastTransitionTime":"2026-02-18T09:05:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:01 crc kubenswrapper[4556]: I0218 09:05:01.942391 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:01 crc kubenswrapper[4556]: I0218 09:05:01.942427 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:01 crc kubenswrapper[4556]: I0218 09:05:01.942438 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:01 crc kubenswrapper[4556]: I0218 09:05:01.942451 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:01 crc kubenswrapper[4556]: I0218 09:05:01.942503 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:01Z","lastTransitionTime":"2026-02-18T09:05:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:02 crc kubenswrapper[4556]: I0218 09:05:02.044540 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:02 crc kubenswrapper[4556]: I0218 09:05:02.044570 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:02 crc kubenswrapper[4556]: I0218 09:05:02.044579 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:02 crc kubenswrapper[4556]: I0218 09:05:02.044592 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:02 crc kubenswrapper[4556]: I0218 09:05:02.044603 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:02Z","lastTransitionTime":"2026-02-18T09:05:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:02 crc kubenswrapper[4556]: I0218 09:05:02.146374 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:02 crc kubenswrapper[4556]: I0218 09:05:02.146412 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:02 crc kubenswrapper[4556]: I0218 09:05:02.146422 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:02 crc kubenswrapper[4556]: I0218 09:05:02.146432 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:02 crc kubenswrapper[4556]: I0218 09:05:02.146439 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:02Z","lastTransitionTime":"2026-02-18T09:05:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:02 crc kubenswrapper[4556]: I0218 09:05:02.247780 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:02 crc kubenswrapper[4556]: I0218 09:05:02.247811 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:02 crc kubenswrapper[4556]: I0218 09:05:02.247820 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:02 crc kubenswrapper[4556]: I0218 09:05:02.247832 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:02 crc kubenswrapper[4556]: I0218 09:05:02.247841 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:02Z","lastTransitionTime":"2026-02-18T09:05:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:02 crc kubenswrapper[4556]: I0218 09:05:02.282339 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cgwv9" Feb 18 09:05:02 crc kubenswrapper[4556]: E0218 09:05:02.282445 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cgwv9" podUID="d2f0e36c-cb29-4b06-bb22-6afd59466cab" Feb 18 09:05:02 crc kubenswrapper[4556]: I0218 09:05:02.283320 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-30 03:04:19.505591524 +0000 UTC Feb 18 09:05:02 crc kubenswrapper[4556]: I0218 09:05:02.349811 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:02 crc kubenswrapper[4556]: I0218 09:05:02.349837 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:02 crc kubenswrapper[4556]: I0218 09:05:02.349846 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:02 crc kubenswrapper[4556]: I0218 09:05:02.349855 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:02 crc kubenswrapper[4556]: I0218 09:05:02.349863 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:02Z","lastTransitionTime":"2026-02-18T09:05:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:02 crc kubenswrapper[4556]: I0218 09:05:02.451066 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:02 crc kubenswrapper[4556]: I0218 09:05:02.451097 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:02 crc kubenswrapper[4556]: I0218 09:05:02.451106 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:02 crc kubenswrapper[4556]: I0218 09:05:02.451117 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:02 crc kubenswrapper[4556]: I0218 09:05:02.451126 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:02Z","lastTransitionTime":"2026-02-18T09:05:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:02 crc kubenswrapper[4556]: I0218 09:05:02.552928 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:02 crc kubenswrapper[4556]: I0218 09:05:02.552957 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:02 crc kubenswrapper[4556]: I0218 09:05:02.552965 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:02 crc kubenswrapper[4556]: I0218 09:05:02.552975 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:02 crc kubenswrapper[4556]: I0218 09:05:02.552983 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:02Z","lastTransitionTime":"2026-02-18T09:05:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:02 crc kubenswrapper[4556]: I0218 09:05:02.654432 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:02 crc kubenswrapper[4556]: I0218 09:05:02.654467 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:02 crc kubenswrapper[4556]: I0218 09:05:02.654476 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:02 crc kubenswrapper[4556]: I0218 09:05:02.654488 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:02 crc kubenswrapper[4556]: I0218 09:05:02.654497 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:02Z","lastTransitionTime":"2026-02-18T09:05:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:02 crc kubenswrapper[4556]: I0218 09:05:02.756340 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:02 crc kubenswrapper[4556]: I0218 09:05:02.756411 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:02 crc kubenswrapper[4556]: I0218 09:05:02.756423 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:02 crc kubenswrapper[4556]: I0218 09:05:02.756448 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:02 crc kubenswrapper[4556]: I0218 09:05:02.756475 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:02Z","lastTransitionTime":"2026-02-18T09:05:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:02 crc kubenswrapper[4556]: I0218 09:05:02.857804 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:02 crc kubenswrapper[4556]: I0218 09:05:02.857854 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:02 crc kubenswrapper[4556]: I0218 09:05:02.857865 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:02 crc kubenswrapper[4556]: I0218 09:05:02.857881 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:02 crc kubenswrapper[4556]: I0218 09:05:02.857893 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:02Z","lastTransitionTime":"2026-02-18T09:05:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:02 crc kubenswrapper[4556]: I0218 09:05:02.959382 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:02 crc kubenswrapper[4556]: I0218 09:05:02.959428 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:02 crc kubenswrapper[4556]: I0218 09:05:02.959441 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:02 crc kubenswrapper[4556]: I0218 09:05:02.959467 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:02 crc kubenswrapper[4556]: I0218 09:05:02.959477 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:02Z","lastTransitionTime":"2026-02-18T09:05:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.061619 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.061657 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.061666 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.061681 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.061690 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:03Z","lastTransitionTime":"2026-02-18T09:05:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.163116 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.163164 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.163187 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.163197 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.163207 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:03Z","lastTransitionTime":"2026-02-18T09:05:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.267482 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.267507 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.267516 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.267527 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.267535 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:03Z","lastTransitionTime":"2026-02-18T09:05:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.282209 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.282259 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.282280 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:05:03 crc kubenswrapper[4556]: E0218 09:05:03.282498 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 09:05:03 crc kubenswrapper[4556]: E0218 09:05:03.282569 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 09:05:03 crc kubenswrapper[4556]: E0218 09:05:03.282397 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.283447 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-16 02:03:12.169144788 +0000 UTC Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.292572 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qfksk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e226b189-cfa9-47fc-b94d-19b5cbe0859f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc2cb92a7735d16b8e5918797e52e1e488d9261a0418ed5ebbb94c033ca9328a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2fx4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qfksk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:03Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.301146 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dac7f27-d3d1-4778-9e54-f273035a1d37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18da54fe071964d8aae8a977f5ee209c31f1c75e1041c44f8d755f58f3e39d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d41b85076e695bbc51a41fce200e0e47c099d88b4d825b4d133d33ed96237e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f76hs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:03Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.310784 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sgcq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9945b792-abbd-4103-a7b4-9496e1cc1b56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5dd402827182bc79283f1ea1c355b60c4ced53ffa9ce6781b0a0db3109359a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac0d9d7b9bc867b32cf6cbeb804c08e67df757e4b22914e2da69a376511cf75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac0d9d7b9bc867b32cf6cbeb804c08e67df757e4b22914e2da69a376511cf75c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b807d53d620e44d3196f68492b9f9665593adac804bc7b91569796e5594c4ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b807d53d620e44d3196f68492b9f9665593adac804bc7b91569796e5594c4ebb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebc044bbd5b0c1e725e6ed46ec2e0015b254070b7518d20815539dcf8c6e4e45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ebc044bbd5b0c1e725e6ed46ec2e0015b254070b7518d20815539dcf8c6e4e45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sgcq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:03Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.317592 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cgwv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2f0e36c-cb29-4b06-bb22-6afd59466cab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpldv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpldv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:40Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cgwv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:03Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.326017 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"302eb215-a4c2-4324-836d-f37debfd620f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a99dc228d1c4cb8aeab2140ec5f37358b65684739f833fbde45822782430a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14f794063db9d33b82b79acc7134b5a2a3435e09d7cffe49d49251d1f7d20402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5eb3d3728da247903bece56bc2481547317e386e9e97b6c69926a3b55dc89b8e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b42d13c1767c29d22e52537ff45561a0fbf8f2069b2c3612e903e8c127e12485\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e865e6a50c03519ce22958ac4fcbb911fc140c330b6d7ffe5c9d7dd381e96\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T09:04:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0218 09:04:15.143085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 09:04:15.144650 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2160144045/tls.crt::/tmp/serving-cert-2160144045/tls.key\\\\\\\"\\\\nI0218 09:04:20.680799 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 09:04:20.682555 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 09:04:20.682573 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 09:04:20.682590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 09:04:20.682594 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 09:04:20.686822 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 09:04:20.686847 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686852 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686856 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 09:04:20.686859 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 09:04:20.686861 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 09:04:20.686864 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 09:04:20.688257 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 09:04:20.690726 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://974e17ae7082631df2d1bcee60c8a5f142bae2659e9d200d537a4d650f401b4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:03Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.336290 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:03Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.345541 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47341f31f94eaf5dd86efbe78ff62873e733b1f86ea895f281b6797770519459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:03Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.351760 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5vcgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d50c39aa-f956-441a-8abe-1d7247a7fd86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ddc5d42163d1fa86f1a59e224adac434c990224e839ef3cb12c6d7689e2b439\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4llw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5vcgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:03Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.363499 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09971522-31bc-4e10-8831-4fb62675e220\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9170bdcbaf1065995d03e6448e40bdced0741c716c9e81f37a72b015132f1161\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c526c0613f19e1a1a584b7129bd56001faf08d0b97fd77feeb69f764ed01f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cd6e758ee396c9000b1bdb6ded2176ec803fa9da6201cdd9b8146940a2c00cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b0412f3330999c3348ac726ac5a0c2d541571f4b13786ca24cb3109a44aa653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://318cb69e50df4c14df9bf5ef57d5c381421a9bdb72135c9b9620f55c0547b410\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0f15d331af9bc2ca3082c1573dd167c66773eb390593844e681f8d0b455d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03acfee4df70e62846f679407e6f9960e44946d5fd4b71d7ea1011729eb8c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b03acfee4df70e62846f679407e6f9960e44946d5fd4b71d7ea1011729eb8c05\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T09:04:48Z\\\",\\\"message\\\":\\\"ddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-controller-manager/kube-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager/kube-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.36\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0218 09:04:48.895203 6239 lb_config.go:1031] Cluster endpoints for openshift-machine-api/machine-api-controllers for network=default are: map[]\\\\nI0218 09:04:48.895220 6239 services_controller.go:443] Built service openshift-machine-api/machine-api-controllers LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.167\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8441, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]s\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-b7fsj_openshift-ovn-kubernetes(09971522-31bc-4e10-8831-4fb62675e220)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aec6030df0a59ad8902d5c694d5cd30d249ca1e895155c57a458e359fa531b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7fsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:03Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.369592 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.369686 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.369751 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.369819 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.369888 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:03Z","lastTransitionTime":"2026-02-18T09:05:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.371579 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4e7c95a-56e9-4e22-83dd-54962e1233b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52ea6a334e8cad154e46eb2f917cdac914557556bae88f922433a9545cb29eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28a600e8daaa2796b2f0cc6ce4112f8d41c611b5b6ca5a7eb0f260dab615caa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1fa733477ee6330835d52333d6e9588626b6fc7ae08ae450dd2e1821447e995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfbda9799e2d2bc16964e79b90fb7087f9e6893d867711612c63aaaa3dd2b141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfbda9799e2d2bc16964e79b90fb7087f9e6893d867711612c63aaaa3dd2b141\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:03Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.377977 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jflp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689d3d82-4071-4fe5-b3e3-b5b4f511580a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d56d65964aaf12feaaa3e5fa3ca779bffdb13fcebfc9f027b563c7dd01f4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bm85s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jflp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:03Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.385836 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:03Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.393845 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d653c21d3d7c875bed37ed6da700dc8c801b95c80d7e361121dfbbd3d05a922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5211d0cd5172d45b92b003e2bf97d7969a65ea448b0491f8f6ea67364925ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:03Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.402592 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6387f58543cf2601a53b0bbe9be3ec7816ba90d5ffdc4465594b37ccac0f6dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:03Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.410008 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:03Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.417075 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bp8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"804f8152-54ae-4486-92e6-f9595c035820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb617e46b8522d73dabef4bbc660e41b737f6e6e62f7ce42dad722490a1041dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phr7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://104b8ff3c41d4304d414e58303c23792a7f0ba580f73a2a8678044741af71380\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phr7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bp8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:03Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.429230 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ae60f3-adab-4740-821a-1570059c7101\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f629262e27020512301a7b53ee836ee98e1e7cdee7a52b0ed506bbaaebd823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f59306c0d55e9d00714ce2ee2c1ed6f2ed33d8b21eecde490e86a93335a14f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c87bc671742a47de6c8e68aaf0c8645f9a8cae7b3d52264f36ae3827b3a8d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2678001ff826712da6716541210f018c74fd3d0218b388d2e7a6f16a1f4027\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6f37933281c927328ecfaf842d97ee8361726768f393726cb39517d2bdc9425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:03Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.437343 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8f6014-2544-4173-a94d-441cf42ba891\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77b0269625b7c6baab44a40373618f4ccfe9e1a028a1bc746b7d60c20ac4efaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac09c421b3101887bfb83107dbe26f2b5856d9b32a9cea92f6233fb6ae218cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce79cfe0242ecf403f989dc6b1c1b8be5ab54eee75ac413a8f9d03cef8bca0ed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://095ecf7e8763fa098b06eb9a8aac9f39451f7762bf20196e0803bbc841610e02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:03Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.471972 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.471997 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.472006 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.472020 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.472028 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:03Z","lastTransitionTime":"2026-02-18T09:05:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.573953 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.573989 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.574001 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.574016 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.574027 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:03Z","lastTransitionTime":"2026-02-18T09:05:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.675839 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.675871 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.675882 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.675895 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.675904 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:03Z","lastTransitionTime":"2026-02-18T09:05:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.778202 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.778234 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.778242 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.778255 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.778264 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:03Z","lastTransitionTime":"2026-02-18T09:05:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.880659 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.880821 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.880887 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.880963 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.881021 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:03Z","lastTransitionTime":"2026-02-18T09:05:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.983250 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.983277 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.983286 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.983297 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:03 crc kubenswrapper[4556]: I0218 09:05:03.983305 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:03Z","lastTransitionTime":"2026-02-18T09:05:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.085094 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.085128 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.085175 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.085188 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.085197 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:04Z","lastTransitionTime":"2026-02-18T09:05:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.186645 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.186700 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.186709 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.186723 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.186732 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:04Z","lastTransitionTime":"2026-02-18T09:05:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.281607 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cgwv9" Feb 18 09:05:04 crc kubenswrapper[4556]: E0218 09:05:04.281724 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cgwv9" podUID="d2f0e36c-cb29-4b06-bb22-6afd59466cab" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.283732 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-20 23:55:09.374559139 +0000 UTC Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.288796 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.288841 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.288851 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.288865 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.288874 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:04Z","lastTransitionTime":"2026-02-18T09:05:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.390758 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.390791 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.390800 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.390812 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.390825 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:04Z","lastTransitionTime":"2026-02-18T09:05:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.492623 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.492647 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.492655 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.492666 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.492674 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:04Z","lastTransitionTime":"2026-02-18T09:05:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.593661 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.593688 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.593696 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.593707 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.593715 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:04Z","lastTransitionTime":"2026-02-18T09:05:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.695626 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.695658 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.695668 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.695681 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.695691 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:04Z","lastTransitionTime":"2026-02-18T09:05:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.725413 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.725438 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.725447 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.725469 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.725476 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:04Z","lastTransitionTime":"2026-02-18T09:05:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:04 crc kubenswrapper[4556]: E0218 09:05:04.734297 4556 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:05:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:05:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:05:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:05:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8758c02e-ef59-46a7-908c-d5c97feb8ceb\\\",\\\"systemUUID\\\":\\\"0f904183-79bf-4790-8551-43a6eb3adbe4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:04Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.736975 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.737007 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.737016 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.737029 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.737038 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:04Z","lastTransitionTime":"2026-02-18T09:05:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:04 crc kubenswrapper[4556]: E0218 09:05:04.745298 4556 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:05:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:05:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:05:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:05:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8758c02e-ef59-46a7-908c-d5c97feb8ceb\\\",\\\"systemUUID\\\":\\\"0f904183-79bf-4790-8551-43a6eb3adbe4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:04Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.747927 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.747955 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.747963 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.747991 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.748002 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:04Z","lastTransitionTime":"2026-02-18T09:05:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:04 crc kubenswrapper[4556]: E0218 09:05:04.756509 4556 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:05:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:05:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:05:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:05:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8758c02e-ef59-46a7-908c-d5c97feb8ceb\\\",\\\"systemUUID\\\":\\\"0f904183-79bf-4790-8551-43a6eb3adbe4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:04Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.759196 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.759222 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.759230 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.759276 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.759284 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:04Z","lastTransitionTime":"2026-02-18T09:05:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:04 crc kubenswrapper[4556]: E0218 09:05:04.766875 4556 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:05:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:05:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:05:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:05:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8758c02e-ef59-46a7-908c-d5c97feb8ceb\\\",\\\"systemUUID\\\":\\\"0f904183-79bf-4790-8551-43a6eb3adbe4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:04Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.769022 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.769062 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.769070 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.769081 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.769088 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:04Z","lastTransitionTime":"2026-02-18T09:05:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:04 crc kubenswrapper[4556]: E0218 09:05:04.777219 4556 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:05:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:05:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:05:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:05:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8758c02e-ef59-46a7-908c-d5c97feb8ceb\\\",\\\"systemUUID\\\":\\\"0f904183-79bf-4790-8551-43a6eb3adbe4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:04Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:04 crc kubenswrapper[4556]: E0218 09:05:04.777319 4556 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.797612 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.797634 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.797641 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.797668 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.797678 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:04Z","lastTransitionTime":"2026-02-18T09:05:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.899850 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.899880 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.899888 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.899899 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:04 crc kubenswrapper[4556]: I0218 09:05:04.899907 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:04Z","lastTransitionTime":"2026-02-18T09:05:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:05 crc kubenswrapper[4556]: I0218 09:05:05.001681 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:05 crc kubenswrapper[4556]: I0218 09:05:05.001710 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:05 crc kubenswrapper[4556]: I0218 09:05:05.001718 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:05 crc kubenswrapper[4556]: I0218 09:05:05.001747 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:05 crc kubenswrapper[4556]: I0218 09:05:05.001757 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:05Z","lastTransitionTime":"2026-02-18T09:05:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:05 crc kubenswrapper[4556]: I0218 09:05:05.103445 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:05 crc kubenswrapper[4556]: I0218 09:05:05.103478 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:05 crc kubenswrapper[4556]: I0218 09:05:05.103486 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:05 crc kubenswrapper[4556]: I0218 09:05:05.103495 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:05 crc kubenswrapper[4556]: I0218 09:05:05.103501 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:05Z","lastTransitionTime":"2026-02-18T09:05:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:05 crc kubenswrapper[4556]: I0218 09:05:05.205820 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:05 crc kubenswrapper[4556]: I0218 09:05:05.205849 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:05 crc kubenswrapper[4556]: I0218 09:05:05.205857 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:05 crc kubenswrapper[4556]: I0218 09:05:05.205866 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:05 crc kubenswrapper[4556]: I0218 09:05:05.205873 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:05Z","lastTransitionTime":"2026-02-18T09:05:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:05 crc kubenswrapper[4556]: I0218 09:05:05.281711 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:05:05 crc kubenswrapper[4556]: E0218 09:05:05.281805 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 09:05:05 crc kubenswrapper[4556]: I0218 09:05:05.282422 4556 scope.go:117] "RemoveContainer" containerID="b03acfee4df70e62846f679407e6f9960e44946d5fd4b71d7ea1011729eb8c05" Feb 18 09:05:05 crc kubenswrapper[4556]: E0218 09:05:05.282597 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-b7fsj_openshift-ovn-kubernetes(09971522-31bc-4e10-8831-4fb62675e220)\"" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" podUID="09971522-31bc-4e10-8831-4fb62675e220" Feb 18 09:05:05 crc kubenswrapper[4556]: I0218 09:05:05.282715 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:05:05 crc kubenswrapper[4556]: E0218 09:05:05.282765 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 09:05:05 crc kubenswrapper[4556]: I0218 09:05:05.282871 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:05:05 crc kubenswrapper[4556]: E0218 09:05:05.282916 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 09:05:05 crc kubenswrapper[4556]: I0218 09:05:05.283840 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-18 09:49:43.812549269 +0000 UTC Feb 18 09:05:05 crc kubenswrapper[4556]: I0218 09:05:05.306955 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:05 crc kubenswrapper[4556]: I0218 09:05:05.306993 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:05 crc kubenswrapper[4556]: I0218 09:05:05.307001 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:05 crc kubenswrapper[4556]: I0218 09:05:05.307009 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:05 crc kubenswrapper[4556]: I0218 09:05:05.307016 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:05Z","lastTransitionTime":"2026-02-18T09:05:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:05 crc kubenswrapper[4556]: I0218 09:05:05.410182 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:05 crc kubenswrapper[4556]: I0218 09:05:05.410220 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:05 crc kubenswrapper[4556]: I0218 09:05:05.410235 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:05 crc kubenswrapper[4556]: I0218 09:05:05.410248 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:05 crc kubenswrapper[4556]: I0218 09:05:05.410256 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:05Z","lastTransitionTime":"2026-02-18T09:05:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:05 crc kubenswrapper[4556]: I0218 09:05:05.511928 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:05 crc kubenswrapper[4556]: I0218 09:05:05.511955 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:05 crc kubenswrapper[4556]: I0218 09:05:05.511979 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:05 crc kubenswrapper[4556]: I0218 09:05:05.511990 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:05 crc kubenswrapper[4556]: I0218 09:05:05.511998 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:05Z","lastTransitionTime":"2026-02-18T09:05:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:05 crc kubenswrapper[4556]: I0218 09:05:05.613022 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:05 crc kubenswrapper[4556]: I0218 09:05:05.613047 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:05 crc kubenswrapper[4556]: I0218 09:05:05.613054 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:05 crc kubenswrapper[4556]: I0218 09:05:05.613065 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:05 crc kubenswrapper[4556]: I0218 09:05:05.613072 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:05Z","lastTransitionTime":"2026-02-18T09:05:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:05 crc kubenswrapper[4556]: I0218 09:05:05.714293 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:05 crc kubenswrapper[4556]: I0218 09:05:05.714336 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:05 crc kubenswrapper[4556]: I0218 09:05:05.714345 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:05 crc kubenswrapper[4556]: I0218 09:05:05.714355 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:05 crc kubenswrapper[4556]: I0218 09:05:05.714362 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:05Z","lastTransitionTime":"2026-02-18T09:05:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:05 crc kubenswrapper[4556]: I0218 09:05:05.816182 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:05 crc kubenswrapper[4556]: I0218 09:05:05.816204 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:05 crc kubenswrapper[4556]: I0218 09:05:05.816212 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:05 crc kubenswrapper[4556]: I0218 09:05:05.816222 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:05 crc kubenswrapper[4556]: I0218 09:05:05.816229 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:05Z","lastTransitionTime":"2026-02-18T09:05:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:05 crc kubenswrapper[4556]: I0218 09:05:05.918024 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:05 crc kubenswrapper[4556]: I0218 09:05:05.918057 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:05 crc kubenswrapper[4556]: I0218 09:05:05.918066 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:05 crc kubenswrapper[4556]: I0218 09:05:05.918078 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:05 crc kubenswrapper[4556]: I0218 09:05:05.918087 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:05Z","lastTransitionTime":"2026-02-18T09:05:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:06 crc kubenswrapper[4556]: I0218 09:05:06.019925 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:06 crc kubenswrapper[4556]: I0218 09:05:06.019947 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:06 crc kubenswrapper[4556]: I0218 09:05:06.019956 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:06 crc kubenswrapper[4556]: I0218 09:05:06.019965 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:06 crc kubenswrapper[4556]: I0218 09:05:06.019973 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:06Z","lastTransitionTime":"2026-02-18T09:05:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:06 crc kubenswrapper[4556]: I0218 09:05:06.121791 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:06 crc kubenswrapper[4556]: I0218 09:05:06.121828 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:06 crc kubenswrapper[4556]: I0218 09:05:06.121838 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:06 crc kubenswrapper[4556]: I0218 09:05:06.121850 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:06 crc kubenswrapper[4556]: I0218 09:05:06.121859 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:06Z","lastTransitionTime":"2026-02-18T09:05:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:06 crc kubenswrapper[4556]: I0218 09:05:06.223796 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:06 crc kubenswrapper[4556]: I0218 09:05:06.223833 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:06 crc kubenswrapper[4556]: I0218 09:05:06.223842 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:06 crc kubenswrapper[4556]: I0218 09:05:06.223855 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:06 crc kubenswrapper[4556]: I0218 09:05:06.223863 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:06Z","lastTransitionTime":"2026-02-18T09:05:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:06 crc kubenswrapper[4556]: I0218 09:05:06.281375 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cgwv9" Feb 18 09:05:06 crc kubenswrapper[4556]: E0218 09:05:06.281494 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cgwv9" podUID="d2f0e36c-cb29-4b06-bb22-6afd59466cab" Feb 18 09:05:06 crc kubenswrapper[4556]: I0218 09:05:06.284637 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-25 22:42:42.88378195 +0000 UTC Feb 18 09:05:06 crc kubenswrapper[4556]: I0218 09:05:06.325737 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:06 crc kubenswrapper[4556]: I0218 09:05:06.325777 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:06 crc kubenswrapper[4556]: I0218 09:05:06.325788 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:06 crc kubenswrapper[4556]: I0218 09:05:06.325798 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:06 crc kubenswrapper[4556]: I0218 09:05:06.325807 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:06Z","lastTransitionTime":"2026-02-18T09:05:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:06 crc kubenswrapper[4556]: I0218 09:05:06.428352 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:06 crc kubenswrapper[4556]: I0218 09:05:06.428387 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:06 crc kubenswrapper[4556]: I0218 09:05:06.428396 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:06 crc kubenswrapper[4556]: I0218 09:05:06.428408 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:06 crc kubenswrapper[4556]: I0218 09:05:06.428416 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:06Z","lastTransitionTime":"2026-02-18T09:05:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:06 crc kubenswrapper[4556]: I0218 09:05:06.530736 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:06 crc kubenswrapper[4556]: I0218 09:05:06.530771 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:06 crc kubenswrapper[4556]: I0218 09:05:06.530782 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:06 crc kubenswrapper[4556]: I0218 09:05:06.530793 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:06 crc kubenswrapper[4556]: I0218 09:05:06.530802 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:06Z","lastTransitionTime":"2026-02-18T09:05:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:06 crc kubenswrapper[4556]: I0218 09:05:06.633126 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:06 crc kubenswrapper[4556]: I0218 09:05:06.633187 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:06 crc kubenswrapper[4556]: I0218 09:05:06.633198 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:06 crc kubenswrapper[4556]: I0218 09:05:06.633213 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:06 crc kubenswrapper[4556]: I0218 09:05:06.633223 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:06Z","lastTransitionTime":"2026-02-18T09:05:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:06 crc kubenswrapper[4556]: I0218 09:05:06.735579 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:06 crc kubenswrapper[4556]: I0218 09:05:06.735605 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:06 crc kubenswrapper[4556]: I0218 09:05:06.735614 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:06 crc kubenswrapper[4556]: I0218 09:05:06.735626 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:06 crc kubenswrapper[4556]: I0218 09:05:06.735634 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:06Z","lastTransitionTime":"2026-02-18T09:05:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:06 crc kubenswrapper[4556]: I0218 09:05:06.837310 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:06 crc kubenswrapper[4556]: I0218 09:05:06.837340 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:06 crc kubenswrapper[4556]: I0218 09:05:06.837347 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:06 crc kubenswrapper[4556]: I0218 09:05:06.837360 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:06 crc kubenswrapper[4556]: I0218 09:05:06.837370 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:06Z","lastTransitionTime":"2026-02-18T09:05:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:06 crc kubenswrapper[4556]: I0218 09:05:06.939570 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:06 crc kubenswrapper[4556]: I0218 09:05:06.939608 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:06 crc kubenswrapper[4556]: I0218 09:05:06.939616 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:06 crc kubenswrapper[4556]: I0218 09:05:06.939630 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:06 crc kubenswrapper[4556]: I0218 09:05:06.939640 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:06Z","lastTransitionTime":"2026-02-18T09:05:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:07 crc kubenswrapper[4556]: I0218 09:05:07.041561 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:07 crc kubenswrapper[4556]: I0218 09:05:07.041593 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:07 crc kubenswrapper[4556]: I0218 09:05:07.041603 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:07 crc kubenswrapper[4556]: I0218 09:05:07.041615 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:07 crc kubenswrapper[4556]: I0218 09:05:07.041625 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:07Z","lastTransitionTime":"2026-02-18T09:05:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:07 crc kubenswrapper[4556]: I0218 09:05:07.143987 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:07 crc kubenswrapper[4556]: I0218 09:05:07.144019 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:07 crc kubenswrapper[4556]: I0218 09:05:07.144028 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:07 crc kubenswrapper[4556]: I0218 09:05:07.144042 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:07 crc kubenswrapper[4556]: I0218 09:05:07.144051 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:07Z","lastTransitionTime":"2026-02-18T09:05:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:07 crc kubenswrapper[4556]: I0218 09:05:07.245760 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:07 crc kubenswrapper[4556]: I0218 09:05:07.245904 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:07 crc kubenswrapper[4556]: I0218 09:05:07.245982 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:07 crc kubenswrapper[4556]: I0218 09:05:07.246043 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:07 crc kubenswrapper[4556]: I0218 09:05:07.246099 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:07Z","lastTransitionTime":"2026-02-18T09:05:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:07 crc kubenswrapper[4556]: I0218 09:05:07.282063 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:05:07 crc kubenswrapper[4556]: I0218 09:05:07.282105 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:05:07 crc kubenswrapper[4556]: I0218 09:05:07.282124 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:05:07 crc kubenswrapper[4556]: E0218 09:05:07.282220 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 09:05:07 crc kubenswrapper[4556]: E0218 09:05:07.282248 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 09:05:07 crc kubenswrapper[4556]: E0218 09:05:07.282353 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 09:05:07 crc kubenswrapper[4556]: I0218 09:05:07.284725 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-07 04:14:22.703556858 +0000 UTC Feb 18 09:05:07 crc kubenswrapper[4556]: I0218 09:05:07.348201 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:07 crc kubenswrapper[4556]: I0218 09:05:07.348226 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:07 crc kubenswrapper[4556]: I0218 09:05:07.348235 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:07 crc kubenswrapper[4556]: I0218 09:05:07.348248 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:07 crc kubenswrapper[4556]: I0218 09:05:07.348257 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:07Z","lastTransitionTime":"2026-02-18T09:05:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:07 crc kubenswrapper[4556]: I0218 09:05:07.449789 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:07 crc kubenswrapper[4556]: I0218 09:05:07.449813 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:07 crc kubenswrapper[4556]: I0218 09:05:07.449821 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:07 crc kubenswrapper[4556]: I0218 09:05:07.449831 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:07 crc kubenswrapper[4556]: I0218 09:05:07.449838 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:07Z","lastTransitionTime":"2026-02-18T09:05:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:07 crc kubenswrapper[4556]: I0218 09:05:07.552065 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:07 crc kubenswrapper[4556]: I0218 09:05:07.552106 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:07 crc kubenswrapper[4556]: I0218 09:05:07.552115 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:07 crc kubenswrapper[4556]: I0218 09:05:07.552131 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:07 crc kubenswrapper[4556]: I0218 09:05:07.552140 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:07Z","lastTransitionTime":"2026-02-18T09:05:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:07 crc kubenswrapper[4556]: I0218 09:05:07.653826 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:07 crc kubenswrapper[4556]: I0218 09:05:07.653851 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:07 crc kubenswrapper[4556]: I0218 09:05:07.653859 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:07 crc kubenswrapper[4556]: I0218 09:05:07.653870 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:07 crc kubenswrapper[4556]: I0218 09:05:07.653878 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:07Z","lastTransitionTime":"2026-02-18T09:05:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:07 crc kubenswrapper[4556]: I0218 09:05:07.755565 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:07 crc kubenswrapper[4556]: I0218 09:05:07.755611 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:07 crc kubenswrapper[4556]: I0218 09:05:07.755621 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:07 crc kubenswrapper[4556]: I0218 09:05:07.755636 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:07 crc kubenswrapper[4556]: I0218 09:05:07.755662 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:07Z","lastTransitionTime":"2026-02-18T09:05:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:07 crc kubenswrapper[4556]: I0218 09:05:07.857386 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:07 crc kubenswrapper[4556]: I0218 09:05:07.857415 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:07 crc kubenswrapper[4556]: I0218 09:05:07.857425 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:07 crc kubenswrapper[4556]: I0218 09:05:07.857437 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:07 crc kubenswrapper[4556]: I0218 09:05:07.857449 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:07Z","lastTransitionTime":"2026-02-18T09:05:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:07 crc kubenswrapper[4556]: I0218 09:05:07.958587 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:07 crc kubenswrapper[4556]: I0218 09:05:07.958640 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:07 crc kubenswrapper[4556]: I0218 09:05:07.958648 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:07 crc kubenswrapper[4556]: I0218 09:05:07.958662 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:07 crc kubenswrapper[4556]: I0218 09:05:07.958673 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:07Z","lastTransitionTime":"2026-02-18T09:05:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:08 crc kubenswrapper[4556]: I0218 09:05:08.060411 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:08 crc kubenswrapper[4556]: I0218 09:05:08.060722 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:08 crc kubenswrapper[4556]: I0218 09:05:08.060739 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:08 crc kubenswrapper[4556]: I0218 09:05:08.060765 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:08 crc kubenswrapper[4556]: I0218 09:05:08.060775 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:08Z","lastTransitionTime":"2026-02-18T09:05:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:08 crc kubenswrapper[4556]: I0218 09:05:08.162907 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:08 crc kubenswrapper[4556]: I0218 09:05:08.162936 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:08 crc kubenswrapper[4556]: I0218 09:05:08.162945 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:08 crc kubenswrapper[4556]: I0218 09:05:08.162958 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:08 crc kubenswrapper[4556]: I0218 09:05:08.162967 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:08Z","lastTransitionTime":"2026-02-18T09:05:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:08 crc kubenswrapper[4556]: I0218 09:05:08.264333 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:08 crc kubenswrapper[4556]: I0218 09:05:08.264438 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:08 crc kubenswrapper[4556]: I0218 09:05:08.264531 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:08 crc kubenswrapper[4556]: I0218 09:05:08.264594 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:08 crc kubenswrapper[4556]: I0218 09:05:08.264658 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:08Z","lastTransitionTime":"2026-02-18T09:05:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:08 crc kubenswrapper[4556]: I0218 09:05:08.281516 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cgwv9" Feb 18 09:05:08 crc kubenswrapper[4556]: E0218 09:05:08.281613 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cgwv9" podUID="d2f0e36c-cb29-4b06-bb22-6afd59466cab" Feb 18 09:05:08 crc kubenswrapper[4556]: I0218 09:05:08.284942 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-13 21:48:56.094602691 +0000 UTC Feb 18 09:05:08 crc kubenswrapper[4556]: I0218 09:05:08.366733 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:08 crc kubenswrapper[4556]: I0218 09:05:08.366760 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:08 crc kubenswrapper[4556]: I0218 09:05:08.366770 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:08 crc kubenswrapper[4556]: I0218 09:05:08.366780 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:08 crc kubenswrapper[4556]: I0218 09:05:08.366789 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:08Z","lastTransitionTime":"2026-02-18T09:05:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:08 crc kubenswrapper[4556]: I0218 09:05:08.468654 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:08 crc kubenswrapper[4556]: I0218 09:05:08.468683 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:08 crc kubenswrapper[4556]: I0218 09:05:08.468692 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:08 crc kubenswrapper[4556]: I0218 09:05:08.468702 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:08 crc kubenswrapper[4556]: I0218 09:05:08.468711 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:08Z","lastTransitionTime":"2026-02-18T09:05:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:08 crc kubenswrapper[4556]: I0218 09:05:08.570765 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:08 crc kubenswrapper[4556]: I0218 09:05:08.570821 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:08 crc kubenswrapper[4556]: I0218 09:05:08.570833 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:08 crc kubenswrapper[4556]: I0218 09:05:08.570848 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:08 crc kubenswrapper[4556]: I0218 09:05:08.570858 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:08Z","lastTransitionTime":"2026-02-18T09:05:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:08 crc kubenswrapper[4556]: I0218 09:05:08.672802 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:08 crc kubenswrapper[4556]: I0218 09:05:08.672832 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:08 crc kubenswrapper[4556]: I0218 09:05:08.672841 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:08 crc kubenswrapper[4556]: I0218 09:05:08.672852 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:08 crc kubenswrapper[4556]: I0218 09:05:08.672861 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:08Z","lastTransitionTime":"2026-02-18T09:05:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:08 crc kubenswrapper[4556]: I0218 09:05:08.775071 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:08 crc kubenswrapper[4556]: I0218 09:05:08.775100 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:08 crc kubenswrapper[4556]: I0218 09:05:08.775108 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:08 crc kubenswrapper[4556]: I0218 09:05:08.775139 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:08 crc kubenswrapper[4556]: I0218 09:05:08.775164 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:08Z","lastTransitionTime":"2026-02-18T09:05:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:08 crc kubenswrapper[4556]: I0218 09:05:08.876562 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:08 crc kubenswrapper[4556]: I0218 09:05:08.876589 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:08 crc kubenswrapper[4556]: I0218 09:05:08.876597 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:08 crc kubenswrapper[4556]: I0218 09:05:08.876624 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:08 crc kubenswrapper[4556]: I0218 09:05:08.876633 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:08Z","lastTransitionTime":"2026-02-18T09:05:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:08 crc kubenswrapper[4556]: I0218 09:05:08.978487 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:08 crc kubenswrapper[4556]: I0218 09:05:08.978519 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:08 crc kubenswrapper[4556]: I0218 09:05:08.978527 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:08 crc kubenswrapper[4556]: I0218 09:05:08.978539 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:08 crc kubenswrapper[4556]: I0218 09:05:08.978548 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:08Z","lastTransitionTime":"2026-02-18T09:05:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:09 crc kubenswrapper[4556]: I0218 09:05:09.080228 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:09 crc kubenswrapper[4556]: I0218 09:05:09.080258 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:09 crc kubenswrapper[4556]: I0218 09:05:09.080266 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:09 crc kubenswrapper[4556]: I0218 09:05:09.080280 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:09 crc kubenswrapper[4556]: I0218 09:05:09.080289 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:09Z","lastTransitionTime":"2026-02-18T09:05:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:09 crc kubenswrapper[4556]: I0218 09:05:09.181928 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:09 crc kubenswrapper[4556]: I0218 09:05:09.181964 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:09 crc kubenswrapper[4556]: I0218 09:05:09.181972 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:09 crc kubenswrapper[4556]: I0218 09:05:09.181982 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:09 crc kubenswrapper[4556]: I0218 09:05:09.181992 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:09Z","lastTransitionTime":"2026-02-18T09:05:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:09 crc kubenswrapper[4556]: I0218 09:05:09.282053 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:05:09 crc kubenswrapper[4556]: I0218 09:05:09.282111 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:05:09 crc kubenswrapper[4556]: E0218 09:05:09.282183 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 09:05:09 crc kubenswrapper[4556]: E0218 09:05:09.282250 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 09:05:09 crc kubenswrapper[4556]: I0218 09:05:09.282294 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:05:09 crc kubenswrapper[4556]: E0218 09:05:09.282395 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 09:05:09 crc kubenswrapper[4556]: I0218 09:05:09.283175 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:09 crc kubenswrapper[4556]: I0218 09:05:09.283206 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:09 crc kubenswrapper[4556]: I0218 09:05:09.283216 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:09 crc kubenswrapper[4556]: I0218 09:05:09.283228 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:09 crc kubenswrapper[4556]: I0218 09:05:09.283238 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:09Z","lastTransitionTime":"2026-02-18T09:05:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:09 crc kubenswrapper[4556]: I0218 09:05:09.285140 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-27 22:20:20.038914547 +0000 UTC Feb 18 09:05:09 crc kubenswrapper[4556]: I0218 09:05:09.385144 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:09 crc kubenswrapper[4556]: I0218 09:05:09.385193 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:09 crc kubenswrapper[4556]: I0218 09:05:09.385202 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:09 crc kubenswrapper[4556]: I0218 09:05:09.385214 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:09 crc kubenswrapper[4556]: I0218 09:05:09.385225 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:09Z","lastTransitionTime":"2026-02-18T09:05:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:09 crc kubenswrapper[4556]: I0218 09:05:09.486609 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:09 crc kubenswrapper[4556]: I0218 09:05:09.486642 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:09 crc kubenswrapper[4556]: I0218 09:05:09.486650 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:09 crc kubenswrapper[4556]: I0218 09:05:09.486663 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:09 crc kubenswrapper[4556]: I0218 09:05:09.486673 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:09Z","lastTransitionTime":"2026-02-18T09:05:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:09 crc kubenswrapper[4556]: I0218 09:05:09.588514 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:09 crc kubenswrapper[4556]: I0218 09:05:09.588549 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:09 crc kubenswrapper[4556]: I0218 09:05:09.588556 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:09 crc kubenswrapper[4556]: I0218 09:05:09.588568 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:09 crc kubenswrapper[4556]: I0218 09:05:09.588579 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:09Z","lastTransitionTime":"2026-02-18T09:05:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:09 crc kubenswrapper[4556]: I0218 09:05:09.691036 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:09 crc kubenswrapper[4556]: I0218 09:05:09.691091 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:09 crc kubenswrapper[4556]: I0218 09:05:09.691100 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:09 crc kubenswrapper[4556]: I0218 09:05:09.691112 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:09 crc kubenswrapper[4556]: I0218 09:05:09.691121 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:09Z","lastTransitionTime":"2026-02-18T09:05:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:09 crc kubenswrapper[4556]: I0218 09:05:09.793356 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:09 crc kubenswrapper[4556]: I0218 09:05:09.793392 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:09 crc kubenswrapper[4556]: I0218 09:05:09.793402 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:09 crc kubenswrapper[4556]: I0218 09:05:09.793417 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:09 crc kubenswrapper[4556]: I0218 09:05:09.793426 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:09Z","lastTransitionTime":"2026-02-18T09:05:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:09 crc kubenswrapper[4556]: I0218 09:05:09.895021 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:09 crc kubenswrapper[4556]: I0218 09:05:09.895055 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:09 crc kubenswrapper[4556]: I0218 09:05:09.895065 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:09 crc kubenswrapper[4556]: I0218 09:05:09.895077 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:09 crc kubenswrapper[4556]: I0218 09:05:09.895087 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:09Z","lastTransitionTime":"2026-02-18T09:05:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:09 crc kubenswrapper[4556]: I0218 09:05:09.996532 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:09 crc kubenswrapper[4556]: I0218 09:05:09.996559 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:09 crc kubenswrapper[4556]: I0218 09:05:09.996568 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:09 crc kubenswrapper[4556]: I0218 09:05:09.996578 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:09 crc kubenswrapper[4556]: I0218 09:05:09.996588 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:09Z","lastTransitionTime":"2026-02-18T09:05:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:10 crc kubenswrapper[4556]: I0218 09:05:10.098142 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:10 crc kubenswrapper[4556]: I0218 09:05:10.098184 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:10 crc kubenswrapper[4556]: I0218 09:05:10.098194 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:10 crc kubenswrapper[4556]: I0218 09:05:10.098207 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:10 crc kubenswrapper[4556]: I0218 09:05:10.098215 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:10Z","lastTransitionTime":"2026-02-18T09:05:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:10 crc kubenswrapper[4556]: I0218 09:05:10.200094 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:10 crc kubenswrapper[4556]: I0218 09:05:10.200124 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:10 crc kubenswrapper[4556]: I0218 09:05:10.200133 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:10 crc kubenswrapper[4556]: I0218 09:05:10.200143 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:10 crc kubenswrapper[4556]: I0218 09:05:10.200174 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:10Z","lastTransitionTime":"2026-02-18T09:05:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:10 crc kubenswrapper[4556]: I0218 09:05:10.281793 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cgwv9" Feb 18 09:05:10 crc kubenswrapper[4556]: E0218 09:05:10.281893 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cgwv9" podUID="d2f0e36c-cb29-4b06-bb22-6afd59466cab" Feb 18 09:05:10 crc kubenswrapper[4556]: I0218 09:05:10.286099 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-24 09:16:42.397498433 +0000 UTC Feb 18 09:05:10 crc kubenswrapper[4556]: I0218 09:05:10.302240 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:10 crc kubenswrapper[4556]: I0218 09:05:10.302263 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:10 crc kubenswrapper[4556]: I0218 09:05:10.302272 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:10 crc kubenswrapper[4556]: I0218 09:05:10.302281 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:10 crc kubenswrapper[4556]: I0218 09:05:10.302290 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:10Z","lastTransitionTime":"2026-02-18T09:05:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:10 crc kubenswrapper[4556]: I0218 09:05:10.403893 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:10 crc kubenswrapper[4556]: I0218 09:05:10.403920 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:10 crc kubenswrapper[4556]: I0218 09:05:10.403929 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:10 crc kubenswrapper[4556]: I0218 09:05:10.403940 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:10 crc kubenswrapper[4556]: I0218 09:05:10.403948 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:10Z","lastTransitionTime":"2026-02-18T09:05:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:10 crc kubenswrapper[4556]: I0218 09:05:10.505426 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:10 crc kubenswrapper[4556]: I0218 09:05:10.505475 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:10 crc kubenswrapper[4556]: I0218 09:05:10.505485 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:10 crc kubenswrapper[4556]: I0218 09:05:10.505500 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:10 crc kubenswrapper[4556]: I0218 09:05:10.505511 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:10Z","lastTransitionTime":"2026-02-18T09:05:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:10 crc kubenswrapper[4556]: I0218 09:05:10.607638 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:10 crc kubenswrapper[4556]: I0218 09:05:10.607660 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:10 crc kubenswrapper[4556]: I0218 09:05:10.607669 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:10 crc kubenswrapper[4556]: I0218 09:05:10.607679 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:10 crc kubenswrapper[4556]: I0218 09:05:10.607687 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:10Z","lastTransitionTime":"2026-02-18T09:05:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:10 crc kubenswrapper[4556]: I0218 09:05:10.709665 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:10 crc kubenswrapper[4556]: I0218 09:05:10.709701 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:10 crc kubenswrapper[4556]: I0218 09:05:10.709710 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:10 crc kubenswrapper[4556]: I0218 09:05:10.709722 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:10 crc kubenswrapper[4556]: I0218 09:05:10.709731 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:10Z","lastTransitionTime":"2026-02-18T09:05:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:10 crc kubenswrapper[4556]: I0218 09:05:10.811561 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:10 crc kubenswrapper[4556]: I0218 09:05:10.811589 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:10 crc kubenswrapper[4556]: I0218 09:05:10.811604 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:10 crc kubenswrapper[4556]: I0218 09:05:10.811617 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:10 crc kubenswrapper[4556]: I0218 09:05:10.811625 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:10Z","lastTransitionTime":"2026-02-18T09:05:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:10 crc kubenswrapper[4556]: I0218 09:05:10.913899 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:10 crc kubenswrapper[4556]: I0218 09:05:10.913942 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:10 crc kubenswrapper[4556]: I0218 09:05:10.913953 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:10 crc kubenswrapper[4556]: I0218 09:05:10.913968 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:10 crc kubenswrapper[4556]: I0218 09:05:10.913978 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:10Z","lastTransitionTime":"2026-02-18T09:05:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:11 crc kubenswrapper[4556]: I0218 09:05:11.015899 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:11 crc kubenswrapper[4556]: I0218 09:05:11.015929 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:11 crc kubenswrapper[4556]: I0218 09:05:11.015938 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:11 crc kubenswrapper[4556]: I0218 09:05:11.015951 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:11 crc kubenswrapper[4556]: I0218 09:05:11.015962 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:11Z","lastTransitionTime":"2026-02-18T09:05:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:11 crc kubenswrapper[4556]: I0218 09:05:11.117633 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:11 crc kubenswrapper[4556]: I0218 09:05:11.117662 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:11 crc kubenswrapper[4556]: I0218 09:05:11.117672 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:11 crc kubenswrapper[4556]: I0218 09:05:11.117683 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:11 crc kubenswrapper[4556]: I0218 09:05:11.117692 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:11Z","lastTransitionTime":"2026-02-18T09:05:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:11 crc kubenswrapper[4556]: I0218 09:05:11.219501 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:11 crc kubenswrapper[4556]: I0218 09:05:11.219533 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:11 crc kubenswrapper[4556]: I0218 09:05:11.219543 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:11 crc kubenswrapper[4556]: I0218 09:05:11.219556 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:11 crc kubenswrapper[4556]: I0218 09:05:11.219566 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:11Z","lastTransitionTime":"2026-02-18T09:05:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:11 crc kubenswrapper[4556]: I0218 09:05:11.282055 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:05:11 crc kubenswrapper[4556]: I0218 09:05:11.282066 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:05:11 crc kubenswrapper[4556]: E0218 09:05:11.282198 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 09:05:11 crc kubenswrapper[4556]: I0218 09:05:11.282173 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:05:11 crc kubenswrapper[4556]: E0218 09:05:11.282268 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 09:05:11 crc kubenswrapper[4556]: E0218 09:05:11.282414 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 09:05:11 crc kubenswrapper[4556]: I0218 09:05:11.286914 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-28 18:42:30.514696922 +0000 UTC Feb 18 09:05:11 crc kubenswrapper[4556]: I0218 09:05:11.321294 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:11 crc kubenswrapper[4556]: I0218 09:05:11.321328 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:11 crc kubenswrapper[4556]: I0218 09:05:11.321338 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:11 crc kubenswrapper[4556]: I0218 09:05:11.321351 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:11 crc kubenswrapper[4556]: I0218 09:05:11.321359 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:11Z","lastTransitionTime":"2026-02-18T09:05:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:11 crc kubenswrapper[4556]: I0218 09:05:11.423366 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:11 crc kubenswrapper[4556]: I0218 09:05:11.423509 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:11 crc kubenswrapper[4556]: I0218 09:05:11.423579 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:11 crc kubenswrapper[4556]: I0218 09:05:11.423653 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:11 crc kubenswrapper[4556]: I0218 09:05:11.423712 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:11Z","lastTransitionTime":"2026-02-18T09:05:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:11 crc kubenswrapper[4556]: I0218 09:05:11.525290 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:11 crc kubenswrapper[4556]: I0218 09:05:11.525480 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:11 crc kubenswrapper[4556]: I0218 09:05:11.525487 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:11 crc kubenswrapper[4556]: I0218 09:05:11.525498 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:11 crc kubenswrapper[4556]: I0218 09:05:11.525506 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:11Z","lastTransitionTime":"2026-02-18T09:05:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:11 crc kubenswrapper[4556]: I0218 09:05:11.627553 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:11 crc kubenswrapper[4556]: I0218 09:05:11.627577 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:11 crc kubenswrapper[4556]: I0218 09:05:11.627585 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:11 crc kubenswrapper[4556]: I0218 09:05:11.627614 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:11 crc kubenswrapper[4556]: I0218 09:05:11.627624 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:11Z","lastTransitionTime":"2026-02-18T09:05:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:11 crc kubenswrapper[4556]: I0218 09:05:11.728896 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:11 crc kubenswrapper[4556]: I0218 09:05:11.728937 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:11 crc kubenswrapper[4556]: I0218 09:05:11.728948 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:11 crc kubenswrapper[4556]: I0218 09:05:11.728963 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:11 crc kubenswrapper[4556]: I0218 09:05:11.728972 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:11Z","lastTransitionTime":"2026-02-18T09:05:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:11 crc kubenswrapper[4556]: I0218 09:05:11.830716 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:11 crc kubenswrapper[4556]: I0218 09:05:11.830748 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:11 crc kubenswrapper[4556]: I0218 09:05:11.830755 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:11 crc kubenswrapper[4556]: I0218 09:05:11.830767 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:11 crc kubenswrapper[4556]: I0218 09:05:11.830778 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:11Z","lastTransitionTime":"2026-02-18T09:05:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:11 crc kubenswrapper[4556]: I0218 09:05:11.932542 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:11 crc kubenswrapper[4556]: I0218 09:05:11.932574 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:11 crc kubenswrapper[4556]: I0218 09:05:11.932602 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:11 crc kubenswrapper[4556]: I0218 09:05:11.932613 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:11 crc kubenswrapper[4556]: I0218 09:05:11.932621 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:11Z","lastTransitionTime":"2026-02-18T09:05:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:12 crc kubenswrapper[4556]: I0218 09:05:12.034385 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:12 crc kubenswrapper[4556]: I0218 09:05:12.034412 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:12 crc kubenswrapper[4556]: I0218 09:05:12.034421 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:12 crc kubenswrapper[4556]: I0218 09:05:12.034432 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:12 crc kubenswrapper[4556]: I0218 09:05:12.034442 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:12Z","lastTransitionTime":"2026-02-18T09:05:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:12 crc kubenswrapper[4556]: I0218 09:05:12.136512 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:12 crc kubenswrapper[4556]: I0218 09:05:12.136539 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:12 crc kubenswrapper[4556]: I0218 09:05:12.136547 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:12 crc kubenswrapper[4556]: I0218 09:05:12.136557 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:12 crc kubenswrapper[4556]: I0218 09:05:12.136564 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:12Z","lastTransitionTime":"2026-02-18T09:05:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:12 crc kubenswrapper[4556]: I0218 09:05:12.238411 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:12 crc kubenswrapper[4556]: I0218 09:05:12.238456 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:12 crc kubenswrapper[4556]: I0218 09:05:12.238467 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:12 crc kubenswrapper[4556]: I0218 09:05:12.238483 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:12 crc kubenswrapper[4556]: I0218 09:05:12.238494 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:12Z","lastTransitionTime":"2026-02-18T09:05:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:12 crc kubenswrapper[4556]: I0218 09:05:12.256127 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d2f0e36c-cb29-4b06-bb22-6afd59466cab-metrics-certs\") pod \"network-metrics-daemon-cgwv9\" (UID: \"d2f0e36c-cb29-4b06-bb22-6afd59466cab\") " pod="openshift-multus/network-metrics-daemon-cgwv9" Feb 18 09:05:12 crc kubenswrapper[4556]: E0218 09:05:12.256261 4556 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 18 09:05:12 crc kubenswrapper[4556]: E0218 09:05:12.256306 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d2f0e36c-cb29-4b06-bb22-6afd59466cab-metrics-certs podName:d2f0e36c-cb29-4b06-bb22-6afd59466cab nodeName:}" failed. No retries permitted until 2026-02-18 09:05:44.256293224 +0000 UTC m=+101.273254204 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d2f0e36c-cb29-4b06-bb22-6afd59466cab-metrics-certs") pod "network-metrics-daemon-cgwv9" (UID: "d2f0e36c-cb29-4b06-bb22-6afd59466cab") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 18 09:05:12 crc kubenswrapper[4556]: I0218 09:05:12.281640 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cgwv9" Feb 18 09:05:12 crc kubenswrapper[4556]: E0218 09:05:12.281739 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cgwv9" podUID="d2f0e36c-cb29-4b06-bb22-6afd59466cab" Feb 18 09:05:12 crc kubenswrapper[4556]: I0218 09:05:12.287036 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-02 13:33:09.677967745 +0000 UTC Feb 18 09:05:12 crc kubenswrapper[4556]: I0218 09:05:12.341044 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:12 crc kubenswrapper[4556]: I0218 09:05:12.341063 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:12 crc kubenswrapper[4556]: I0218 09:05:12.341072 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:12 crc kubenswrapper[4556]: I0218 09:05:12.341082 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:12 crc kubenswrapper[4556]: I0218 09:05:12.341091 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:12Z","lastTransitionTime":"2026-02-18T09:05:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:12 crc kubenswrapper[4556]: I0218 09:05:12.442690 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:12 crc kubenswrapper[4556]: I0218 09:05:12.442762 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:12 crc kubenswrapper[4556]: I0218 09:05:12.442778 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:12 crc kubenswrapper[4556]: I0218 09:05:12.442803 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:12 crc kubenswrapper[4556]: I0218 09:05:12.442816 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:12Z","lastTransitionTime":"2026-02-18T09:05:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:12 crc kubenswrapper[4556]: I0218 09:05:12.544211 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:12 crc kubenswrapper[4556]: I0218 09:05:12.544269 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:12 crc kubenswrapper[4556]: I0218 09:05:12.544281 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:12 crc kubenswrapper[4556]: I0218 09:05:12.544301 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:12 crc kubenswrapper[4556]: I0218 09:05:12.544314 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:12Z","lastTransitionTime":"2026-02-18T09:05:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:12 crc kubenswrapper[4556]: I0218 09:05:12.646839 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:12 crc kubenswrapper[4556]: I0218 09:05:12.646887 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:12 crc kubenswrapper[4556]: I0218 09:05:12.646901 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:12 crc kubenswrapper[4556]: I0218 09:05:12.646920 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:12 crc kubenswrapper[4556]: I0218 09:05:12.646933 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:12Z","lastTransitionTime":"2026-02-18T09:05:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:12 crc kubenswrapper[4556]: I0218 09:05:12.748970 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:12 crc kubenswrapper[4556]: I0218 09:05:12.749013 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:12 crc kubenswrapper[4556]: I0218 09:05:12.749024 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:12 crc kubenswrapper[4556]: I0218 09:05:12.749039 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:12 crc kubenswrapper[4556]: I0218 09:05:12.749052 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:12Z","lastTransitionTime":"2026-02-18T09:05:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:12 crc kubenswrapper[4556]: I0218 09:05:12.851604 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:12 crc kubenswrapper[4556]: I0218 09:05:12.851634 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:12 crc kubenswrapper[4556]: I0218 09:05:12.851642 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:12 crc kubenswrapper[4556]: I0218 09:05:12.851654 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:12 crc kubenswrapper[4556]: I0218 09:05:12.851663 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:12Z","lastTransitionTime":"2026-02-18T09:05:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:12 crc kubenswrapper[4556]: I0218 09:05:12.953665 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:12 crc kubenswrapper[4556]: I0218 09:05:12.953692 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:12 crc kubenswrapper[4556]: I0218 09:05:12.953701 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:12 crc kubenswrapper[4556]: I0218 09:05:12.953712 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:12 crc kubenswrapper[4556]: I0218 09:05:12.953722 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:12Z","lastTransitionTime":"2026-02-18T09:05:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.055676 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.055732 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.055744 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.055764 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.055774 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:13Z","lastTransitionTime":"2026-02-18T09:05:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.157857 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.157896 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.157905 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.157921 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.157931 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:13Z","lastTransitionTime":"2026-02-18T09:05:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.260977 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.261018 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.261026 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.261042 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.261068 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:13Z","lastTransitionTime":"2026-02-18T09:05:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.281513 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.281621 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.281666 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:05:13 crc kubenswrapper[4556]: E0218 09:05:13.281645 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 09:05:13 crc kubenswrapper[4556]: E0218 09:05:13.281732 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 09:05:13 crc kubenswrapper[4556]: E0218 09:05:13.281860 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.287100 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-17 16:19:39.504541558 +0000 UTC Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.290137 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4e7c95a-56e9-4e22-83dd-54962e1233b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52ea6a334e8cad154e46eb2f917cdac914557556bae88f922433a9545cb29eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28a600e8daaa2796b2f0cc6ce4112f8d41c611b5b6ca5a7eb0f260dab615caa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1fa733477ee6330835d52333d6e9588626b6fc7ae08ae450dd2e1821447e995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfbda9799e2d2bc16964e79b90fb7087f9e6893d867711612c63aaaa3dd2b141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfbda9799e2d2bc16964e79b90fb7087f9e6893d867711612c63aaaa3dd2b141\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:13Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.298642 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47341f31f94eaf5dd86efbe78ff62873e733b1f86ea895f281b6797770519459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:13Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.305480 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5vcgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d50c39aa-f956-441a-8abe-1d7247a7fd86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ddc5d42163d1fa86f1a59e224adac434c990224e839ef3cb12c6d7689e2b439\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4llw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5vcgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:13Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.320703 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09971522-31bc-4e10-8831-4fb62675e220\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9170bdcbaf1065995d03e6448e40bdced0741c716c9e81f37a72b015132f1161\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c526c0613f19e1a1a584b7129bd56001faf08d0b97fd77feeb69f764ed01f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cd6e758ee396c9000b1bdb6ded2176ec803fa9da6201cdd9b8146940a2c00cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b0412f3330999c3348ac726ac5a0c2d541571f4b13786ca24cb3109a44aa653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://318cb69e50df4c14df9bf5ef57d5c381421a9bdb72135c9b9620f55c0547b410\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0f15d331af9bc2ca3082c1573dd167c66773eb390593844e681f8d0b455d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03acfee4df70e62846f679407e6f9960e44946d5fd4b71d7ea1011729eb8c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b03acfee4df70e62846f679407e6f9960e44946d5fd4b71d7ea1011729eb8c05\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T09:04:48Z\\\",\\\"message\\\":\\\"ddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-controller-manager/kube-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager/kube-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.36\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0218 09:04:48.895203 6239 lb_config.go:1031] Cluster endpoints for openshift-machine-api/machine-api-controllers for network=default are: map[]\\\\nI0218 09:04:48.895220 6239 services_controller.go:443] Built service openshift-machine-api/machine-api-controllers LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.167\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8441, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]s\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-b7fsj_openshift-ovn-kubernetes(09971522-31bc-4e10-8831-4fb62675e220)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aec6030df0a59ad8902d5c694d5cd30d249ca1e895155c57a458e359fa531b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7fsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:13Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.328052 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jflp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689d3d82-4071-4fe5-b3e3-b5b4f511580a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d56d65964aaf12feaaa3e5fa3ca779bffdb13fcebfc9f027b563c7dd01f4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bm85s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jflp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:13Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.346249 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ae60f3-adab-4740-821a-1570059c7101\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f629262e27020512301a7b53ee836ee98e1e7cdee7a52b0ed506bbaaebd823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f59306c0d55e9d00714ce2ee2c1ed6f2ed33d8b21eecde490e86a93335a14f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c87bc671742a47de6c8e68aaf0c8645f9a8cae7b3d52264f36ae3827b3a8d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2678001ff826712da6716541210f018c74fd3d0218b388d2e7a6f16a1f4027\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6f37933281c927328ecfaf842d97ee8361726768f393726cb39517d2bdc9425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:13Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.355427 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8f6014-2544-4173-a94d-441cf42ba891\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77b0269625b7c6baab44a40373618f4ccfe9e1a028a1bc746b7d60c20ac4efaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac09c421b3101887bfb83107dbe26f2b5856d9b32a9cea92f6233fb6ae218cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce79cfe0242ecf403f989dc6b1c1b8be5ab54eee75ac413a8f9d03cef8bca0ed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://095ecf7e8763fa098b06eb9a8aac9f39451f7762bf20196e0803bbc841610e02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:13Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.362725 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.362757 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.362767 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.362781 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.362790 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:13Z","lastTransitionTime":"2026-02-18T09:05:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.365393 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:13Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.377110 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d653c21d3d7c875bed37ed6da700dc8c801b95c80d7e361121dfbbd3d05a922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5211d0cd5172d45b92b003e2bf97d7969a65ea448b0491f8f6ea67364925ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:13Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.387683 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6387f58543cf2601a53b0bbe9be3ec7816ba90d5ffdc4465594b37ccac0f6dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:13Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.397659 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:13Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.406449 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bp8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"804f8152-54ae-4486-92e6-f9595c035820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb617e46b8522d73dabef4bbc660e41b737f6e6e62f7ce42dad722490a1041dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phr7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://104b8ff3c41d4304d414e58303c23792a7f0ba580f73a2a8678044741af71380\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phr7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bp8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:13Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.418610 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"302eb215-a4c2-4324-836d-f37debfd620f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a99dc228d1c4cb8aeab2140ec5f37358b65684739f833fbde45822782430a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14f794063db9d33b82b79acc7134b5a2a3435e09d7cffe49d49251d1f7d20402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5eb3d3728da247903bece56bc2481547317e386e9e97b6c69926a3b55dc89b8e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b42d13c1767c29d22e52537ff45561a0fbf8f2069b2c3612e903e8c127e12485\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e865e6a50c03519ce22958ac4fcbb911fc140c330b6d7ffe5c9d7dd381e96\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T09:04:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0218 09:04:15.143085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 09:04:15.144650 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2160144045/tls.crt::/tmp/serving-cert-2160144045/tls.key\\\\\\\"\\\\nI0218 09:04:20.680799 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 09:04:20.682555 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 09:04:20.682573 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 09:04:20.682590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 09:04:20.682594 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 09:04:20.686822 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 09:04:20.686847 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686852 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686856 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 09:04:20.686859 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 09:04:20.686861 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 09:04:20.686864 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 09:04:20.688257 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 09:04:20.690726 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://974e17ae7082631df2d1bcee60c8a5f142bae2659e9d200d537a4d650f401b4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:13Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.429053 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:13Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.441720 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qfksk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e226b189-cfa9-47fc-b94d-19b5cbe0859f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc2cb92a7735d16b8e5918797e52e1e488d9261a0418ed5ebbb94c033ca9328a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2fx4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qfksk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:13Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.451351 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dac7f27-d3d1-4778-9e54-f273035a1d37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18da54fe071964d8aae8a977f5ee209c31f1c75e1041c44f8d755f58f3e39d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d41b85076e695bbc51a41fce200e0e47c099d88b4d825b4d133d33ed96237e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f76hs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:13Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.461587 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sgcq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9945b792-abbd-4103-a7b4-9496e1cc1b56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5dd402827182bc79283f1ea1c355b60c4ced53ffa9ce6781b0a0db3109359a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac0d9d7b9bc867b32cf6cbeb804c08e67df757e4b22914e2da69a376511cf75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac0d9d7b9bc867b32cf6cbeb804c08e67df757e4b22914e2da69a376511cf75c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b807d53d620e44d3196f68492b9f9665593adac804bc7b91569796e5594c4ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b807d53d620e44d3196f68492b9f9665593adac804bc7b91569796e5594c4ebb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebc044bbd5b0c1e725e6ed46ec2e0015b254070b7518d20815539dcf8c6e4e45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ebc044bbd5b0c1e725e6ed46ec2e0015b254070b7518d20815539dcf8c6e4e45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sgcq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:13Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.465010 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.465218 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.465251 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.465285 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.465296 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:13Z","lastTransitionTime":"2026-02-18T09:05:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.469687 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cgwv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2f0e36c-cb29-4b06-bb22-6afd59466cab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpldv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpldv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:40Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cgwv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:13Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.565146 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-qfksk_e226b189-cfa9-47fc-b94d-19b5cbe0859f/kube-multus/0.log" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.565210 4556 generic.go:334] "Generic (PLEG): container finished" podID="e226b189-cfa9-47fc-b94d-19b5cbe0859f" containerID="fc2cb92a7735d16b8e5918797e52e1e488d9261a0418ed5ebbb94c033ca9328a" exitCode=1 Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.565241 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-qfksk" event={"ID":"e226b189-cfa9-47fc-b94d-19b5cbe0859f","Type":"ContainerDied","Data":"fc2cb92a7735d16b8e5918797e52e1e488d9261a0418ed5ebbb94c033ca9328a"} Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.565602 4556 scope.go:117] "RemoveContainer" containerID="fc2cb92a7735d16b8e5918797e52e1e488d9261a0418ed5ebbb94c033ca9328a" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.567215 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.567259 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.567271 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.567285 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.567296 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:13Z","lastTransitionTime":"2026-02-18T09:05:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.576939 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dac7f27-d3d1-4778-9e54-f273035a1d37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18da54fe071964d8aae8a977f5ee209c31f1c75e1041c44f8d755f58f3e39d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d41b85076e695bbc51a41fce200e0e47c099d88b4d825b4d133d33ed96237e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f76hs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:13Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.590311 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sgcq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9945b792-abbd-4103-a7b4-9496e1cc1b56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5dd402827182bc79283f1ea1c355b60c4ced53ffa9ce6781b0a0db3109359a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac0d9d7b9bc867b32cf6cbeb804c08e67df757e4b22914e2da69a376511cf75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac0d9d7b9bc867b32cf6cbeb804c08e67df757e4b22914e2da69a376511cf75c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b807d53d620e44d3196f68492b9f9665593adac804bc7b91569796e5594c4ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b807d53d620e44d3196f68492b9f9665593adac804bc7b91569796e5594c4ebb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebc044bbd5b0c1e725e6ed46ec2e0015b254070b7518d20815539dcf8c6e4e45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ebc044bbd5b0c1e725e6ed46ec2e0015b254070b7518d20815539dcf8c6e4e45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sgcq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:13Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.598236 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cgwv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2f0e36c-cb29-4b06-bb22-6afd59466cab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpldv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpldv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:40Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cgwv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:13Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.608517 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"302eb215-a4c2-4324-836d-f37debfd620f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a99dc228d1c4cb8aeab2140ec5f37358b65684739f833fbde45822782430a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14f794063db9d33b82b79acc7134b5a2a3435e09d7cffe49d49251d1f7d20402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5eb3d3728da247903bece56bc2481547317e386e9e97b6c69926a3b55dc89b8e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b42d13c1767c29d22e52537ff45561a0fbf8f2069b2c3612e903e8c127e12485\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e865e6a50c03519ce22958ac4fcbb911fc140c330b6d7ffe5c9d7dd381e96\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T09:04:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0218 09:04:15.143085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 09:04:15.144650 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2160144045/tls.crt::/tmp/serving-cert-2160144045/tls.key\\\\\\\"\\\\nI0218 09:04:20.680799 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 09:04:20.682555 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 09:04:20.682573 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 09:04:20.682590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 09:04:20.682594 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 09:04:20.686822 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 09:04:20.686847 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686852 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686856 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 09:04:20.686859 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 09:04:20.686861 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 09:04:20.686864 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 09:04:20.688257 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 09:04:20.690726 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://974e17ae7082631df2d1bcee60c8a5f142bae2659e9d200d537a4d650f401b4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:13Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.617611 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:13Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.631213 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qfksk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e226b189-cfa9-47fc-b94d-19b5cbe0859f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc2cb92a7735d16b8e5918797e52e1e488d9261a0418ed5ebbb94c033ca9328a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc2cb92a7735d16b8e5918797e52e1e488d9261a0418ed5ebbb94c033ca9328a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T09:05:12Z\\\",\\\"message\\\":\\\"2026-02-18T09:04:27+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_3fa1df1f-8463-485c-a4c9-cd0c6dd4d227\\\\n2026-02-18T09:04:27+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_3fa1df1f-8463-485c-a4c9-cd0c6dd4d227 to /host/opt/cni/bin/\\\\n2026-02-18T09:04:27Z [verbose] multus-daemon started\\\\n2026-02-18T09:04:27Z [verbose] Readiness Indicator file check\\\\n2026-02-18T09:05:12Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2fx4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qfksk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:13Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.641353 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5vcgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d50c39aa-f956-441a-8abe-1d7247a7fd86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ddc5d42163d1fa86f1a59e224adac434c990224e839ef3cb12c6d7689e2b439\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4llw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5vcgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:13Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.657711 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09971522-31bc-4e10-8831-4fb62675e220\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9170bdcbaf1065995d03e6448e40bdced0741c716c9e81f37a72b015132f1161\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c526c0613f19e1a1a584b7129bd56001faf08d0b97fd77feeb69f764ed01f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cd6e758ee396c9000b1bdb6ded2176ec803fa9da6201cdd9b8146940a2c00cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b0412f3330999c3348ac726ac5a0c2d541571f4b13786ca24cb3109a44aa653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://318cb69e50df4c14df9bf5ef57d5c381421a9bdb72135c9b9620f55c0547b410\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0f15d331af9bc2ca3082c1573dd167c66773eb390593844e681f8d0b455d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03acfee4df70e62846f679407e6f9960e44946d5fd4b71d7ea1011729eb8c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b03acfee4df70e62846f679407e6f9960e44946d5fd4b71d7ea1011729eb8c05\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T09:04:48Z\\\",\\\"message\\\":\\\"ddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-controller-manager/kube-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager/kube-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.36\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0218 09:04:48.895203 6239 lb_config.go:1031] Cluster endpoints for openshift-machine-api/machine-api-controllers for network=default are: map[]\\\\nI0218 09:04:48.895220 6239 services_controller.go:443] Built service openshift-machine-api/machine-api-controllers LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.167\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8441, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]s\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-b7fsj_openshift-ovn-kubernetes(09971522-31bc-4e10-8831-4fb62675e220)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aec6030df0a59ad8902d5c694d5cd30d249ca1e895155c57a458e359fa531b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7fsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:13Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.668119 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4e7c95a-56e9-4e22-83dd-54962e1233b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52ea6a334e8cad154e46eb2f917cdac914557556bae88f922433a9545cb29eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28a600e8daaa2796b2f0cc6ce4112f8d41c611b5b6ca5a7eb0f260dab615caa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1fa733477ee6330835d52333d6e9588626b6fc7ae08ae450dd2e1821447e995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfbda9799e2d2bc16964e79b90fb7087f9e6893d867711612c63aaaa3dd2b141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfbda9799e2d2bc16964e79b90fb7087f9e6893d867711612c63aaaa3dd2b141\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:13Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.669907 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.669949 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.669964 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.669985 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.669999 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:13Z","lastTransitionTime":"2026-02-18T09:05:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.680131 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47341f31f94eaf5dd86efbe78ff62873e733b1f86ea895f281b6797770519459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:13Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.689222 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jflp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689d3d82-4071-4fe5-b3e3-b5b4f511580a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d56d65964aaf12feaaa3e5fa3ca779bffdb13fcebfc9f027b563c7dd01f4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bm85s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jflp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:13Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.698636 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d653c21d3d7c875bed37ed6da700dc8c801b95c80d7e361121dfbbd3d05a922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5211d0cd5172d45b92b003e2bf97d7969a65ea448b0491f8f6ea67364925ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:13Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.708054 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6387f58543cf2601a53b0bbe9be3ec7816ba90d5ffdc4465594b37ccac0f6dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:13Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.717699 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:13Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.726012 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bp8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"804f8152-54ae-4486-92e6-f9595c035820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb617e46b8522d73dabef4bbc660e41b737f6e6e62f7ce42dad722490a1041dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phr7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://104b8ff3c41d4304d414e58303c23792a7f0ba580f73a2a8678044741af71380\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phr7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bp8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:13Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.748684 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ae60f3-adab-4740-821a-1570059c7101\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f629262e27020512301a7b53ee836ee98e1e7cdee7a52b0ed506bbaaebd823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f59306c0d55e9d00714ce2ee2c1ed6f2ed33d8b21eecde490e86a93335a14f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c87bc671742a47de6c8e68aaf0c8645f9a8cae7b3d52264f36ae3827b3a8d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2678001ff826712da6716541210f018c74fd3d0218b388d2e7a6f16a1f4027\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6f37933281c927328ecfaf842d97ee8361726768f393726cb39517d2bdc9425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:13Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.758051 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8f6014-2544-4173-a94d-441cf42ba891\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77b0269625b7c6baab44a40373618f4ccfe9e1a028a1bc746b7d60c20ac4efaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac09c421b3101887bfb83107dbe26f2b5856d9b32a9cea92f6233fb6ae218cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce79cfe0242ecf403f989dc6b1c1b8be5ab54eee75ac413a8f9d03cef8bca0ed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://095ecf7e8763fa098b06eb9a8aac9f39451f7762bf20196e0803bbc841610e02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:13Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.768956 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:13Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.773479 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.773530 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.773542 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.773563 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.773576 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:13Z","lastTransitionTime":"2026-02-18T09:05:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.875395 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.875423 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.875443 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.875458 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.875469 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:13Z","lastTransitionTime":"2026-02-18T09:05:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.978358 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.978403 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.978414 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.978445 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:13 crc kubenswrapper[4556]: I0218 09:05:13.978473 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:13Z","lastTransitionTime":"2026-02-18T09:05:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.081030 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.081064 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.081073 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.081090 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.081102 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:14Z","lastTransitionTime":"2026-02-18T09:05:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.183231 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.183280 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.183291 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.183317 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.183334 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:14Z","lastTransitionTime":"2026-02-18T09:05:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.282110 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cgwv9" Feb 18 09:05:14 crc kubenswrapper[4556]: E0218 09:05:14.282549 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cgwv9" podUID="d2f0e36c-cb29-4b06-bb22-6afd59466cab" Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.287272 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.287372 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.287951 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.288032 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.287204 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 18:08:13.059936088 +0000 UTC Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.288100 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:14Z","lastTransitionTime":"2026-02-18T09:05:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.391390 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.391458 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.391474 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.391500 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.391514 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:14Z","lastTransitionTime":"2026-02-18T09:05:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.493719 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.493775 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.493803 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.493819 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.493829 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:14Z","lastTransitionTime":"2026-02-18T09:05:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.569492 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-qfksk_e226b189-cfa9-47fc-b94d-19b5cbe0859f/kube-multus/0.log" Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.569549 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-qfksk" event={"ID":"e226b189-cfa9-47fc-b94d-19b5cbe0859f","Type":"ContainerStarted","Data":"23f488cd416c6d49acf193e27beba9e9f1dcfb36430f9ae540742c2e2bd0f144"} Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.581487 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cgwv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2f0e36c-cb29-4b06-bb22-6afd59466cab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpldv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpldv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:40Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cgwv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:14Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.592559 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"302eb215-a4c2-4324-836d-f37debfd620f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a99dc228d1c4cb8aeab2140ec5f37358b65684739f833fbde45822782430a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14f794063db9d33b82b79acc7134b5a2a3435e09d7cffe49d49251d1f7d20402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5eb3d3728da247903bece56bc2481547317e386e9e97b6c69926a3b55dc89b8e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b42d13c1767c29d22e52537ff45561a0fbf8f2069b2c3612e903e8c127e12485\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e865e6a50c03519ce22958ac4fcbb911fc140c330b6d7ffe5c9d7dd381e96\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T09:04:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0218 09:04:15.143085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 09:04:15.144650 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2160144045/tls.crt::/tmp/serving-cert-2160144045/tls.key\\\\\\\"\\\\nI0218 09:04:20.680799 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 09:04:20.682555 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 09:04:20.682573 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 09:04:20.682590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 09:04:20.682594 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 09:04:20.686822 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 09:04:20.686847 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686852 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686856 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 09:04:20.686859 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 09:04:20.686861 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 09:04:20.686864 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 09:04:20.688257 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 09:04:20.690726 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://974e17ae7082631df2d1bcee60c8a5f142bae2659e9d200d537a4d650f401b4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:14Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.595788 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.595814 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.595825 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.595838 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.595848 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:14Z","lastTransitionTime":"2026-02-18T09:05:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.603375 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:14Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.613213 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qfksk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e226b189-cfa9-47fc-b94d-19b5cbe0859f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23f488cd416c6d49acf193e27beba9e9f1dcfb36430f9ae540742c2e2bd0f144\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc2cb92a7735d16b8e5918797e52e1e488d9261a0418ed5ebbb94c033ca9328a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T09:05:12Z\\\",\\\"message\\\":\\\"2026-02-18T09:04:27+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_3fa1df1f-8463-485c-a4c9-cd0c6dd4d227\\\\n2026-02-18T09:04:27+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_3fa1df1f-8463-485c-a4c9-cd0c6dd4d227 to /host/opt/cni/bin/\\\\n2026-02-18T09:04:27Z [verbose] multus-daemon started\\\\n2026-02-18T09:04:27Z [verbose] Readiness Indicator file check\\\\n2026-02-18T09:05:12Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:05:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2fx4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qfksk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:14Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.622704 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dac7f27-d3d1-4778-9e54-f273035a1d37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18da54fe071964d8aae8a977f5ee209c31f1c75e1041c44f8d755f58f3e39d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d41b85076e695bbc51a41fce200e0e47c099d88b4d825b4d133d33ed96237e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f76hs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:14Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.637115 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sgcq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9945b792-abbd-4103-a7b4-9496e1cc1b56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5dd402827182bc79283f1ea1c355b60c4ced53ffa9ce6781b0a0db3109359a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac0d9d7b9bc867b32cf6cbeb804c08e67df757e4b22914e2da69a376511cf75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac0d9d7b9bc867b32cf6cbeb804c08e67df757e4b22914e2da69a376511cf75c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b807d53d620e44d3196f68492b9f9665593adac804bc7b91569796e5594c4ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b807d53d620e44d3196f68492b9f9665593adac804bc7b91569796e5594c4ebb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebc044bbd5b0c1e725e6ed46ec2e0015b254070b7518d20815539dcf8c6e4e45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ebc044bbd5b0c1e725e6ed46ec2e0015b254070b7518d20815539dcf8c6e4e45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sgcq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:14Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.647405 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4e7c95a-56e9-4e22-83dd-54962e1233b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52ea6a334e8cad154e46eb2f917cdac914557556bae88f922433a9545cb29eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28a600e8daaa2796b2f0cc6ce4112f8d41c611b5b6ca5a7eb0f260dab615caa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1fa733477ee6330835d52333d6e9588626b6fc7ae08ae450dd2e1821447e995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfbda9799e2d2bc16964e79b90fb7087f9e6893d867711612c63aaaa3dd2b141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfbda9799e2d2bc16964e79b90fb7087f9e6893d867711612c63aaaa3dd2b141\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:14Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.656191 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47341f31f94eaf5dd86efbe78ff62873e733b1f86ea895f281b6797770519459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:14Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.663758 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5vcgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d50c39aa-f956-441a-8abe-1d7247a7fd86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ddc5d42163d1fa86f1a59e224adac434c990224e839ef3cb12c6d7689e2b439\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4llw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5vcgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:14Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.678788 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09971522-31bc-4e10-8831-4fb62675e220\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9170bdcbaf1065995d03e6448e40bdced0741c716c9e81f37a72b015132f1161\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c526c0613f19e1a1a584b7129bd56001faf08d0b97fd77feeb69f764ed01f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cd6e758ee396c9000b1bdb6ded2176ec803fa9da6201cdd9b8146940a2c00cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b0412f3330999c3348ac726ac5a0c2d541571f4b13786ca24cb3109a44aa653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://318cb69e50df4c14df9bf5ef57d5c381421a9bdb72135c9b9620f55c0547b410\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0f15d331af9bc2ca3082c1573dd167c66773eb390593844e681f8d0b455d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03acfee4df70e62846f679407e6f9960e44946d5fd4b71d7ea1011729eb8c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b03acfee4df70e62846f679407e6f9960e44946d5fd4b71d7ea1011729eb8c05\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T09:04:48Z\\\",\\\"message\\\":\\\"ddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-controller-manager/kube-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager/kube-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.36\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0218 09:04:48.895203 6239 lb_config.go:1031] Cluster endpoints for openshift-machine-api/machine-api-controllers for network=default are: map[]\\\\nI0218 09:04:48.895220 6239 services_controller.go:443] Built service openshift-machine-api/machine-api-controllers LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.167\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8441, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]s\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-b7fsj_openshift-ovn-kubernetes(09971522-31bc-4e10-8831-4fb62675e220)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aec6030df0a59ad8902d5c694d5cd30d249ca1e895155c57a458e359fa531b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7fsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:14Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.686931 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jflp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689d3d82-4071-4fe5-b3e3-b5b4f511580a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d56d65964aaf12feaaa3e5fa3ca779bffdb13fcebfc9f027b563c7dd01f4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bm85s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jflp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:14Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.697422 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:14Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.698050 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.698142 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.698228 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.698297 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.698363 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:14Z","lastTransitionTime":"2026-02-18T09:05:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.707086 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bp8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"804f8152-54ae-4486-92e6-f9595c035820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb617e46b8522d73dabef4bbc660e41b737f6e6e62f7ce42dad722490a1041dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phr7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://104b8ff3c41d4304d414e58303c23792a7f0ba580f73a2a8678044741af71380\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phr7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bp8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:14Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.723087 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ae60f3-adab-4740-821a-1570059c7101\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f629262e27020512301a7b53ee836ee98e1e7cdee7a52b0ed506bbaaebd823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f59306c0d55e9d00714ce2ee2c1ed6f2ed33d8b21eecde490e86a93335a14f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c87bc671742a47de6c8e68aaf0c8645f9a8cae7b3d52264f36ae3827b3a8d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2678001ff826712da6716541210f018c74fd3d0218b388d2e7a6f16a1f4027\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6f37933281c927328ecfaf842d97ee8361726768f393726cb39517d2bdc9425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:14Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.733106 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8f6014-2544-4173-a94d-441cf42ba891\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77b0269625b7c6baab44a40373618f4ccfe9e1a028a1bc746b7d60c20ac4efaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac09c421b3101887bfb83107dbe26f2b5856d9b32a9cea92f6233fb6ae218cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce79cfe0242ecf403f989dc6b1c1b8be5ab54eee75ac413a8f9d03cef8bca0ed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://095ecf7e8763fa098b06eb9a8aac9f39451f7762bf20196e0803bbc841610e02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:14Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.742928 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:14Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.752247 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d653c21d3d7c875bed37ed6da700dc8c801b95c80d7e361121dfbbd3d05a922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5211d0cd5172d45b92b003e2bf97d7969a65ea448b0491f8f6ea67364925ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:14Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.761495 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6387f58543cf2601a53b0bbe9be3ec7816ba90d5ffdc4465594b37ccac0f6dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:14Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.800712 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.800751 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.800763 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.800778 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.800789 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:14Z","lastTransitionTime":"2026-02-18T09:05:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.902897 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.902934 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.902945 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.902970 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:14 crc kubenswrapper[4556]: I0218 09:05:14.902983 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:14Z","lastTransitionTime":"2026-02-18T09:05:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.005110 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.005143 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.005171 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.005187 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.005199 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:15Z","lastTransitionTime":"2026-02-18T09:05:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.107071 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.107101 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.107115 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.107130 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.107140 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:15Z","lastTransitionTime":"2026-02-18T09:05:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.127649 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.127749 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.127761 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.127775 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.127786 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:15Z","lastTransitionTime":"2026-02-18T09:05:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:15 crc kubenswrapper[4556]: E0218 09:05:15.138552 4556 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:05:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:05:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:05:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:05:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8758c02e-ef59-46a7-908c-d5c97feb8ceb\\\",\\\"systemUUID\\\":\\\"0f904183-79bf-4790-8551-43a6eb3adbe4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:15Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.141874 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.142023 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.142109 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.142209 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.142284 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:15Z","lastTransitionTime":"2026-02-18T09:05:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:15 crc kubenswrapper[4556]: E0218 09:05:15.152769 4556 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:05:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:05:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:05:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:05:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8758c02e-ef59-46a7-908c-d5c97feb8ceb\\\",\\\"systemUUID\\\":\\\"0f904183-79bf-4790-8551-43a6eb3adbe4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:15Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.158625 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.158659 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.158668 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.158685 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.158697 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:15Z","lastTransitionTime":"2026-02-18T09:05:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:15 crc kubenswrapper[4556]: E0218 09:05:15.167624 4556 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:05:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:05:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:05:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:05:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8758c02e-ef59-46a7-908c-d5c97feb8ceb\\\",\\\"systemUUID\\\":\\\"0f904183-79bf-4790-8551-43a6eb3adbe4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:15Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.171025 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.171063 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.171075 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.171120 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.171133 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:15Z","lastTransitionTime":"2026-02-18T09:05:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:15 crc kubenswrapper[4556]: E0218 09:05:15.182193 4556 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:05:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:05:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:05:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:05:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8758c02e-ef59-46a7-908c-d5c97feb8ceb\\\",\\\"systemUUID\\\":\\\"0f904183-79bf-4790-8551-43a6eb3adbe4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:15Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.185415 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.185465 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.185480 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.185496 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.185506 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:15Z","lastTransitionTime":"2026-02-18T09:05:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:15 crc kubenswrapper[4556]: E0218 09:05:15.193935 4556 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:05:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:05:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:05:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:05:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8758c02e-ef59-46a7-908c-d5c97feb8ceb\\\",\\\"systemUUID\\\":\\\"0f904183-79bf-4790-8551-43a6eb3adbe4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:15Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:15 crc kubenswrapper[4556]: E0218 09:05:15.194051 4556 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.209184 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.209208 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.209218 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.209230 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.209242 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:15Z","lastTransitionTime":"2026-02-18T09:05:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.282024 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.282052 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.282080 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:05:15 crc kubenswrapper[4556]: E0218 09:05:15.282148 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 09:05:15 crc kubenswrapper[4556]: E0218 09:05:15.282233 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 09:05:15 crc kubenswrapper[4556]: E0218 09:05:15.282297 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.288317 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 00:49:40.576671966 +0000 UTC Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.310713 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.310756 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.310769 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.310785 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.310796 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:15Z","lastTransitionTime":"2026-02-18T09:05:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.412598 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.412632 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.412643 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.412658 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.412670 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:15Z","lastTransitionTime":"2026-02-18T09:05:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.514142 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.514189 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.514200 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.514216 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.514226 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:15Z","lastTransitionTime":"2026-02-18T09:05:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.615872 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.615903 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.615913 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.615925 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.615936 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:15Z","lastTransitionTime":"2026-02-18T09:05:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.717802 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.717836 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.717845 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.717855 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.717864 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:15Z","lastTransitionTime":"2026-02-18T09:05:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.819670 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.819692 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.819700 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.819712 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.819719 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:15Z","lastTransitionTime":"2026-02-18T09:05:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.921608 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.921635 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.921644 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.921653 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:15 crc kubenswrapper[4556]: I0218 09:05:15.921662 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:15Z","lastTransitionTime":"2026-02-18T09:05:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.022905 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.022934 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.022943 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.022953 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.022977 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:16Z","lastTransitionTime":"2026-02-18T09:05:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.124815 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.124848 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.124857 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.124873 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.124883 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:16Z","lastTransitionTime":"2026-02-18T09:05:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.226221 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.226250 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.226259 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.226269 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.226276 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:16Z","lastTransitionTime":"2026-02-18T09:05:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.282173 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cgwv9" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.282605 4556 scope.go:117] "RemoveContainer" containerID="b03acfee4df70e62846f679407e6f9960e44946d5fd4b71d7ea1011729eb8c05" Feb 18 09:05:16 crc kubenswrapper[4556]: E0218 09:05:16.282623 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cgwv9" podUID="d2f0e36c-cb29-4b06-bb22-6afd59466cab" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.289197 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-31 14:20:53.800399192 +0000 UTC Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.292641 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.328264 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.328293 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.328303 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.328316 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.328326 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:16Z","lastTransitionTime":"2026-02-18T09:05:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.430690 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.430716 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.430727 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.430739 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.430749 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:16Z","lastTransitionTime":"2026-02-18T09:05:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.532716 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.532762 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.532772 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.532786 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.532794 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:16Z","lastTransitionTime":"2026-02-18T09:05:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.576709 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-b7fsj_09971522-31bc-4e10-8831-4fb62675e220/ovnkube-controller/2.log" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.580449 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" event={"ID":"09971522-31bc-4e10-8831-4fb62675e220","Type":"ContainerStarted","Data":"c2f16fe4bf2d659398ec141ba12d2dda32727810b03b96bfaf491602656027cb"} Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.580956 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.591771 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"302eb215-a4c2-4324-836d-f37debfd620f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a99dc228d1c4cb8aeab2140ec5f37358b65684739f833fbde45822782430a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14f794063db9d33b82b79acc7134b5a2a3435e09d7cffe49d49251d1f7d20402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5eb3d3728da247903bece56bc2481547317e386e9e97b6c69926a3b55dc89b8e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b42d13c1767c29d22e52537ff45561a0fbf8f2069b2c3612e903e8c127e12485\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e865e6a50c03519ce22958ac4fcbb911fc140c330b6d7ffe5c9d7dd381e96\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T09:04:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0218 09:04:15.143085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 09:04:15.144650 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2160144045/tls.crt::/tmp/serving-cert-2160144045/tls.key\\\\\\\"\\\\nI0218 09:04:20.680799 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 09:04:20.682555 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 09:04:20.682573 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 09:04:20.682590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 09:04:20.682594 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 09:04:20.686822 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 09:04:20.686847 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686852 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686856 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 09:04:20.686859 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 09:04:20.686861 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 09:04:20.686864 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 09:04:20.688257 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 09:04:20.690726 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://974e17ae7082631df2d1bcee60c8a5f142bae2659e9d200d537a4d650f401b4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:16Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.601022 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:16Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.609736 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qfksk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e226b189-cfa9-47fc-b94d-19b5cbe0859f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23f488cd416c6d49acf193e27beba9e9f1dcfb36430f9ae540742c2e2bd0f144\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc2cb92a7735d16b8e5918797e52e1e488d9261a0418ed5ebbb94c033ca9328a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T09:05:12Z\\\",\\\"message\\\":\\\"2026-02-18T09:04:27+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_3fa1df1f-8463-485c-a4c9-cd0c6dd4d227\\\\n2026-02-18T09:04:27+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_3fa1df1f-8463-485c-a4c9-cd0c6dd4d227 to /host/opt/cni/bin/\\\\n2026-02-18T09:04:27Z [verbose] multus-daemon started\\\\n2026-02-18T09:04:27Z [verbose] Readiness Indicator file check\\\\n2026-02-18T09:05:12Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:05:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2fx4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qfksk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:16Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.617088 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dac7f27-d3d1-4778-9e54-f273035a1d37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18da54fe071964d8aae8a977f5ee209c31f1c75e1041c44f8d755f58f3e39d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d41b85076e695bbc51a41fce200e0e47c099d88b4d825b4d133d33ed96237e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f76hs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:16Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.629942 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sgcq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9945b792-abbd-4103-a7b4-9496e1cc1b56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5dd402827182bc79283f1ea1c355b60c4ced53ffa9ce6781b0a0db3109359a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac0d9d7b9bc867b32cf6cbeb804c08e67df757e4b22914e2da69a376511cf75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac0d9d7b9bc867b32cf6cbeb804c08e67df757e4b22914e2da69a376511cf75c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b807d53d620e44d3196f68492b9f9665593adac804bc7b91569796e5594c4ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b807d53d620e44d3196f68492b9f9665593adac804bc7b91569796e5594c4ebb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebc044bbd5b0c1e725e6ed46ec2e0015b254070b7518d20815539dcf8c6e4e45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ebc044bbd5b0c1e725e6ed46ec2e0015b254070b7518d20815539dcf8c6e4e45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sgcq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:16Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.635350 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.635383 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.635392 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.635407 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.635425 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:16Z","lastTransitionTime":"2026-02-18T09:05:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.637849 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cgwv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2f0e36c-cb29-4b06-bb22-6afd59466cab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpldv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpldv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:40Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cgwv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:16Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.645921 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4e7c95a-56e9-4e22-83dd-54962e1233b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52ea6a334e8cad154e46eb2f917cdac914557556bae88f922433a9545cb29eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28a600e8daaa2796b2f0cc6ce4112f8d41c611b5b6ca5a7eb0f260dab615caa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1fa733477ee6330835d52333d6e9588626b6fc7ae08ae450dd2e1821447e995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfbda9799e2d2bc16964e79b90fb7087f9e6893d867711612c63aaaa3dd2b141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfbda9799e2d2bc16964e79b90fb7087f9e6893d867711612c63aaaa3dd2b141\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:16Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.652610 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64e03bf0-d2fc-40ba-a468-4d9a2b43afc7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ac8fc80676085932e05da4ff18e892eebaec46d096c7c343b667864f6c5e3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6e0422278544bddfec087d7b0a94bedc55a7ac67ff280a36030528eb87fd72b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6e0422278544bddfec087d7b0a94bedc55a7ac67ff280a36030528eb87fd72b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:16Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.660349 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47341f31f94eaf5dd86efbe78ff62873e733b1f86ea895f281b6797770519459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:16Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.667078 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5vcgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d50c39aa-f956-441a-8abe-1d7247a7fd86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ddc5d42163d1fa86f1a59e224adac434c990224e839ef3cb12c6d7689e2b439\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4llw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5vcgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:16Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.679393 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09971522-31bc-4e10-8831-4fb62675e220\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9170bdcbaf1065995d03e6448e40bdced0741c716c9e81f37a72b015132f1161\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c526c0613f19e1a1a584b7129bd56001faf08d0b97fd77feeb69f764ed01f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cd6e758ee396c9000b1bdb6ded2176ec803fa9da6201cdd9b8146940a2c00cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b0412f3330999c3348ac726ac5a0c2d541571f4b13786ca24cb3109a44aa653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://318cb69e50df4c14df9bf5ef57d5c381421a9bdb72135c9b9620f55c0547b410\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0f15d331af9bc2ca3082c1573dd167c66773eb390593844e681f8d0b455d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2f16fe4bf2d659398ec141ba12d2dda32727810b03b96bfaf491602656027cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b03acfee4df70e62846f679407e6f9960e44946d5fd4b71d7ea1011729eb8c05\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T09:04:48Z\\\",\\\"message\\\":\\\"ddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-controller-manager/kube-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager/kube-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.36\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0218 09:04:48.895203 6239 lb_config.go:1031] Cluster endpoints for openshift-machine-api/machine-api-controllers for network=default are: map[]\\\\nI0218 09:04:48.895220 6239 services_controller.go:443] Built service openshift-machine-api/machine-api-controllers LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.167\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8441, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]s\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aec6030df0a59ad8902d5c694d5cd30d249ca1e895155c57a458e359fa531b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7fsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:16Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.686430 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jflp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689d3d82-4071-4fe5-b3e3-b5b4f511580a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d56d65964aaf12feaaa3e5fa3ca779bffdb13fcebfc9f027b563c7dd01f4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bm85s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jflp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:16Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.700054 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ae60f3-adab-4740-821a-1570059c7101\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f629262e27020512301a7b53ee836ee98e1e7cdee7a52b0ed506bbaaebd823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f59306c0d55e9d00714ce2ee2c1ed6f2ed33d8b21eecde490e86a93335a14f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c87bc671742a47de6c8e68aaf0c8645f9a8cae7b3d52264f36ae3827b3a8d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2678001ff826712da6716541210f018c74fd3d0218b388d2e7a6f16a1f4027\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6f37933281c927328ecfaf842d97ee8361726768f393726cb39517d2bdc9425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:16Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.708970 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8f6014-2544-4173-a94d-441cf42ba891\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77b0269625b7c6baab44a40373618f4ccfe9e1a028a1bc746b7d60c20ac4efaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac09c421b3101887bfb83107dbe26f2b5856d9b32a9cea92f6233fb6ae218cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce79cfe0242ecf403f989dc6b1c1b8be5ab54eee75ac413a8f9d03cef8bca0ed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://095ecf7e8763fa098b06eb9a8aac9f39451f7762bf20196e0803bbc841610e02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:16Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.737909 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.737945 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.737955 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.737971 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.737980 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:16Z","lastTransitionTime":"2026-02-18T09:05:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.741208 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:16Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.757461 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d653c21d3d7c875bed37ed6da700dc8c801b95c80d7e361121dfbbd3d05a922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5211d0cd5172d45b92b003e2bf97d7969a65ea448b0491f8f6ea67364925ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:16Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.767958 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6387f58543cf2601a53b0bbe9be3ec7816ba90d5ffdc4465594b37ccac0f6dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:16Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.777238 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:16Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.790120 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bp8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"804f8152-54ae-4486-92e6-f9595c035820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb617e46b8522d73dabef4bbc660e41b737f6e6e62f7ce42dad722490a1041dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phr7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://104b8ff3c41d4304d414e58303c23792a7f0ba580f73a2a8678044741af71380\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phr7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bp8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:16Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.840112 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.840170 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.840182 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.840198 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.840208 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:16Z","lastTransitionTime":"2026-02-18T09:05:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.942058 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.942096 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.942105 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.942119 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:16 crc kubenswrapper[4556]: I0218 09:05:16.942135 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:16Z","lastTransitionTime":"2026-02-18T09:05:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.043968 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.044001 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.044024 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.044037 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.044045 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:17Z","lastTransitionTime":"2026-02-18T09:05:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.145739 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.145766 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.145774 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.145785 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.145792 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:17Z","lastTransitionTime":"2026-02-18T09:05:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.247814 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.247842 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.247857 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.247869 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.247877 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:17Z","lastTransitionTime":"2026-02-18T09:05:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.281398 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.281459 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.281483 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:05:17 crc kubenswrapper[4556]: E0218 09:05:17.281545 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 09:05:17 crc kubenswrapper[4556]: E0218 09:05:17.281618 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 09:05:17 crc kubenswrapper[4556]: E0218 09:05:17.281703 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.289633 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 05:38:35.685392884 +0000 UTC Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.349265 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.349292 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.349301 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.349311 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.349320 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:17Z","lastTransitionTime":"2026-02-18T09:05:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.451220 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.451247 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.451255 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.451267 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.451275 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:17Z","lastTransitionTime":"2026-02-18T09:05:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.552957 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.552993 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.553002 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.553014 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.553025 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:17Z","lastTransitionTime":"2026-02-18T09:05:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.584476 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-b7fsj_09971522-31bc-4e10-8831-4fb62675e220/ovnkube-controller/3.log" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.584921 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-b7fsj_09971522-31bc-4e10-8831-4fb62675e220/ovnkube-controller/2.log" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.586904 4556 generic.go:334] "Generic (PLEG): container finished" podID="09971522-31bc-4e10-8831-4fb62675e220" containerID="c2f16fe4bf2d659398ec141ba12d2dda32727810b03b96bfaf491602656027cb" exitCode=1 Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.586931 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" event={"ID":"09971522-31bc-4e10-8831-4fb62675e220","Type":"ContainerDied","Data":"c2f16fe4bf2d659398ec141ba12d2dda32727810b03b96bfaf491602656027cb"} Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.586957 4556 scope.go:117] "RemoveContainer" containerID="b03acfee4df70e62846f679407e6f9960e44946d5fd4b71d7ea1011729eb8c05" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.587350 4556 scope.go:117] "RemoveContainer" containerID="c2f16fe4bf2d659398ec141ba12d2dda32727810b03b96bfaf491602656027cb" Feb 18 09:05:17 crc kubenswrapper[4556]: E0218 09:05:17.587533 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-b7fsj_openshift-ovn-kubernetes(09971522-31bc-4e10-8831-4fb62675e220)\"" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" podUID="09971522-31bc-4e10-8831-4fb62675e220" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.596622 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:17Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.605784 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qfksk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e226b189-cfa9-47fc-b94d-19b5cbe0859f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23f488cd416c6d49acf193e27beba9e9f1dcfb36430f9ae540742c2e2bd0f144\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc2cb92a7735d16b8e5918797e52e1e488d9261a0418ed5ebbb94c033ca9328a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T09:05:12Z\\\",\\\"message\\\":\\\"2026-02-18T09:04:27+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_3fa1df1f-8463-485c-a4c9-cd0c6dd4d227\\\\n2026-02-18T09:04:27+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_3fa1df1f-8463-485c-a4c9-cd0c6dd4d227 to /host/opt/cni/bin/\\\\n2026-02-18T09:04:27Z [verbose] multus-daemon started\\\\n2026-02-18T09:04:27Z [verbose] Readiness Indicator file check\\\\n2026-02-18T09:05:12Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:05:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2fx4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qfksk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:17Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.613464 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dac7f27-d3d1-4778-9e54-f273035a1d37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18da54fe071964d8aae8a977f5ee209c31f1c75e1041c44f8d755f58f3e39d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d41b85076e695bbc51a41fce200e0e47c099d88b4d825b4d133d33ed96237e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f76hs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:17Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.622695 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sgcq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9945b792-abbd-4103-a7b4-9496e1cc1b56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5dd402827182bc79283f1ea1c355b60c4ced53ffa9ce6781b0a0db3109359a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac0d9d7b9bc867b32cf6cbeb804c08e67df757e4b22914e2da69a376511cf75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac0d9d7b9bc867b32cf6cbeb804c08e67df757e4b22914e2da69a376511cf75c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b807d53d620e44d3196f68492b9f9665593adac804bc7b91569796e5594c4ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b807d53d620e44d3196f68492b9f9665593adac804bc7b91569796e5594c4ebb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebc044bbd5b0c1e725e6ed46ec2e0015b254070b7518d20815539dcf8c6e4e45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ebc044bbd5b0c1e725e6ed46ec2e0015b254070b7518d20815539dcf8c6e4e45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sgcq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:17Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.629579 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cgwv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2f0e36c-cb29-4b06-bb22-6afd59466cab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpldv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpldv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:40Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cgwv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:17Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.638251 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"302eb215-a4c2-4324-836d-f37debfd620f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a99dc228d1c4cb8aeab2140ec5f37358b65684739f833fbde45822782430a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14f794063db9d33b82b79acc7134b5a2a3435e09d7cffe49d49251d1f7d20402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5eb3d3728da247903bece56bc2481547317e386e9e97b6c69926a3b55dc89b8e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b42d13c1767c29d22e52537ff45561a0fbf8f2069b2c3612e903e8c127e12485\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e865e6a50c03519ce22958ac4fcbb911fc140c330b6d7ffe5c9d7dd381e96\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T09:04:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0218 09:04:15.143085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 09:04:15.144650 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2160144045/tls.crt::/tmp/serving-cert-2160144045/tls.key\\\\\\\"\\\\nI0218 09:04:20.680799 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 09:04:20.682555 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 09:04:20.682573 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 09:04:20.682590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 09:04:20.682594 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 09:04:20.686822 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 09:04:20.686847 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686852 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686856 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 09:04:20.686859 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 09:04:20.686861 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 09:04:20.686864 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 09:04:20.688257 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 09:04:20.690726 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://974e17ae7082631df2d1bcee60c8a5f142bae2659e9d200d537a4d650f401b4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:17Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.644877 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64e03bf0-d2fc-40ba-a468-4d9a2b43afc7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ac8fc80676085932e05da4ff18e892eebaec46d096c7c343b667864f6c5e3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6e0422278544bddfec087d7b0a94bedc55a7ac67ff280a36030528eb87fd72b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6e0422278544bddfec087d7b0a94bedc55a7ac67ff280a36030528eb87fd72b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:17Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.652763 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47341f31f94eaf5dd86efbe78ff62873e733b1f86ea895f281b6797770519459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:17Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.655149 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.655190 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.655198 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.655211 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.655220 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:17Z","lastTransitionTime":"2026-02-18T09:05:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.659442 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5vcgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d50c39aa-f956-441a-8abe-1d7247a7fd86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ddc5d42163d1fa86f1a59e224adac434c990224e839ef3cb12c6d7689e2b439\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4llw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5vcgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:17Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.671199 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09971522-31bc-4e10-8831-4fb62675e220\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9170bdcbaf1065995d03e6448e40bdced0741c716c9e81f37a72b015132f1161\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c526c0613f19e1a1a584b7129bd56001faf08d0b97fd77feeb69f764ed01f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cd6e758ee396c9000b1bdb6ded2176ec803fa9da6201cdd9b8146940a2c00cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b0412f3330999c3348ac726ac5a0c2d541571f4b13786ca24cb3109a44aa653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://318cb69e50df4c14df9bf5ef57d5c381421a9bdb72135c9b9620f55c0547b410\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0f15d331af9bc2ca3082c1573dd167c66773eb390593844e681f8d0b455d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2f16fe4bf2d659398ec141ba12d2dda32727810b03b96bfaf491602656027cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b03acfee4df70e62846f679407e6f9960e44946d5fd4b71d7ea1011729eb8c05\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T09:04:48Z\\\",\\\"message\\\":\\\"ddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-controller-manager/kube-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager/kube-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.36\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0218 09:04:48.895203 6239 lb_config.go:1031] Cluster endpoints for openshift-machine-api/machine-api-controllers for network=default are: map[]\\\\nI0218 09:04:48.895220 6239 services_controller.go:443] Built service openshift-machine-api/machine-api-controllers LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.167\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8441, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]s\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2f16fe4bf2d659398ec141ba12d2dda32727810b03b96bfaf491602656027cb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T09:05:16Z\\\",\\\"message\\\":\\\"d to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:16Z is after 2025-08-24T17:21:41Z]\\\\nI0218 09:05:16.892059 6632 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/package-server-manager-metrics]} name:Service_openshift-operator-lifecycle-manager/package-server-manager-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.110:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {f9232b32-e89f-4c8e-acc4-c6801b70dcb0}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:NB_Global Row:map[] Rows:[] Columns:[] Mutations:[{Column:nb_cfg Mutator:+= Value:1}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {6011affd-30a6-4be6-872d-e4cf1ca780cf}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0218 09:05:16.891870 6632 handler.go:208] Rem\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aec6030df0a59ad8902d5c694d5cd30d249ca1e895155c57a458e359fa531b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7fsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:17Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.678373 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4e7c95a-56e9-4e22-83dd-54962e1233b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52ea6a334e8cad154e46eb2f917cdac914557556bae88f922433a9545cb29eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28a600e8daaa2796b2f0cc6ce4112f8d41c611b5b6ca5a7eb0f260dab615caa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1fa733477ee6330835d52333d6e9588626b6fc7ae08ae450dd2e1821447e995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfbda9799e2d2bc16964e79b90fb7087f9e6893d867711612c63aaaa3dd2b141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfbda9799e2d2bc16964e79b90fb7087f9e6893d867711612c63aaaa3dd2b141\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:17Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.684696 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jflp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689d3d82-4071-4fe5-b3e3-b5b4f511580a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d56d65964aaf12feaaa3e5fa3ca779bffdb13fcebfc9f027b563c7dd01f4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bm85s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jflp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:17Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.692633 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8f6014-2544-4173-a94d-441cf42ba891\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77b0269625b7c6baab44a40373618f4ccfe9e1a028a1bc746b7d60c20ac4efaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac09c421b3101887bfb83107dbe26f2b5856d9b32a9cea92f6233fb6ae218cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce79cfe0242ecf403f989dc6b1c1b8be5ab54eee75ac413a8f9d03cef8bca0ed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://095ecf7e8763fa098b06eb9a8aac9f39451f7762bf20196e0803bbc841610e02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:17Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.700749 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:17Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.711515 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d653c21d3d7c875bed37ed6da700dc8c801b95c80d7e361121dfbbd3d05a922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5211d0cd5172d45b92b003e2bf97d7969a65ea448b0491f8f6ea67364925ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:17Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.719773 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6387f58543cf2601a53b0bbe9be3ec7816ba90d5ffdc4465594b37ccac0f6dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:17Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.727518 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:17Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.734299 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bp8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"804f8152-54ae-4486-92e6-f9595c035820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb617e46b8522d73dabef4bbc660e41b737f6e6e62f7ce42dad722490a1041dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phr7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://104b8ff3c41d4304d414e58303c23792a7f0ba580f73a2a8678044741af71380\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phr7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bp8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:17Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.746597 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ae60f3-adab-4740-821a-1570059c7101\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f629262e27020512301a7b53ee836ee98e1e7cdee7a52b0ed506bbaaebd823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f59306c0d55e9d00714ce2ee2c1ed6f2ed33d8b21eecde490e86a93335a14f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c87bc671742a47de6c8e68aaf0c8645f9a8cae7b3d52264f36ae3827b3a8d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2678001ff826712da6716541210f018c74fd3d0218b388d2e7a6f16a1f4027\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6f37933281c927328ecfaf842d97ee8361726768f393726cb39517d2bdc9425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:17Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.756765 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.756791 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.756799 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.756811 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.756819 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:17Z","lastTransitionTime":"2026-02-18T09:05:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.858392 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.858424 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.858432 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.858442 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.858464 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:17Z","lastTransitionTime":"2026-02-18T09:05:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.960434 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.960465 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.960474 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.960486 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:17 crc kubenswrapper[4556]: I0218 09:05:17.960496 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:17Z","lastTransitionTime":"2026-02-18T09:05:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.068628 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.068883 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.068893 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.068905 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.068917 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:18Z","lastTransitionTime":"2026-02-18T09:05:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.170682 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.170715 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.170724 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.170737 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.170746 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:18Z","lastTransitionTime":"2026-02-18T09:05:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.272776 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.272809 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.272817 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.272828 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.272838 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:18Z","lastTransitionTime":"2026-02-18T09:05:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.282037 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cgwv9" Feb 18 09:05:18 crc kubenswrapper[4556]: E0218 09:05:18.282144 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cgwv9" podUID="d2f0e36c-cb29-4b06-bb22-6afd59466cab" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.290250 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-15 01:06:25.714024421 +0000 UTC Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.374313 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.374340 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.374348 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.374358 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.374366 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:18Z","lastTransitionTime":"2026-02-18T09:05:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.476553 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.476574 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.476582 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.476592 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.476600 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:18Z","lastTransitionTime":"2026-02-18T09:05:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.578013 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.578034 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.578042 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.578051 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.578060 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:18Z","lastTransitionTime":"2026-02-18T09:05:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.595329 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-b7fsj_09971522-31bc-4e10-8831-4fb62675e220/ovnkube-controller/3.log" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.597708 4556 scope.go:117] "RemoveContainer" containerID="c2f16fe4bf2d659398ec141ba12d2dda32727810b03b96bfaf491602656027cb" Feb 18 09:05:18 crc kubenswrapper[4556]: E0218 09:05:18.597838 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-b7fsj_openshift-ovn-kubernetes(09971522-31bc-4e10-8831-4fb62675e220)\"" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" podUID="09971522-31bc-4e10-8831-4fb62675e220" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.607503 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qfksk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e226b189-cfa9-47fc-b94d-19b5cbe0859f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23f488cd416c6d49acf193e27beba9e9f1dcfb36430f9ae540742c2e2bd0f144\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc2cb92a7735d16b8e5918797e52e1e488d9261a0418ed5ebbb94c033ca9328a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T09:05:12Z\\\",\\\"message\\\":\\\"2026-02-18T09:04:27+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_3fa1df1f-8463-485c-a4c9-cd0c6dd4d227\\\\n2026-02-18T09:04:27+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_3fa1df1f-8463-485c-a4c9-cd0c6dd4d227 to /host/opt/cni/bin/\\\\n2026-02-18T09:04:27Z [verbose] multus-daemon started\\\\n2026-02-18T09:04:27Z [verbose] Readiness Indicator file check\\\\n2026-02-18T09:05:12Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:05:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2fx4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qfksk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:18Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.615313 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dac7f27-d3d1-4778-9e54-f273035a1d37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18da54fe071964d8aae8a977f5ee209c31f1c75e1041c44f8d755f58f3e39d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d41b85076e695bbc51a41fce200e0e47c099d88b4d825b4d133d33ed96237e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f76hs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:18Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.624523 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sgcq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9945b792-abbd-4103-a7b4-9496e1cc1b56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5dd402827182bc79283f1ea1c355b60c4ced53ffa9ce6781b0a0db3109359a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac0d9d7b9bc867b32cf6cbeb804c08e67df757e4b22914e2da69a376511cf75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac0d9d7b9bc867b32cf6cbeb804c08e67df757e4b22914e2da69a376511cf75c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b807d53d620e44d3196f68492b9f9665593adac804bc7b91569796e5594c4ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b807d53d620e44d3196f68492b9f9665593adac804bc7b91569796e5594c4ebb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebc044bbd5b0c1e725e6ed46ec2e0015b254070b7518d20815539dcf8c6e4e45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ebc044bbd5b0c1e725e6ed46ec2e0015b254070b7518d20815539dcf8c6e4e45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sgcq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:18Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.630917 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cgwv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2f0e36c-cb29-4b06-bb22-6afd59466cab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpldv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpldv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:40Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cgwv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:18Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.639258 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"302eb215-a4c2-4324-836d-f37debfd620f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a99dc228d1c4cb8aeab2140ec5f37358b65684739f833fbde45822782430a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14f794063db9d33b82b79acc7134b5a2a3435e09d7cffe49d49251d1f7d20402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5eb3d3728da247903bece56bc2481547317e386e9e97b6c69926a3b55dc89b8e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b42d13c1767c29d22e52537ff45561a0fbf8f2069b2c3612e903e8c127e12485\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e865e6a50c03519ce22958ac4fcbb911fc140c330b6d7ffe5c9d7dd381e96\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T09:04:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0218 09:04:15.143085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 09:04:15.144650 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2160144045/tls.crt::/tmp/serving-cert-2160144045/tls.key\\\\\\\"\\\\nI0218 09:04:20.680799 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 09:04:20.682555 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 09:04:20.682573 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 09:04:20.682590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 09:04:20.682594 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 09:04:20.686822 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 09:04:20.686847 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686852 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686856 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 09:04:20.686859 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 09:04:20.686861 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 09:04:20.686864 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 09:04:20.688257 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 09:04:20.690726 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://974e17ae7082631df2d1bcee60c8a5f142bae2659e9d200d537a4d650f401b4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:18Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.646813 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:18Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.654064 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47341f31f94eaf5dd86efbe78ff62873e733b1f86ea895f281b6797770519459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:18Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.660757 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5vcgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d50c39aa-f956-441a-8abe-1d7247a7fd86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ddc5d42163d1fa86f1a59e224adac434c990224e839ef3cb12c6d7689e2b439\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4llw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5vcgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:18Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.672518 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09971522-31bc-4e10-8831-4fb62675e220\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9170bdcbaf1065995d03e6448e40bdced0741c716c9e81f37a72b015132f1161\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c526c0613f19e1a1a584b7129bd56001faf08d0b97fd77feeb69f764ed01f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cd6e758ee396c9000b1bdb6ded2176ec803fa9da6201cdd9b8146940a2c00cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b0412f3330999c3348ac726ac5a0c2d541571f4b13786ca24cb3109a44aa653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://318cb69e50df4c14df9bf5ef57d5c381421a9bdb72135c9b9620f55c0547b410\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0f15d331af9bc2ca3082c1573dd167c66773eb390593844e681f8d0b455d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2f16fe4bf2d659398ec141ba12d2dda32727810b03b96bfaf491602656027cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2f16fe4bf2d659398ec141ba12d2dda32727810b03b96bfaf491602656027cb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T09:05:16Z\\\",\\\"message\\\":\\\"d to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:16Z is after 2025-08-24T17:21:41Z]\\\\nI0218 09:05:16.892059 6632 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/package-server-manager-metrics]} name:Service_openshift-operator-lifecycle-manager/package-server-manager-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.110:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {f9232b32-e89f-4c8e-acc4-c6801b70dcb0}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:NB_Global Row:map[] Rows:[] Columns:[] Mutations:[{Column:nb_cfg Mutator:+= Value:1}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {6011affd-30a6-4be6-872d-e4cf1ca780cf}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0218 09:05:16.891870 6632 handler.go:208] Rem\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:05:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-b7fsj_openshift-ovn-kubernetes(09971522-31bc-4e10-8831-4fb62675e220)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aec6030df0a59ad8902d5c694d5cd30d249ca1e895155c57a458e359fa531b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7fsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:18Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.679583 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.679614 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.679625 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.679641 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.679651 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:18Z","lastTransitionTime":"2026-02-18T09:05:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.679682 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4e7c95a-56e9-4e22-83dd-54962e1233b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52ea6a334e8cad154e46eb2f917cdac914557556bae88f922433a9545cb29eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28a600e8daaa2796b2f0cc6ce4112f8d41c611b5b6ca5a7eb0f260dab615caa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1fa733477ee6330835d52333d6e9588626b6fc7ae08ae450dd2e1821447e995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfbda9799e2d2bc16964e79b90fb7087f9e6893d867711612c63aaaa3dd2b141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfbda9799e2d2bc16964e79b90fb7087f9e6893d867711612c63aaaa3dd2b141\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:18Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.686035 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64e03bf0-d2fc-40ba-a468-4d9a2b43afc7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ac8fc80676085932e05da4ff18e892eebaec46d096c7c343b667864f6c5e3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6e0422278544bddfec087d7b0a94bedc55a7ac67ff280a36030528eb87fd72b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6e0422278544bddfec087d7b0a94bedc55a7ac67ff280a36030528eb87fd72b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:18Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.692107 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jflp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689d3d82-4071-4fe5-b3e3-b5b4f511580a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d56d65964aaf12feaaa3e5fa3ca779bffdb13fcebfc9f027b563c7dd01f4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bm85s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jflp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:18Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.700244 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:18Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.708139 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d653c21d3d7c875bed37ed6da700dc8c801b95c80d7e361121dfbbd3d05a922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5211d0cd5172d45b92b003e2bf97d7969a65ea448b0491f8f6ea67364925ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:18Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.716439 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6387f58543cf2601a53b0bbe9be3ec7816ba90d5ffdc4465594b37ccac0f6dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:18Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.724464 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:18Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.731547 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bp8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"804f8152-54ae-4486-92e6-f9595c035820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb617e46b8522d73dabef4bbc660e41b737f6e6e62f7ce42dad722490a1041dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phr7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://104b8ff3c41d4304d414e58303c23792a7f0ba580f73a2a8678044741af71380\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phr7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bp8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:18Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.743861 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ae60f3-adab-4740-821a-1570059c7101\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f629262e27020512301a7b53ee836ee98e1e7cdee7a52b0ed506bbaaebd823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f59306c0d55e9d00714ce2ee2c1ed6f2ed33d8b21eecde490e86a93335a14f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c87bc671742a47de6c8e68aaf0c8645f9a8cae7b3d52264f36ae3827b3a8d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2678001ff826712da6716541210f018c74fd3d0218b388d2e7a6f16a1f4027\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6f37933281c927328ecfaf842d97ee8361726768f393726cb39517d2bdc9425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:18Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.751729 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8f6014-2544-4173-a94d-441cf42ba891\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77b0269625b7c6baab44a40373618f4ccfe9e1a028a1bc746b7d60c20ac4efaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac09c421b3101887bfb83107dbe26f2b5856d9b32a9cea92f6233fb6ae218cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce79cfe0242ecf403f989dc6b1c1b8be5ab54eee75ac413a8f9d03cef8bca0ed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://095ecf7e8763fa098b06eb9a8aac9f39451f7762bf20196e0803bbc841610e02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:18Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.781610 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.781644 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.781653 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.781666 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.781675 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:18Z","lastTransitionTime":"2026-02-18T09:05:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.883015 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.883075 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.883086 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.883099 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.883109 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:18Z","lastTransitionTime":"2026-02-18T09:05:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.984564 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.984590 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.984598 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.984609 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:18 crc kubenswrapper[4556]: I0218 09:05:18.984617 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:18Z","lastTransitionTime":"2026-02-18T09:05:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:19 crc kubenswrapper[4556]: I0218 09:05:19.086358 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:19 crc kubenswrapper[4556]: I0218 09:05:19.086385 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:19 crc kubenswrapper[4556]: I0218 09:05:19.086393 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:19 crc kubenswrapper[4556]: I0218 09:05:19.086413 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:19 crc kubenswrapper[4556]: I0218 09:05:19.086421 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:19Z","lastTransitionTime":"2026-02-18T09:05:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:19 crc kubenswrapper[4556]: I0218 09:05:19.190206 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:19 crc kubenswrapper[4556]: I0218 09:05:19.190589 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:19 crc kubenswrapper[4556]: I0218 09:05:19.190602 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:19 crc kubenswrapper[4556]: I0218 09:05:19.190615 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:19 crc kubenswrapper[4556]: I0218 09:05:19.190624 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:19Z","lastTransitionTime":"2026-02-18T09:05:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:19 crc kubenswrapper[4556]: I0218 09:05:19.282038 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:05:19 crc kubenswrapper[4556]: I0218 09:05:19.282079 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:05:19 crc kubenswrapper[4556]: I0218 09:05:19.282082 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:05:19 crc kubenswrapper[4556]: E0218 09:05:19.282134 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 09:05:19 crc kubenswrapper[4556]: E0218 09:05:19.282243 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 09:05:19 crc kubenswrapper[4556]: E0218 09:05:19.282288 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 09:05:19 crc kubenswrapper[4556]: I0218 09:05:19.290798 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-27 17:55:41.004677448 +0000 UTC Feb 18 09:05:19 crc kubenswrapper[4556]: I0218 09:05:19.291704 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:19 crc kubenswrapper[4556]: I0218 09:05:19.291727 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:19 crc kubenswrapper[4556]: I0218 09:05:19.291735 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:19 crc kubenswrapper[4556]: I0218 09:05:19.291745 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:19 crc kubenswrapper[4556]: I0218 09:05:19.291754 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:19Z","lastTransitionTime":"2026-02-18T09:05:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:19 crc kubenswrapper[4556]: I0218 09:05:19.393615 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:19 crc kubenswrapper[4556]: I0218 09:05:19.393646 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:19 crc kubenswrapper[4556]: I0218 09:05:19.393655 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:19 crc kubenswrapper[4556]: I0218 09:05:19.393668 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:19 crc kubenswrapper[4556]: I0218 09:05:19.393676 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:19Z","lastTransitionTime":"2026-02-18T09:05:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:19 crc kubenswrapper[4556]: I0218 09:05:19.495802 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:19 crc kubenswrapper[4556]: I0218 09:05:19.495930 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:19 crc kubenswrapper[4556]: I0218 09:05:19.496010 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:19 crc kubenswrapper[4556]: I0218 09:05:19.496070 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:19 crc kubenswrapper[4556]: I0218 09:05:19.496132 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:19Z","lastTransitionTime":"2026-02-18T09:05:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:19 crc kubenswrapper[4556]: I0218 09:05:19.597879 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:19 crc kubenswrapper[4556]: I0218 09:05:19.597975 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:19 crc kubenswrapper[4556]: I0218 09:05:19.598031 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:19 crc kubenswrapper[4556]: I0218 09:05:19.598088 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:19 crc kubenswrapper[4556]: I0218 09:05:19.598161 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:19Z","lastTransitionTime":"2026-02-18T09:05:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:19 crc kubenswrapper[4556]: I0218 09:05:19.699761 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:19 crc kubenswrapper[4556]: I0218 09:05:19.699792 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:19 crc kubenswrapper[4556]: I0218 09:05:19.699800 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:19 crc kubenswrapper[4556]: I0218 09:05:19.699809 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:19 crc kubenswrapper[4556]: I0218 09:05:19.699817 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:19Z","lastTransitionTime":"2026-02-18T09:05:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:19 crc kubenswrapper[4556]: I0218 09:05:19.801302 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:19 crc kubenswrapper[4556]: I0218 09:05:19.801336 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:19 crc kubenswrapper[4556]: I0218 09:05:19.801345 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:19 crc kubenswrapper[4556]: I0218 09:05:19.801357 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:19 crc kubenswrapper[4556]: I0218 09:05:19.801367 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:19Z","lastTransitionTime":"2026-02-18T09:05:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:19 crc kubenswrapper[4556]: I0218 09:05:19.903010 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:19 crc kubenswrapper[4556]: I0218 09:05:19.903059 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:19 crc kubenswrapper[4556]: I0218 09:05:19.903069 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:19 crc kubenswrapper[4556]: I0218 09:05:19.903081 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:19 crc kubenswrapper[4556]: I0218 09:05:19.903093 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:19Z","lastTransitionTime":"2026-02-18T09:05:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:20 crc kubenswrapper[4556]: I0218 09:05:20.004624 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:20 crc kubenswrapper[4556]: I0218 09:05:20.004655 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:20 crc kubenswrapper[4556]: I0218 09:05:20.004664 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:20 crc kubenswrapper[4556]: I0218 09:05:20.004675 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:20 crc kubenswrapper[4556]: I0218 09:05:20.004684 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:20Z","lastTransitionTime":"2026-02-18T09:05:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:20 crc kubenswrapper[4556]: I0218 09:05:20.106436 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:20 crc kubenswrapper[4556]: I0218 09:05:20.106467 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:20 crc kubenswrapper[4556]: I0218 09:05:20.106492 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:20 crc kubenswrapper[4556]: I0218 09:05:20.106505 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:20 crc kubenswrapper[4556]: I0218 09:05:20.106514 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:20Z","lastTransitionTime":"2026-02-18T09:05:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:20 crc kubenswrapper[4556]: I0218 09:05:20.207609 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:20 crc kubenswrapper[4556]: I0218 09:05:20.207649 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:20 crc kubenswrapper[4556]: I0218 09:05:20.207659 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:20 crc kubenswrapper[4556]: I0218 09:05:20.207673 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:20 crc kubenswrapper[4556]: I0218 09:05:20.207683 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:20Z","lastTransitionTime":"2026-02-18T09:05:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:20 crc kubenswrapper[4556]: I0218 09:05:20.281708 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cgwv9" Feb 18 09:05:20 crc kubenswrapper[4556]: E0218 09:05:20.281817 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cgwv9" podUID="d2f0e36c-cb29-4b06-bb22-6afd59466cab" Feb 18 09:05:20 crc kubenswrapper[4556]: I0218 09:05:20.291176 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-11 16:39:57.47205822 +0000 UTC Feb 18 09:05:20 crc kubenswrapper[4556]: I0218 09:05:20.309334 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:20 crc kubenswrapper[4556]: I0218 09:05:20.309380 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:20 crc kubenswrapper[4556]: I0218 09:05:20.309391 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:20 crc kubenswrapper[4556]: I0218 09:05:20.309410 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:20 crc kubenswrapper[4556]: I0218 09:05:20.309418 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:20Z","lastTransitionTime":"2026-02-18T09:05:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:20 crc kubenswrapper[4556]: I0218 09:05:20.410517 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:20 crc kubenswrapper[4556]: I0218 09:05:20.410549 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:20 crc kubenswrapper[4556]: I0218 09:05:20.410558 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:20 crc kubenswrapper[4556]: I0218 09:05:20.410569 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:20 crc kubenswrapper[4556]: I0218 09:05:20.410577 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:20Z","lastTransitionTime":"2026-02-18T09:05:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:20 crc kubenswrapper[4556]: I0218 09:05:20.512184 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:20 crc kubenswrapper[4556]: I0218 09:05:20.512223 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:20 crc kubenswrapper[4556]: I0218 09:05:20.512232 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:20 crc kubenswrapper[4556]: I0218 09:05:20.512245 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:20 crc kubenswrapper[4556]: I0218 09:05:20.512254 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:20Z","lastTransitionTime":"2026-02-18T09:05:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:20 crc kubenswrapper[4556]: I0218 09:05:20.613892 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:20 crc kubenswrapper[4556]: I0218 09:05:20.613930 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:20 crc kubenswrapper[4556]: I0218 09:05:20.613941 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:20 crc kubenswrapper[4556]: I0218 09:05:20.613954 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:20 crc kubenswrapper[4556]: I0218 09:05:20.613963 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:20Z","lastTransitionTime":"2026-02-18T09:05:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:20 crc kubenswrapper[4556]: I0218 09:05:20.715774 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:20 crc kubenswrapper[4556]: I0218 09:05:20.715808 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:20 crc kubenswrapper[4556]: I0218 09:05:20.715816 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:20 crc kubenswrapper[4556]: I0218 09:05:20.715828 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:20 crc kubenswrapper[4556]: I0218 09:05:20.715839 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:20Z","lastTransitionTime":"2026-02-18T09:05:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:20 crc kubenswrapper[4556]: I0218 09:05:20.817714 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:20 crc kubenswrapper[4556]: I0218 09:05:20.817752 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:20 crc kubenswrapper[4556]: I0218 09:05:20.817763 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:20 crc kubenswrapper[4556]: I0218 09:05:20.817777 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:20 crc kubenswrapper[4556]: I0218 09:05:20.817787 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:20Z","lastTransitionTime":"2026-02-18T09:05:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:20 crc kubenswrapper[4556]: I0218 09:05:20.919925 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:20 crc kubenswrapper[4556]: I0218 09:05:20.919964 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:20 crc kubenswrapper[4556]: I0218 09:05:20.919972 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:20 crc kubenswrapper[4556]: I0218 09:05:20.919986 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:20 crc kubenswrapper[4556]: I0218 09:05:20.919996 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:20Z","lastTransitionTime":"2026-02-18T09:05:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:21 crc kubenswrapper[4556]: I0218 09:05:21.021483 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:21 crc kubenswrapper[4556]: I0218 09:05:21.021511 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:21 crc kubenswrapper[4556]: I0218 09:05:21.021519 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:21 crc kubenswrapper[4556]: I0218 09:05:21.021531 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:21 crc kubenswrapper[4556]: I0218 09:05:21.021539 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:21Z","lastTransitionTime":"2026-02-18T09:05:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:21 crc kubenswrapper[4556]: I0218 09:05:21.123475 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:21 crc kubenswrapper[4556]: I0218 09:05:21.123516 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:21 crc kubenswrapper[4556]: I0218 09:05:21.123528 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:21 crc kubenswrapper[4556]: I0218 09:05:21.123540 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:21 crc kubenswrapper[4556]: I0218 09:05:21.123549 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:21Z","lastTransitionTime":"2026-02-18T09:05:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:21 crc kubenswrapper[4556]: I0218 09:05:21.225858 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:21 crc kubenswrapper[4556]: I0218 09:05:21.225892 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:21 crc kubenswrapper[4556]: I0218 09:05:21.225901 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:21 crc kubenswrapper[4556]: I0218 09:05:21.225913 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:21 crc kubenswrapper[4556]: I0218 09:05:21.225921 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:21Z","lastTransitionTime":"2026-02-18T09:05:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:21 crc kubenswrapper[4556]: I0218 09:05:21.281932 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:05:21 crc kubenswrapper[4556]: I0218 09:05:21.281967 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:05:21 crc kubenswrapper[4556]: E0218 09:05:21.282042 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 09:05:21 crc kubenswrapper[4556]: I0218 09:05:21.282148 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:05:21 crc kubenswrapper[4556]: E0218 09:05:21.282259 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 09:05:21 crc kubenswrapper[4556]: E0218 09:05:21.282308 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 09:05:21 crc kubenswrapper[4556]: I0218 09:05:21.291789 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-07 19:33:51.037405851 +0000 UTC Feb 18 09:05:21 crc kubenswrapper[4556]: I0218 09:05:21.327958 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:21 crc kubenswrapper[4556]: I0218 09:05:21.327984 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:21 crc kubenswrapper[4556]: I0218 09:05:21.327993 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:21 crc kubenswrapper[4556]: I0218 09:05:21.328001 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:21 crc kubenswrapper[4556]: I0218 09:05:21.328008 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:21Z","lastTransitionTime":"2026-02-18T09:05:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:21 crc kubenswrapper[4556]: I0218 09:05:21.430056 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:21 crc kubenswrapper[4556]: I0218 09:05:21.430084 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:21 crc kubenswrapper[4556]: I0218 09:05:21.430091 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:21 crc kubenswrapper[4556]: I0218 09:05:21.430100 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:21 crc kubenswrapper[4556]: I0218 09:05:21.430108 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:21Z","lastTransitionTime":"2026-02-18T09:05:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:21 crc kubenswrapper[4556]: I0218 09:05:21.532616 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:21 crc kubenswrapper[4556]: I0218 09:05:21.532652 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:21 crc kubenswrapper[4556]: I0218 09:05:21.532661 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:21 crc kubenswrapper[4556]: I0218 09:05:21.532672 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:21 crc kubenswrapper[4556]: I0218 09:05:21.532680 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:21Z","lastTransitionTime":"2026-02-18T09:05:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:21 crc kubenswrapper[4556]: I0218 09:05:21.633919 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:21 crc kubenswrapper[4556]: I0218 09:05:21.633955 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:21 crc kubenswrapper[4556]: I0218 09:05:21.633964 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:21 crc kubenswrapper[4556]: I0218 09:05:21.633977 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:21 crc kubenswrapper[4556]: I0218 09:05:21.633987 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:21Z","lastTransitionTime":"2026-02-18T09:05:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:21 crc kubenswrapper[4556]: I0218 09:05:21.736515 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:21 crc kubenswrapper[4556]: I0218 09:05:21.736548 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:21 crc kubenswrapper[4556]: I0218 09:05:21.736558 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:21 crc kubenswrapper[4556]: I0218 09:05:21.736571 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:21 crc kubenswrapper[4556]: I0218 09:05:21.736580 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:21Z","lastTransitionTime":"2026-02-18T09:05:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:21 crc kubenswrapper[4556]: I0218 09:05:21.838681 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:21 crc kubenswrapper[4556]: I0218 09:05:21.838714 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:21 crc kubenswrapper[4556]: I0218 09:05:21.838722 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:21 crc kubenswrapper[4556]: I0218 09:05:21.838734 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:21 crc kubenswrapper[4556]: I0218 09:05:21.838743 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:21Z","lastTransitionTime":"2026-02-18T09:05:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:21 crc kubenswrapper[4556]: I0218 09:05:21.940951 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:21 crc kubenswrapper[4556]: I0218 09:05:21.940985 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:21 crc kubenswrapper[4556]: I0218 09:05:21.940993 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:21 crc kubenswrapper[4556]: I0218 09:05:21.941005 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:21 crc kubenswrapper[4556]: I0218 09:05:21.941014 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:21Z","lastTransitionTime":"2026-02-18T09:05:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:22 crc kubenswrapper[4556]: I0218 09:05:22.043056 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:22 crc kubenswrapper[4556]: I0218 09:05:22.043086 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:22 crc kubenswrapper[4556]: I0218 09:05:22.043094 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:22 crc kubenswrapper[4556]: I0218 09:05:22.043104 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:22 crc kubenswrapper[4556]: I0218 09:05:22.043111 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:22Z","lastTransitionTime":"2026-02-18T09:05:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:22 crc kubenswrapper[4556]: I0218 09:05:22.145565 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:22 crc kubenswrapper[4556]: I0218 09:05:22.145601 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:22 crc kubenswrapper[4556]: I0218 09:05:22.145608 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:22 crc kubenswrapper[4556]: I0218 09:05:22.145622 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:22 crc kubenswrapper[4556]: I0218 09:05:22.145631 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:22Z","lastTransitionTime":"2026-02-18T09:05:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:22 crc kubenswrapper[4556]: I0218 09:05:22.247790 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:22 crc kubenswrapper[4556]: I0218 09:05:22.247826 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:22 crc kubenswrapper[4556]: I0218 09:05:22.247835 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:22 crc kubenswrapper[4556]: I0218 09:05:22.247848 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:22 crc kubenswrapper[4556]: I0218 09:05:22.247858 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:22Z","lastTransitionTime":"2026-02-18T09:05:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:22 crc kubenswrapper[4556]: I0218 09:05:22.282003 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cgwv9" Feb 18 09:05:22 crc kubenswrapper[4556]: E0218 09:05:22.282105 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cgwv9" podUID="d2f0e36c-cb29-4b06-bb22-6afd59466cab" Feb 18 09:05:22 crc kubenswrapper[4556]: I0218 09:05:22.292255 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-08 03:47:02.069776361 +0000 UTC Feb 18 09:05:22 crc kubenswrapper[4556]: I0218 09:05:22.349222 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:22 crc kubenswrapper[4556]: I0218 09:05:22.349250 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:22 crc kubenswrapper[4556]: I0218 09:05:22.349259 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:22 crc kubenswrapper[4556]: I0218 09:05:22.349269 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:22 crc kubenswrapper[4556]: I0218 09:05:22.349278 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:22Z","lastTransitionTime":"2026-02-18T09:05:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:22 crc kubenswrapper[4556]: I0218 09:05:22.451302 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:22 crc kubenswrapper[4556]: I0218 09:05:22.451332 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:22 crc kubenswrapper[4556]: I0218 09:05:22.451340 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:22 crc kubenswrapper[4556]: I0218 09:05:22.451353 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:22 crc kubenswrapper[4556]: I0218 09:05:22.451362 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:22Z","lastTransitionTime":"2026-02-18T09:05:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:22 crc kubenswrapper[4556]: I0218 09:05:22.553488 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:22 crc kubenswrapper[4556]: I0218 09:05:22.553515 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:22 crc kubenswrapper[4556]: I0218 09:05:22.553523 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:22 crc kubenswrapper[4556]: I0218 09:05:22.553533 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:22 crc kubenswrapper[4556]: I0218 09:05:22.553542 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:22Z","lastTransitionTime":"2026-02-18T09:05:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:22 crc kubenswrapper[4556]: I0218 09:05:22.654883 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:22 crc kubenswrapper[4556]: I0218 09:05:22.654930 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:22 crc kubenswrapper[4556]: I0218 09:05:22.654940 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:22 crc kubenswrapper[4556]: I0218 09:05:22.654952 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:22 crc kubenswrapper[4556]: I0218 09:05:22.654960 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:22Z","lastTransitionTime":"2026-02-18T09:05:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:22 crc kubenswrapper[4556]: I0218 09:05:22.756146 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:22 crc kubenswrapper[4556]: I0218 09:05:22.756189 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:22 crc kubenswrapper[4556]: I0218 09:05:22.756198 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:22 crc kubenswrapper[4556]: I0218 09:05:22.756210 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:22 crc kubenswrapper[4556]: I0218 09:05:22.756220 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:22Z","lastTransitionTime":"2026-02-18T09:05:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:22 crc kubenswrapper[4556]: I0218 09:05:22.858043 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:22 crc kubenswrapper[4556]: I0218 09:05:22.858075 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:22 crc kubenswrapper[4556]: I0218 09:05:22.858083 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:22 crc kubenswrapper[4556]: I0218 09:05:22.858094 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:22 crc kubenswrapper[4556]: I0218 09:05:22.858104 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:22Z","lastTransitionTime":"2026-02-18T09:05:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:22 crc kubenswrapper[4556]: I0218 09:05:22.959997 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:22 crc kubenswrapper[4556]: I0218 09:05:22.960029 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:22 crc kubenswrapper[4556]: I0218 09:05:22.960055 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:22 crc kubenswrapper[4556]: I0218 09:05:22.960066 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:22 crc kubenswrapper[4556]: I0218 09:05:22.960076 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:22Z","lastTransitionTime":"2026-02-18T09:05:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.062041 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.062079 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.062089 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.062102 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.062111 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:23Z","lastTransitionTime":"2026-02-18T09:05:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.164365 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.164413 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.164425 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.164440 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.164450 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:23Z","lastTransitionTime":"2026-02-18T09:05:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.266297 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.266437 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.266496 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.266555 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.266611 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:23Z","lastTransitionTime":"2026-02-18T09:05:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.281788 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:05:23 crc kubenswrapper[4556]: E0218 09:05:23.281969 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.281867 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.281802 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:05:23 crc kubenswrapper[4556]: E0218 09:05:23.282433 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 09:05:23 crc kubenswrapper[4556]: E0218 09:05:23.282559 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.290034 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5vcgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d50c39aa-f956-441a-8abe-1d7247a7fd86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ddc5d42163d1fa86f1a59e224adac434c990224e839ef3cb12c6d7689e2b439\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4llw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5vcgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:23Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.292508 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-03 01:10:59.23281906 +0000 UTC Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.303098 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09971522-31bc-4e10-8831-4fb62675e220\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9170bdcbaf1065995d03e6448e40bdced0741c716c9e81f37a72b015132f1161\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c526c0613f19e1a1a584b7129bd56001faf08d0b97fd77feeb69f764ed01f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cd6e758ee396c9000b1bdb6ded2176ec803fa9da6201cdd9b8146940a2c00cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b0412f3330999c3348ac726ac5a0c2d541571f4b13786ca24cb3109a44aa653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://318cb69e50df4c14df9bf5ef57d5c381421a9bdb72135c9b9620f55c0547b410\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0f15d331af9bc2ca3082c1573dd167c66773eb390593844e681f8d0b455d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2f16fe4bf2d659398ec141ba12d2dda32727810b03b96bfaf491602656027cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2f16fe4bf2d659398ec141ba12d2dda32727810b03b96bfaf491602656027cb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T09:05:16Z\\\",\\\"message\\\":\\\"d to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:16Z is after 2025-08-24T17:21:41Z]\\\\nI0218 09:05:16.892059 6632 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/package-server-manager-metrics]} name:Service_openshift-operator-lifecycle-manager/package-server-manager-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.110:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {f9232b32-e89f-4c8e-acc4-c6801b70dcb0}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:NB_Global Row:map[] Rows:[] Columns:[] Mutations:[{Column:nb_cfg Mutator:+= Value:1}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {6011affd-30a6-4be6-872d-e4cf1ca780cf}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0218 09:05:16.891870 6632 handler.go:208] Rem\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:05:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-b7fsj_openshift-ovn-kubernetes(09971522-31bc-4e10-8831-4fb62675e220)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aec6030df0a59ad8902d5c694d5cd30d249ca1e895155c57a458e359fa531b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7fsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:23Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.310685 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4e7c95a-56e9-4e22-83dd-54962e1233b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52ea6a334e8cad154e46eb2f917cdac914557556bae88f922433a9545cb29eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28a600e8daaa2796b2f0cc6ce4112f8d41c611b5b6ca5a7eb0f260dab615caa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1fa733477ee6330835d52333d6e9588626b6fc7ae08ae450dd2e1821447e995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfbda9799e2d2bc16964e79b90fb7087f9e6893d867711612c63aaaa3dd2b141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfbda9799e2d2bc16964e79b90fb7087f9e6893d867711612c63aaaa3dd2b141\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:23Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.317585 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64e03bf0-d2fc-40ba-a468-4d9a2b43afc7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ac8fc80676085932e05da4ff18e892eebaec46d096c7c343b667864f6c5e3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6e0422278544bddfec087d7b0a94bedc55a7ac67ff280a36030528eb87fd72b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6e0422278544bddfec087d7b0a94bedc55a7ac67ff280a36030528eb87fd72b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:23Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.324615 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47341f31f94eaf5dd86efbe78ff62873e733b1f86ea895f281b6797770519459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:23Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.331364 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jflp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689d3d82-4071-4fe5-b3e3-b5b4f511580a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d56d65964aaf12feaaa3e5fa3ca779bffdb13fcebfc9f027b563c7dd01f4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bm85s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jflp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:23Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.340139 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d653c21d3d7c875bed37ed6da700dc8c801b95c80d7e361121dfbbd3d05a922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5211d0cd5172d45b92b003e2bf97d7969a65ea448b0491f8f6ea67364925ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:23Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.350536 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6387f58543cf2601a53b0bbe9be3ec7816ba90d5ffdc4465594b37ccac0f6dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:23Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.357935 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:23Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.364714 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bp8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"804f8152-54ae-4486-92e6-f9595c035820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb617e46b8522d73dabef4bbc660e41b737f6e6e62f7ce42dad722490a1041dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phr7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://104b8ff3c41d4304d414e58303c23792a7f0ba580f73a2a8678044741af71380\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phr7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bp8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:23Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.367758 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.367790 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.367798 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.367808 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.367817 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:23Z","lastTransitionTime":"2026-02-18T09:05:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.376657 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ae60f3-adab-4740-821a-1570059c7101\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f629262e27020512301a7b53ee836ee98e1e7cdee7a52b0ed506bbaaebd823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f59306c0d55e9d00714ce2ee2c1ed6f2ed33d8b21eecde490e86a93335a14f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c87bc671742a47de6c8e68aaf0c8645f9a8cae7b3d52264f36ae3827b3a8d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2678001ff826712da6716541210f018c74fd3d0218b388d2e7a6f16a1f4027\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6f37933281c927328ecfaf842d97ee8361726768f393726cb39517d2bdc9425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:23Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.384414 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8f6014-2544-4173-a94d-441cf42ba891\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77b0269625b7c6baab44a40373618f4ccfe9e1a028a1bc746b7d60c20ac4efaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac09c421b3101887bfb83107dbe26f2b5856d9b32a9cea92f6233fb6ae218cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce79cfe0242ecf403f989dc6b1c1b8be5ab54eee75ac413a8f9d03cef8bca0ed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://095ecf7e8763fa098b06eb9a8aac9f39451f7762bf20196e0803bbc841610e02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:23Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.392471 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:23Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.400629 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dac7f27-d3d1-4778-9e54-f273035a1d37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18da54fe071964d8aae8a977f5ee209c31f1c75e1041c44f8d755f58f3e39d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d41b85076e695bbc51a41fce200e0e47c099d88b4d825b4d133d33ed96237e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f76hs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:23Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.409643 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sgcq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9945b792-abbd-4103-a7b4-9496e1cc1b56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5dd402827182bc79283f1ea1c355b60c4ced53ffa9ce6781b0a0db3109359a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac0d9d7b9bc867b32cf6cbeb804c08e67df757e4b22914e2da69a376511cf75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac0d9d7b9bc867b32cf6cbeb804c08e67df757e4b22914e2da69a376511cf75c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b807d53d620e44d3196f68492b9f9665593adac804bc7b91569796e5594c4ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b807d53d620e44d3196f68492b9f9665593adac804bc7b91569796e5594c4ebb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebc044bbd5b0c1e725e6ed46ec2e0015b254070b7518d20815539dcf8c6e4e45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ebc044bbd5b0c1e725e6ed46ec2e0015b254070b7518d20815539dcf8c6e4e45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sgcq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:23Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.418440 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cgwv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2f0e36c-cb29-4b06-bb22-6afd59466cab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpldv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpldv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:40Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cgwv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:23Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.429404 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"302eb215-a4c2-4324-836d-f37debfd620f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a99dc228d1c4cb8aeab2140ec5f37358b65684739f833fbde45822782430a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14f794063db9d33b82b79acc7134b5a2a3435e09d7cffe49d49251d1f7d20402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5eb3d3728da247903bece56bc2481547317e386e9e97b6c69926a3b55dc89b8e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b42d13c1767c29d22e52537ff45561a0fbf8f2069b2c3612e903e8c127e12485\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e865e6a50c03519ce22958ac4fcbb911fc140c330b6d7ffe5c9d7dd381e96\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T09:04:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0218 09:04:15.143085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 09:04:15.144650 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2160144045/tls.crt::/tmp/serving-cert-2160144045/tls.key\\\\\\\"\\\\nI0218 09:04:20.680799 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 09:04:20.682555 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 09:04:20.682573 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 09:04:20.682590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 09:04:20.682594 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 09:04:20.686822 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 09:04:20.686847 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686852 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686856 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 09:04:20.686859 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 09:04:20.686861 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 09:04:20.686864 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 09:04:20.688257 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 09:04:20.690726 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://974e17ae7082631df2d1bcee60c8a5f142bae2659e9d200d537a4d650f401b4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:23Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.436794 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:23Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.444645 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qfksk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e226b189-cfa9-47fc-b94d-19b5cbe0859f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23f488cd416c6d49acf193e27beba9e9f1dcfb36430f9ae540742c2e2bd0f144\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc2cb92a7735d16b8e5918797e52e1e488d9261a0418ed5ebbb94c033ca9328a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T09:05:12Z\\\",\\\"message\\\":\\\"2026-02-18T09:04:27+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_3fa1df1f-8463-485c-a4c9-cd0c6dd4d227\\\\n2026-02-18T09:04:27+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_3fa1df1f-8463-485c-a4c9-cd0c6dd4d227 to /host/opt/cni/bin/\\\\n2026-02-18T09:04:27Z [verbose] multus-daemon started\\\\n2026-02-18T09:04:27Z [verbose] Readiness Indicator file check\\\\n2026-02-18T09:05:12Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:05:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2fx4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qfksk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:23Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.469993 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.470020 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.470028 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.470038 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.470047 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:23Z","lastTransitionTime":"2026-02-18T09:05:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.571211 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.571322 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.571387 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.571454 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.571511 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:23Z","lastTransitionTime":"2026-02-18T09:05:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.672812 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.672867 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.672877 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.672905 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.672918 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:23Z","lastTransitionTime":"2026-02-18T09:05:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.774533 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.774642 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.774707 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.774766 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.774817 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:23Z","lastTransitionTime":"2026-02-18T09:05:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.876452 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.876487 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.876498 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.876511 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.876519 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:23Z","lastTransitionTime":"2026-02-18T09:05:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.978310 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.978342 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.978353 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.978365 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:23 crc kubenswrapper[4556]: I0218 09:05:23.978385 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:23Z","lastTransitionTime":"2026-02-18T09:05:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:24 crc kubenswrapper[4556]: I0218 09:05:24.080184 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:24 crc kubenswrapper[4556]: I0218 09:05:24.080210 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:24 crc kubenswrapper[4556]: I0218 09:05:24.080219 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:24 crc kubenswrapper[4556]: I0218 09:05:24.080230 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:24 crc kubenswrapper[4556]: I0218 09:05:24.080237 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:24Z","lastTransitionTime":"2026-02-18T09:05:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:24 crc kubenswrapper[4556]: I0218 09:05:24.182232 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:24 crc kubenswrapper[4556]: I0218 09:05:24.182280 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:24 crc kubenswrapper[4556]: I0218 09:05:24.182289 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:24 crc kubenswrapper[4556]: I0218 09:05:24.182300 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:24 crc kubenswrapper[4556]: I0218 09:05:24.182308 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:24Z","lastTransitionTime":"2026-02-18T09:05:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:24 crc kubenswrapper[4556]: I0218 09:05:24.282314 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cgwv9" Feb 18 09:05:24 crc kubenswrapper[4556]: E0218 09:05:24.282425 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cgwv9" podUID="d2f0e36c-cb29-4b06-bb22-6afd59466cab" Feb 18 09:05:24 crc kubenswrapper[4556]: I0218 09:05:24.283402 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:24 crc kubenswrapper[4556]: I0218 09:05:24.283427 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:24 crc kubenswrapper[4556]: I0218 09:05:24.283437 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:24 crc kubenswrapper[4556]: I0218 09:05:24.283449 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:24 crc kubenswrapper[4556]: I0218 09:05:24.283457 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:24Z","lastTransitionTime":"2026-02-18T09:05:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:24 crc kubenswrapper[4556]: I0218 09:05:24.292685 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-15 06:42:16.532784477 +0000 UTC Feb 18 09:05:24 crc kubenswrapper[4556]: I0218 09:05:24.384684 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:24 crc kubenswrapper[4556]: I0218 09:05:24.384710 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:24 crc kubenswrapper[4556]: I0218 09:05:24.384720 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:24 crc kubenswrapper[4556]: I0218 09:05:24.384757 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:24 crc kubenswrapper[4556]: I0218 09:05:24.384766 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:24Z","lastTransitionTime":"2026-02-18T09:05:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:24 crc kubenswrapper[4556]: I0218 09:05:24.486359 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:24 crc kubenswrapper[4556]: I0218 09:05:24.486414 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:24 crc kubenswrapper[4556]: I0218 09:05:24.486446 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:24 crc kubenswrapper[4556]: I0218 09:05:24.486458 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:24 crc kubenswrapper[4556]: I0218 09:05:24.486467 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:24Z","lastTransitionTime":"2026-02-18T09:05:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:24 crc kubenswrapper[4556]: I0218 09:05:24.588399 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:24 crc kubenswrapper[4556]: I0218 09:05:24.588426 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:24 crc kubenswrapper[4556]: I0218 09:05:24.588455 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:24 crc kubenswrapper[4556]: I0218 09:05:24.588466 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:24 crc kubenswrapper[4556]: I0218 09:05:24.588473 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:24Z","lastTransitionTime":"2026-02-18T09:05:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:24 crc kubenswrapper[4556]: I0218 09:05:24.690065 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:24 crc kubenswrapper[4556]: I0218 09:05:24.690093 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:24 crc kubenswrapper[4556]: I0218 09:05:24.690101 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:24 crc kubenswrapper[4556]: I0218 09:05:24.690113 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:24 crc kubenswrapper[4556]: I0218 09:05:24.690123 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:24Z","lastTransitionTime":"2026-02-18T09:05:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:24 crc kubenswrapper[4556]: I0218 09:05:24.791486 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:24 crc kubenswrapper[4556]: I0218 09:05:24.791578 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:24 crc kubenswrapper[4556]: I0218 09:05:24.791647 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:24 crc kubenswrapper[4556]: I0218 09:05:24.791711 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:24 crc kubenswrapper[4556]: I0218 09:05:24.791763 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:24Z","lastTransitionTime":"2026-02-18T09:05:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:24 crc kubenswrapper[4556]: I0218 09:05:24.893209 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:24 crc kubenswrapper[4556]: I0218 09:05:24.893233 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:24 crc kubenswrapper[4556]: I0218 09:05:24.893242 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:24 crc kubenswrapper[4556]: I0218 09:05:24.893412 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:24 crc kubenswrapper[4556]: I0218 09:05:24.893435 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:24Z","lastTransitionTime":"2026-02-18T09:05:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:24 crc kubenswrapper[4556]: I0218 09:05:24.999907 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:24 crc kubenswrapper[4556]: I0218 09:05:24.999931 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:24 crc kubenswrapper[4556]: I0218 09:05:24.999939 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:24 crc kubenswrapper[4556]: I0218 09:05:24.999948 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:24.999956 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:24Z","lastTransitionTime":"2026-02-18T09:05:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.102134 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.102185 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.102198 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.102210 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.102219 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:25Z","lastTransitionTime":"2026-02-18T09:05:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.171825 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.171888 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.171918 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:05:25 crc kubenswrapper[4556]: E0218 09:05:25.171944 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 09:06:29.171920445 +0000 UTC m=+146.188881455 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:05:25 crc kubenswrapper[4556]: E0218 09:05:25.171970 4556 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 18 09:05:25 crc kubenswrapper[4556]: E0218 09:05:25.172013 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-18 09:06:29.17200264 +0000 UTC m=+146.188963620 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 18 09:05:25 crc kubenswrapper[4556]: E0218 09:05:25.172032 4556 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 18 09:05:25 crc kubenswrapper[4556]: E0218 09:05:25.172066 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-18 09:06:29.17205626 +0000 UTC m=+146.189017241 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.204234 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.204256 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.204267 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.204279 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.204286 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:25Z","lastTransitionTime":"2026-02-18T09:05:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.273234 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.273278 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:05:25 crc kubenswrapper[4556]: E0218 09:05:25.273378 4556 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 18 09:05:25 crc kubenswrapper[4556]: E0218 09:05:25.273395 4556 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 18 09:05:25 crc kubenswrapper[4556]: E0218 09:05:25.273404 4556 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 18 09:05:25 crc kubenswrapper[4556]: E0218 09:05:25.273431 4556 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 18 09:05:25 crc kubenswrapper[4556]: E0218 09:05:25.273442 4556 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 09:05:25 crc kubenswrapper[4556]: E0218 09:05:25.273472 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-18 09:06:29.273464591 +0000 UTC m=+146.290425570 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 09:05:25 crc kubenswrapper[4556]: E0218 09:05:25.273407 4556 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 09:05:25 crc kubenswrapper[4556]: E0218 09:05:25.273535 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-18 09:06:29.273527118 +0000 UTC m=+146.290488098 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.282314 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.282343 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:05:25 crc kubenswrapper[4556]: E0218 09:05:25.282412 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.282426 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:05:25 crc kubenswrapper[4556]: E0218 09:05:25.282533 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 09:05:25 crc kubenswrapper[4556]: E0218 09:05:25.282638 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.293001 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-16 13:40:51.094080227 +0000 UTC Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.306236 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.306262 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.306271 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.306280 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.306288 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:25Z","lastTransitionTime":"2026-02-18T09:05:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.333434 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.333471 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.333481 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.333494 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.333502 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:25Z","lastTransitionTime":"2026-02-18T09:05:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:25 crc kubenswrapper[4556]: E0218 09:05:25.341957 4556 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:05:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:05:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:05:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:05:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8758c02e-ef59-46a7-908c-d5c97feb8ceb\\\",\\\"systemUUID\\\":\\\"0f904183-79bf-4790-8551-43a6eb3adbe4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:25Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.344165 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.344193 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.344203 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.344215 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.344223 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:25Z","lastTransitionTime":"2026-02-18T09:05:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:25 crc kubenswrapper[4556]: E0218 09:05:25.352670 4556 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:05:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:05:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:05:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:05:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8758c02e-ef59-46a7-908c-d5c97feb8ceb\\\",\\\"systemUUID\\\":\\\"0f904183-79bf-4790-8551-43a6eb3adbe4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:25Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.354793 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.354875 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.354932 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.354996 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.355056 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:25Z","lastTransitionTime":"2026-02-18T09:05:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:25 crc kubenswrapper[4556]: E0218 09:05:25.362973 4556 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:05:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:05:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:05:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:05:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8758c02e-ef59-46a7-908c-d5c97feb8ceb\\\",\\\"systemUUID\\\":\\\"0f904183-79bf-4790-8551-43a6eb3adbe4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:25Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.365070 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.365178 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.365248 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.365307 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.365381 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:25Z","lastTransitionTime":"2026-02-18T09:05:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:25 crc kubenswrapper[4556]: E0218 09:05:25.372916 4556 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:05:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:05:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:05:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:05:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8758c02e-ef59-46a7-908c-d5c97feb8ceb\\\",\\\"systemUUID\\\":\\\"0f904183-79bf-4790-8551-43a6eb3adbe4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:25Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.374886 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.374907 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.374915 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.374926 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.374934 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:25Z","lastTransitionTime":"2026-02-18T09:05:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:25 crc kubenswrapper[4556]: E0218 09:05:25.382706 4556 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:05:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:05:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:05:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T09:05:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8758c02e-ef59-46a7-908c-d5c97feb8ceb\\\",\\\"systemUUID\\\":\\\"0f904183-79bf-4790-8551-43a6eb3adbe4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:25Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:25 crc kubenswrapper[4556]: E0218 09:05:25.382807 4556 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.408759 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.408852 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.408909 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.408984 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.409041 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:25Z","lastTransitionTime":"2026-02-18T09:05:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.510814 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.510855 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.510869 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.510881 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.510892 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:25Z","lastTransitionTime":"2026-02-18T09:05:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.612249 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.612380 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.612456 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.612512 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.612571 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:25Z","lastTransitionTime":"2026-02-18T09:05:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.713954 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.713988 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.714000 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.714012 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.714021 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:25Z","lastTransitionTime":"2026-02-18T09:05:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.815975 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.816015 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.816026 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.816039 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.816048 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:25Z","lastTransitionTime":"2026-02-18T09:05:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.918003 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.918031 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.918039 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.918048 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:25 crc kubenswrapper[4556]: I0218 09:05:25.918056 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:25Z","lastTransitionTime":"2026-02-18T09:05:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:26 crc kubenswrapper[4556]: I0218 09:05:26.019746 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:26 crc kubenswrapper[4556]: I0218 09:05:26.019779 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:26 crc kubenswrapper[4556]: I0218 09:05:26.019806 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:26 crc kubenswrapper[4556]: I0218 09:05:26.019818 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:26 crc kubenswrapper[4556]: I0218 09:05:26.019827 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:26Z","lastTransitionTime":"2026-02-18T09:05:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:26 crc kubenswrapper[4556]: I0218 09:05:26.121461 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:26 crc kubenswrapper[4556]: I0218 09:05:26.121493 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:26 crc kubenswrapper[4556]: I0218 09:05:26.121502 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:26 crc kubenswrapper[4556]: I0218 09:05:26.121529 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:26 crc kubenswrapper[4556]: I0218 09:05:26.121538 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:26Z","lastTransitionTime":"2026-02-18T09:05:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:26 crc kubenswrapper[4556]: I0218 09:05:26.223594 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:26 crc kubenswrapper[4556]: I0218 09:05:26.223624 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:26 crc kubenswrapper[4556]: I0218 09:05:26.223632 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:26 crc kubenswrapper[4556]: I0218 09:05:26.223643 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:26 crc kubenswrapper[4556]: I0218 09:05:26.223651 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:26Z","lastTransitionTime":"2026-02-18T09:05:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:26 crc kubenswrapper[4556]: I0218 09:05:26.281720 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cgwv9" Feb 18 09:05:26 crc kubenswrapper[4556]: E0218 09:05:26.281820 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cgwv9" podUID="d2f0e36c-cb29-4b06-bb22-6afd59466cab" Feb 18 09:05:26 crc kubenswrapper[4556]: I0218 09:05:26.293799 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 09:42:33.993491287 +0000 UTC Feb 18 09:05:26 crc kubenswrapper[4556]: I0218 09:05:26.325041 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:26 crc kubenswrapper[4556]: I0218 09:05:26.325066 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:26 crc kubenswrapper[4556]: I0218 09:05:26.325083 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:26 crc kubenswrapper[4556]: I0218 09:05:26.325092 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:26 crc kubenswrapper[4556]: I0218 09:05:26.325100 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:26Z","lastTransitionTime":"2026-02-18T09:05:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:26 crc kubenswrapper[4556]: I0218 09:05:26.426841 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:26 crc kubenswrapper[4556]: I0218 09:05:26.426936 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:26 crc kubenswrapper[4556]: I0218 09:05:26.427004 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:26 crc kubenswrapper[4556]: I0218 09:05:26.427061 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:26 crc kubenswrapper[4556]: I0218 09:05:26.427117 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:26Z","lastTransitionTime":"2026-02-18T09:05:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:26 crc kubenswrapper[4556]: I0218 09:05:26.528463 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:26 crc kubenswrapper[4556]: I0218 09:05:26.528487 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:26 crc kubenswrapper[4556]: I0218 09:05:26.528496 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:26 crc kubenswrapper[4556]: I0218 09:05:26.528507 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:26 crc kubenswrapper[4556]: I0218 09:05:26.528515 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:26Z","lastTransitionTime":"2026-02-18T09:05:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:26 crc kubenswrapper[4556]: I0218 09:05:26.629484 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:26 crc kubenswrapper[4556]: I0218 09:05:26.629575 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:26 crc kubenswrapper[4556]: I0218 09:05:26.629655 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:26 crc kubenswrapper[4556]: I0218 09:05:26.629731 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:26 crc kubenswrapper[4556]: I0218 09:05:26.629785 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:26Z","lastTransitionTime":"2026-02-18T09:05:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:26 crc kubenswrapper[4556]: I0218 09:05:26.731363 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:26 crc kubenswrapper[4556]: I0218 09:05:26.731398 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:26 crc kubenswrapper[4556]: I0218 09:05:26.731407 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:26 crc kubenswrapper[4556]: I0218 09:05:26.731420 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:26 crc kubenswrapper[4556]: I0218 09:05:26.731434 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:26Z","lastTransitionTime":"2026-02-18T09:05:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:26 crc kubenswrapper[4556]: I0218 09:05:26.832956 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:26 crc kubenswrapper[4556]: I0218 09:05:26.833006 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:26 crc kubenswrapper[4556]: I0218 09:05:26.833014 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:26 crc kubenswrapper[4556]: I0218 09:05:26.833030 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:26 crc kubenswrapper[4556]: I0218 09:05:26.833039 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:26Z","lastTransitionTime":"2026-02-18T09:05:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:26 crc kubenswrapper[4556]: I0218 09:05:26.934656 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:26 crc kubenswrapper[4556]: I0218 09:05:26.934764 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:26 crc kubenswrapper[4556]: I0218 09:05:26.934823 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:26 crc kubenswrapper[4556]: I0218 09:05:26.934880 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:26 crc kubenswrapper[4556]: I0218 09:05:26.934932 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:26Z","lastTransitionTime":"2026-02-18T09:05:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:27 crc kubenswrapper[4556]: I0218 09:05:27.036447 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:27 crc kubenswrapper[4556]: I0218 09:05:27.036487 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:27 crc kubenswrapper[4556]: I0218 09:05:27.036500 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:27 crc kubenswrapper[4556]: I0218 09:05:27.036529 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:27 crc kubenswrapper[4556]: I0218 09:05:27.036538 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:27Z","lastTransitionTime":"2026-02-18T09:05:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:27 crc kubenswrapper[4556]: I0218 09:05:27.138417 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:27 crc kubenswrapper[4556]: I0218 09:05:27.138447 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:27 crc kubenswrapper[4556]: I0218 09:05:27.138455 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:27 crc kubenswrapper[4556]: I0218 09:05:27.138466 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:27 crc kubenswrapper[4556]: I0218 09:05:27.138474 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:27Z","lastTransitionTime":"2026-02-18T09:05:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:27 crc kubenswrapper[4556]: I0218 09:05:27.240135 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:27 crc kubenswrapper[4556]: I0218 09:05:27.240184 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:27 crc kubenswrapper[4556]: I0218 09:05:27.240193 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:27 crc kubenswrapper[4556]: I0218 09:05:27.240204 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:27 crc kubenswrapper[4556]: I0218 09:05:27.240211 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:27Z","lastTransitionTime":"2026-02-18T09:05:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:27 crc kubenswrapper[4556]: I0218 09:05:27.281647 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:05:27 crc kubenswrapper[4556]: E0218 09:05:27.281732 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 09:05:27 crc kubenswrapper[4556]: I0218 09:05:27.281835 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:05:27 crc kubenswrapper[4556]: E0218 09:05:27.281985 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 09:05:27 crc kubenswrapper[4556]: I0218 09:05:27.281862 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:05:27 crc kubenswrapper[4556]: E0218 09:05:27.282216 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 09:05:27 crc kubenswrapper[4556]: I0218 09:05:27.293917 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-08 19:03:50.61392306 +0000 UTC Feb 18 09:05:27 crc kubenswrapper[4556]: I0218 09:05:27.342333 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:27 crc kubenswrapper[4556]: I0218 09:05:27.342373 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:27 crc kubenswrapper[4556]: I0218 09:05:27.342382 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:27 crc kubenswrapper[4556]: I0218 09:05:27.342396 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:27 crc kubenswrapper[4556]: I0218 09:05:27.342405 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:27Z","lastTransitionTime":"2026-02-18T09:05:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:27 crc kubenswrapper[4556]: I0218 09:05:27.444089 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:27 crc kubenswrapper[4556]: I0218 09:05:27.444193 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:27 crc kubenswrapper[4556]: I0218 09:05:27.444204 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:27 crc kubenswrapper[4556]: I0218 09:05:27.444229 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:27 crc kubenswrapper[4556]: I0218 09:05:27.444639 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:27Z","lastTransitionTime":"2026-02-18T09:05:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:27 crc kubenswrapper[4556]: I0218 09:05:27.546445 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:27 crc kubenswrapper[4556]: I0218 09:05:27.546499 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:27 crc kubenswrapper[4556]: I0218 09:05:27.546511 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:27 crc kubenswrapper[4556]: I0218 09:05:27.546521 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:27 crc kubenswrapper[4556]: I0218 09:05:27.546529 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:27Z","lastTransitionTime":"2026-02-18T09:05:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:27 crc kubenswrapper[4556]: I0218 09:05:27.648624 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:27 crc kubenswrapper[4556]: I0218 09:05:27.648647 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:27 crc kubenswrapper[4556]: I0218 09:05:27.648654 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:27 crc kubenswrapper[4556]: I0218 09:05:27.648662 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:27 crc kubenswrapper[4556]: I0218 09:05:27.648669 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:27Z","lastTransitionTime":"2026-02-18T09:05:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:27 crc kubenswrapper[4556]: I0218 09:05:27.750117 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:27 crc kubenswrapper[4556]: I0218 09:05:27.750145 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:27 crc kubenswrapper[4556]: I0218 09:05:27.750172 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:27 crc kubenswrapper[4556]: I0218 09:05:27.750182 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:27 crc kubenswrapper[4556]: I0218 09:05:27.750189 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:27Z","lastTransitionTime":"2026-02-18T09:05:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:27 crc kubenswrapper[4556]: I0218 09:05:27.851471 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:27 crc kubenswrapper[4556]: I0218 09:05:27.851505 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:27 crc kubenswrapper[4556]: I0218 09:05:27.851513 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:27 crc kubenswrapper[4556]: I0218 09:05:27.851544 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:27 crc kubenswrapper[4556]: I0218 09:05:27.851554 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:27Z","lastTransitionTime":"2026-02-18T09:05:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:27 crc kubenswrapper[4556]: I0218 09:05:27.953095 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:27 crc kubenswrapper[4556]: I0218 09:05:27.953140 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:27 crc kubenswrapper[4556]: I0218 09:05:27.953182 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:27 crc kubenswrapper[4556]: I0218 09:05:27.953200 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:27 crc kubenswrapper[4556]: I0218 09:05:27.953212 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:27Z","lastTransitionTime":"2026-02-18T09:05:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:28 crc kubenswrapper[4556]: I0218 09:05:28.055195 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:28 crc kubenswrapper[4556]: I0218 09:05:28.055222 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:28 crc kubenswrapper[4556]: I0218 09:05:28.055231 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:28 crc kubenswrapper[4556]: I0218 09:05:28.055240 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:28 crc kubenswrapper[4556]: I0218 09:05:28.055261 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:28Z","lastTransitionTime":"2026-02-18T09:05:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:28 crc kubenswrapper[4556]: I0218 09:05:28.157308 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:28 crc kubenswrapper[4556]: I0218 09:05:28.157337 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:28 crc kubenswrapper[4556]: I0218 09:05:28.157353 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:28 crc kubenswrapper[4556]: I0218 09:05:28.157363 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:28 crc kubenswrapper[4556]: I0218 09:05:28.157370 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:28Z","lastTransitionTime":"2026-02-18T09:05:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:28 crc kubenswrapper[4556]: I0218 09:05:28.258924 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:28 crc kubenswrapper[4556]: I0218 09:05:28.258950 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:28 crc kubenswrapper[4556]: I0218 09:05:28.258959 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:28 crc kubenswrapper[4556]: I0218 09:05:28.258968 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:28 crc kubenswrapper[4556]: I0218 09:05:28.258978 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:28Z","lastTransitionTime":"2026-02-18T09:05:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:28 crc kubenswrapper[4556]: I0218 09:05:28.282262 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cgwv9" Feb 18 09:05:28 crc kubenswrapper[4556]: E0218 09:05:28.282361 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cgwv9" podUID="d2f0e36c-cb29-4b06-bb22-6afd59466cab" Feb 18 09:05:28 crc kubenswrapper[4556]: I0218 09:05:28.294569 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-25 18:08:25.901603273 +0000 UTC Feb 18 09:05:28 crc kubenswrapper[4556]: I0218 09:05:28.360844 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:28 crc kubenswrapper[4556]: I0218 09:05:28.360892 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:28 crc kubenswrapper[4556]: I0218 09:05:28.360901 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:28 crc kubenswrapper[4556]: I0218 09:05:28.360912 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:28 crc kubenswrapper[4556]: I0218 09:05:28.360920 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:28Z","lastTransitionTime":"2026-02-18T09:05:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:28 crc kubenswrapper[4556]: I0218 09:05:28.463180 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:28 crc kubenswrapper[4556]: I0218 09:05:28.463283 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:28 crc kubenswrapper[4556]: I0218 09:05:28.463363 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:28 crc kubenswrapper[4556]: I0218 09:05:28.463422 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:28 crc kubenswrapper[4556]: I0218 09:05:28.463480 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:28Z","lastTransitionTime":"2026-02-18T09:05:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:28 crc kubenswrapper[4556]: I0218 09:05:28.565325 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:28 crc kubenswrapper[4556]: I0218 09:05:28.565510 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:28 crc kubenswrapper[4556]: I0218 09:05:28.565582 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:28 crc kubenswrapper[4556]: I0218 09:05:28.565649 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:28 crc kubenswrapper[4556]: I0218 09:05:28.565713 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:28Z","lastTransitionTime":"2026-02-18T09:05:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:28 crc kubenswrapper[4556]: I0218 09:05:28.667921 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:28 crc kubenswrapper[4556]: I0218 09:05:28.668063 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:28 crc kubenswrapper[4556]: I0218 09:05:28.668137 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:28 crc kubenswrapper[4556]: I0218 09:05:28.668221 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:28 crc kubenswrapper[4556]: I0218 09:05:28.668281 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:28Z","lastTransitionTime":"2026-02-18T09:05:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:28 crc kubenswrapper[4556]: I0218 09:05:28.770004 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:28 crc kubenswrapper[4556]: I0218 09:05:28.770105 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:28 crc kubenswrapper[4556]: I0218 09:05:28.770184 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:28 crc kubenswrapper[4556]: I0218 09:05:28.770261 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:28 crc kubenswrapper[4556]: I0218 09:05:28.770316 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:28Z","lastTransitionTime":"2026-02-18T09:05:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:28 crc kubenswrapper[4556]: I0218 09:05:28.872033 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:28 crc kubenswrapper[4556]: I0218 09:05:28.872123 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:28 crc kubenswrapper[4556]: I0218 09:05:28.872212 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:28 crc kubenswrapper[4556]: I0218 09:05:28.872277 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:28 crc kubenswrapper[4556]: I0218 09:05:28.872350 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:28Z","lastTransitionTime":"2026-02-18T09:05:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:28 crc kubenswrapper[4556]: I0218 09:05:28.973544 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:28 crc kubenswrapper[4556]: I0218 09:05:28.973587 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:28 crc kubenswrapper[4556]: I0218 09:05:28.973619 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:28 crc kubenswrapper[4556]: I0218 09:05:28.973635 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:28 crc kubenswrapper[4556]: I0218 09:05:28.973646 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:28Z","lastTransitionTime":"2026-02-18T09:05:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:29 crc kubenswrapper[4556]: I0218 09:05:29.075489 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:29 crc kubenswrapper[4556]: I0218 09:05:29.075517 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:29 crc kubenswrapper[4556]: I0218 09:05:29.075526 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:29 crc kubenswrapper[4556]: I0218 09:05:29.075553 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:29 crc kubenswrapper[4556]: I0218 09:05:29.075562 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:29Z","lastTransitionTime":"2026-02-18T09:05:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:29 crc kubenswrapper[4556]: I0218 09:05:29.177541 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:29 crc kubenswrapper[4556]: I0218 09:05:29.177570 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:29 crc kubenswrapper[4556]: I0218 09:05:29.177579 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:29 crc kubenswrapper[4556]: I0218 09:05:29.177592 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:29 crc kubenswrapper[4556]: I0218 09:05:29.177601 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:29Z","lastTransitionTime":"2026-02-18T09:05:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:29 crc kubenswrapper[4556]: I0218 09:05:29.279218 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:29 crc kubenswrapper[4556]: I0218 09:05:29.279258 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:29 crc kubenswrapper[4556]: I0218 09:05:29.279267 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:29 crc kubenswrapper[4556]: I0218 09:05:29.279277 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:29 crc kubenswrapper[4556]: I0218 09:05:29.279285 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:29Z","lastTransitionTime":"2026-02-18T09:05:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:29 crc kubenswrapper[4556]: I0218 09:05:29.281683 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:05:29 crc kubenswrapper[4556]: E0218 09:05:29.281784 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 09:05:29 crc kubenswrapper[4556]: I0218 09:05:29.281954 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:05:29 crc kubenswrapper[4556]: E0218 09:05:29.282023 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 09:05:29 crc kubenswrapper[4556]: I0218 09:05:29.282125 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:05:29 crc kubenswrapper[4556]: E0218 09:05:29.282228 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 09:05:29 crc kubenswrapper[4556]: I0218 09:05:29.294866 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-29 05:12:18.628000418 +0000 UTC Feb 18 09:05:29 crc kubenswrapper[4556]: I0218 09:05:29.380861 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:29 crc kubenswrapper[4556]: I0218 09:05:29.380889 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:29 crc kubenswrapper[4556]: I0218 09:05:29.380898 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:29 crc kubenswrapper[4556]: I0218 09:05:29.380909 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:29 crc kubenswrapper[4556]: I0218 09:05:29.380917 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:29Z","lastTransitionTime":"2026-02-18T09:05:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:29 crc kubenswrapper[4556]: I0218 09:05:29.482674 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:29 crc kubenswrapper[4556]: I0218 09:05:29.482855 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:29 crc kubenswrapper[4556]: I0218 09:05:29.482987 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:29 crc kubenswrapper[4556]: I0218 09:05:29.483099 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:29 crc kubenswrapper[4556]: I0218 09:05:29.483226 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:29Z","lastTransitionTime":"2026-02-18T09:05:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:29 crc kubenswrapper[4556]: I0218 09:05:29.584851 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:29 crc kubenswrapper[4556]: I0218 09:05:29.584880 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:29 crc kubenswrapper[4556]: I0218 09:05:29.584888 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:29 crc kubenswrapper[4556]: I0218 09:05:29.584898 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:29 crc kubenswrapper[4556]: I0218 09:05:29.584906 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:29Z","lastTransitionTime":"2026-02-18T09:05:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:29 crc kubenswrapper[4556]: I0218 09:05:29.689208 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:29 crc kubenswrapper[4556]: I0218 09:05:29.689241 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:29 crc kubenswrapper[4556]: I0218 09:05:29.689251 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:29 crc kubenswrapper[4556]: I0218 09:05:29.689320 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:29 crc kubenswrapper[4556]: I0218 09:05:29.689351 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:29Z","lastTransitionTime":"2026-02-18T09:05:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:29 crc kubenswrapper[4556]: I0218 09:05:29.790669 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:29 crc kubenswrapper[4556]: I0218 09:05:29.790696 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:29 crc kubenswrapper[4556]: I0218 09:05:29.790705 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:29 crc kubenswrapper[4556]: I0218 09:05:29.790715 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:29 crc kubenswrapper[4556]: I0218 09:05:29.790722 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:29Z","lastTransitionTime":"2026-02-18T09:05:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:29 crc kubenswrapper[4556]: I0218 09:05:29.892431 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:29 crc kubenswrapper[4556]: I0218 09:05:29.892460 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:29 crc kubenswrapper[4556]: I0218 09:05:29.892468 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:29 crc kubenswrapper[4556]: I0218 09:05:29.892478 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:29 crc kubenswrapper[4556]: I0218 09:05:29.892486 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:29Z","lastTransitionTime":"2026-02-18T09:05:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:29 crc kubenswrapper[4556]: I0218 09:05:29.993737 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:29 crc kubenswrapper[4556]: I0218 09:05:29.993765 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:29 crc kubenswrapper[4556]: I0218 09:05:29.993775 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:29 crc kubenswrapper[4556]: I0218 09:05:29.993791 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:29 crc kubenswrapper[4556]: I0218 09:05:29.993800 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:29Z","lastTransitionTime":"2026-02-18T09:05:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:30 crc kubenswrapper[4556]: I0218 09:05:30.095589 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:30 crc kubenswrapper[4556]: I0218 09:05:30.095637 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:30 crc kubenswrapper[4556]: I0218 09:05:30.095647 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:30 crc kubenswrapper[4556]: I0218 09:05:30.095656 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:30 crc kubenswrapper[4556]: I0218 09:05:30.095666 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:30Z","lastTransitionTime":"2026-02-18T09:05:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:30 crc kubenswrapper[4556]: I0218 09:05:30.196933 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:30 crc kubenswrapper[4556]: I0218 09:05:30.196960 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:30 crc kubenswrapper[4556]: I0218 09:05:30.196969 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:30 crc kubenswrapper[4556]: I0218 09:05:30.196977 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:30 crc kubenswrapper[4556]: I0218 09:05:30.197001 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:30Z","lastTransitionTime":"2026-02-18T09:05:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:30 crc kubenswrapper[4556]: I0218 09:05:30.281858 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cgwv9" Feb 18 09:05:30 crc kubenswrapper[4556]: E0218 09:05:30.281943 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cgwv9" podUID="d2f0e36c-cb29-4b06-bb22-6afd59466cab" Feb 18 09:05:30 crc kubenswrapper[4556]: I0218 09:05:30.295004 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-03 08:55:55.918306106 +0000 UTC Feb 18 09:05:30 crc kubenswrapper[4556]: I0218 09:05:30.298021 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:30 crc kubenswrapper[4556]: I0218 09:05:30.298053 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:30 crc kubenswrapper[4556]: I0218 09:05:30.298062 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:30 crc kubenswrapper[4556]: I0218 09:05:30.298071 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:30 crc kubenswrapper[4556]: I0218 09:05:30.298080 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:30Z","lastTransitionTime":"2026-02-18T09:05:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:30 crc kubenswrapper[4556]: I0218 09:05:30.400236 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:30 crc kubenswrapper[4556]: I0218 09:05:30.400274 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:30 crc kubenswrapper[4556]: I0218 09:05:30.400284 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:30 crc kubenswrapper[4556]: I0218 09:05:30.400298 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:30 crc kubenswrapper[4556]: I0218 09:05:30.400308 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:30Z","lastTransitionTime":"2026-02-18T09:05:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:30 crc kubenswrapper[4556]: I0218 09:05:30.501734 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:30 crc kubenswrapper[4556]: I0218 09:05:30.501762 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:30 crc kubenswrapper[4556]: I0218 09:05:30.501771 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:30 crc kubenswrapper[4556]: I0218 09:05:30.501782 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:30 crc kubenswrapper[4556]: I0218 09:05:30.501790 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:30Z","lastTransitionTime":"2026-02-18T09:05:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:30 crc kubenswrapper[4556]: I0218 09:05:30.603216 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:30 crc kubenswrapper[4556]: I0218 09:05:30.603247 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:30 crc kubenswrapper[4556]: I0218 09:05:30.603256 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:30 crc kubenswrapper[4556]: I0218 09:05:30.603267 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:30 crc kubenswrapper[4556]: I0218 09:05:30.603274 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:30Z","lastTransitionTime":"2026-02-18T09:05:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:30 crc kubenswrapper[4556]: I0218 09:05:30.704563 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:30 crc kubenswrapper[4556]: I0218 09:05:30.704592 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:30 crc kubenswrapper[4556]: I0218 09:05:30.704599 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:30 crc kubenswrapper[4556]: I0218 09:05:30.704610 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:30 crc kubenswrapper[4556]: I0218 09:05:30.704620 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:30Z","lastTransitionTime":"2026-02-18T09:05:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:30 crc kubenswrapper[4556]: I0218 09:05:30.806540 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:30 crc kubenswrapper[4556]: I0218 09:05:30.806579 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:30 crc kubenswrapper[4556]: I0218 09:05:30.806589 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:30 crc kubenswrapper[4556]: I0218 09:05:30.806605 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:30 crc kubenswrapper[4556]: I0218 09:05:30.806617 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:30Z","lastTransitionTime":"2026-02-18T09:05:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:30 crc kubenswrapper[4556]: I0218 09:05:30.908657 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:30 crc kubenswrapper[4556]: I0218 09:05:30.908691 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:30 crc kubenswrapper[4556]: I0218 09:05:30.908701 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:30 crc kubenswrapper[4556]: I0218 09:05:30.908714 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:30 crc kubenswrapper[4556]: I0218 09:05:30.908726 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:30Z","lastTransitionTime":"2026-02-18T09:05:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:31 crc kubenswrapper[4556]: I0218 09:05:31.010247 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:31 crc kubenswrapper[4556]: I0218 09:05:31.010298 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:31 crc kubenswrapper[4556]: I0218 09:05:31.010306 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:31 crc kubenswrapper[4556]: I0218 09:05:31.010319 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:31 crc kubenswrapper[4556]: I0218 09:05:31.010336 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:31Z","lastTransitionTime":"2026-02-18T09:05:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:31 crc kubenswrapper[4556]: I0218 09:05:31.112269 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:31 crc kubenswrapper[4556]: I0218 09:05:31.112306 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:31 crc kubenswrapper[4556]: I0218 09:05:31.112315 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:31 crc kubenswrapper[4556]: I0218 09:05:31.112336 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:31 crc kubenswrapper[4556]: I0218 09:05:31.112344 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:31Z","lastTransitionTime":"2026-02-18T09:05:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:31 crc kubenswrapper[4556]: I0218 09:05:31.214454 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:31 crc kubenswrapper[4556]: I0218 09:05:31.214492 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:31 crc kubenswrapper[4556]: I0218 09:05:31.214502 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:31 crc kubenswrapper[4556]: I0218 09:05:31.214517 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:31 crc kubenswrapper[4556]: I0218 09:05:31.214525 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:31Z","lastTransitionTime":"2026-02-18T09:05:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:31 crc kubenswrapper[4556]: I0218 09:05:31.282237 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:05:31 crc kubenswrapper[4556]: I0218 09:05:31.282237 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:05:31 crc kubenswrapper[4556]: I0218 09:05:31.282320 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:05:31 crc kubenswrapper[4556]: E0218 09:05:31.282454 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 09:05:31 crc kubenswrapper[4556]: E0218 09:05:31.282564 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 09:05:31 crc kubenswrapper[4556]: E0218 09:05:31.282606 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 09:05:31 crc kubenswrapper[4556]: I0218 09:05:31.295887 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-09 23:44:20.480348693 +0000 UTC Feb 18 09:05:31 crc kubenswrapper[4556]: I0218 09:05:31.316113 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:31 crc kubenswrapper[4556]: I0218 09:05:31.316145 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:31 crc kubenswrapper[4556]: I0218 09:05:31.316175 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:31 crc kubenswrapper[4556]: I0218 09:05:31.316189 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:31 crc kubenswrapper[4556]: I0218 09:05:31.316197 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:31Z","lastTransitionTime":"2026-02-18T09:05:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:31 crc kubenswrapper[4556]: I0218 09:05:31.417961 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:31 crc kubenswrapper[4556]: I0218 09:05:31.417994 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:31 crc kubenswrapper[4556]: I0218 09:05:31.418003 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:31 crc kubenswrapper[4556]: I0218 09:05:31.418016 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:31 crc kubenswrapper[4556]: I0218 09:05:31.418025 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:31Z","lastTransitionTime":"2026-02-18T09:05:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:31 crc kubenswrapper[4556]: I0218 09:05:31.519952 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:31 crc kubenswrapper[4556]: I0218 09:05:31.519983 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:31 crc kubenswrapper[4556]: I0218 09:05:31.519994 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:31 crc kubenswrapper[4556]: I0218 09:05:31.520005 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:31 crc kubenswrapper[4556]: I0218 09:05:31.520014 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:31Z","lastTransitionTime":"2026-02-18T09:05:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:31 crc kubenswrapper[4556]: I0218 09:05:31.621004 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:31 crc kubenswrapper[4556]: I0218 09:05:31.621032 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:31 crc kubenswrapper[4556]: I0218 09:05:31.621041 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:31 crc kubenswrapper[4556]: I0218 09:05:31.621051 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:31 crc kubenswrapper[4556]: I0218 09:05:31.621065 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:31Z","lastTransitionTime":"2026-02-18T09:05:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:31 crc kubenswrapper[4556]: I0218 09:05:31.722498 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:31 crc kubenswrapper[4556]: I0218 09:05:31.722524 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:31 crc kubenswrapper[4556]: I0218 09:05:31.722553 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:31 crc kubenswrapper[4556]: I0218 09:05:31.722564 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:31 crc kubenswrapper[4556]: I0218 09:05:31.722572 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:31Z","lastTransitionTime":"2026-02-18T09:05:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:31 crc kubenswrapper[4556]: I0218 09:05:31.823685 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:31 crc kubenswrapper[4556]: I0218 09:05:31.823713 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:31 crc kubenswrapper[4556]: I0218 09:05:31.823721 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:31 crc kubenswrapper[4556]: I0218 09:05:31.823733 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:31 crc kubenswrapper[4556]: I0218 09:05:31.823741 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:31Z","lastTransitionTime":"2026-02-18T09:05:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:31 crc kubenswrapper[4556]: I0218 09:05:31.925442 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:31 crc kubenswrapper[4556]: I0218 09:05:31.925469 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:31 crc kubenswrapper[4556]: I0218 09:05:31.925477 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:31 crc kubenswrapper[4556]: I0218 09:05:31.925487 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:31 crc kubenswrapper[4556]: I0218 09:05:31.925494 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:31Z","lastTransitionTime":"2026-02-18T09:05:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:32 crc kubenswrapper[4556]: I0218 09:05:32.027382 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:32 crc kubenswrapper[4556]: I0218 09:05:32.027431 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:32 crc kubenswrapper[4556]: I0218 09:05:32.027441 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:32 crc kubenswrapper[4556]: I0218 09:05:32.027451 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:32 crc kubenswrapper[4556]: I0218 09:05:32.027461 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:32Z","lastTransitionTime":"2026-02-18T09:05:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:32 crc kubenswrapper[4556]: I0218 09:05:32.129454 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:32 crc kubenswrapper[4556]: I0218 09:05:32.129489 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:32 crc kubenswrapper[4556]: I0218 09:05:32.129499 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:32 crc kubenswrapper[4556]: I0218 09:05:32.129511 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:32 crc kubenswrapper[4556]: I0218 09:05:32.129521 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:32Z","lastTransitionTime":"2026-02-18T09:05:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:32 crc kubenswrapper[4556]: I0218 09:05:32.231302 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:32 crc kubenswrapper[4556]: I0218 09:05:32.231338 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:32 crc kubenswrapper[4556]: I0218 09:05:32.231347 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:32 crc kubenswrapper[4556]: I0218 09:05:32.231357 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:32 crc kubenswrapper[4556]: I0218 09:05:32.231366 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:32Z","lastTransitionTime":"2026-02-18T09:05:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:32 crc kubenswrapper[4556]: I0218 09:05:32.282169 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cgwv9" Feb 18 09:05:32 crc kubenswrapper[4556]: E0218 09:05:32.282496 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cgwv9" podUID="d2f0e36c-cb29-4b06-bb22-6afd59466cab" Feb 18 09:05:32 crc kubenswrapper[4556]: I0218 09:05:32.282586 4556 scope.go:117] "RemoveContainer" containerID="c2f16fe4bf2d659398ec141ba12d2dda32727810b03b96bfaf491602656027cb" Feb 18 09:05:32 crc kubenswrapper[4556]: E0218 09:05:32.282710 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-b7fsj_openshift-ovn-kubernetes(09971522-31bc-4e10-8831-4fb62675e220)\"" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" podUID="09971522-31bc-4e10-8831-4fb62675e220" Feb 18 09:05:32 crc kubenswrapper[4556]: I0218 09:05:32.296302 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-19 19:35:30.630028631 +0000 UTC Feb 18 09:05:32 crc kubenswrapper[4556]: I0218 09:05:32.332868 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:32 crc kubenswrapper[4556]: I0218 09:05:32.332892 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:32 crc kubenswrapper[4556]: I0218 09:05:32.332900 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:32 crc kubenswrapper[4556]: I0218 09:05:32.332909 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:32 crc kubenswrapper[4556]: I0218 09:05:32.332917 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:32Z","lastTransitionTime":"2026-02-18T09:05:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:32 crc kubenswrapper[4556]: I0218 09:05:32.435061 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:32 crc kubenswrapper[4556]: I0218 09:05:32.435085 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:32 crc kubenswrapper[4556]: I0218 09:05:32.435144 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:32 crc kubenswrapper[4556]: I0218 09:05:32.435175 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:32 crc kubenswrapper[4556]: I0218 09:05:32.435186 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:32Z","lastTransitionTime":"2026-02-18T09:05:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:32 crc kubenswrapper[4556]: I0218 09:05:32.536292 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:32 crc kubenswrapper[4556]: I0218 09:05:32.536327 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:32 crc kubenswrapper[4556]: I0218 09:05:32.536336 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:32 crc kubenswrapper[4556]: I0218 09:05:32.536346 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:32 crc kubenswrapper[4556]: I0218 09:05:32.536352 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:32Z","lastTransitionTime":"2026-02-18T09:05:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:32 crc kubenswrapper[4556]: I0218 09:05:32.639797 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:32 crc kubenswrapper[4556]: I0218 09:05:32.639823 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:32 crc kubenswrapper[4556]: I0218 09:05:32.639833 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:32 crc kubenswrapper[4556]: I0218 09:05:32.639844 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:32 crc kubenswrapper[4556]: I0218 09:05:32.639853 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:32Z","lastTransitionTime":"2026-02-18T09:05:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:32 crc kubenswrapper[4556]: I0218 09:05:32.742012 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:32 crc kubenswrapper[4556]: I0218 09:05:32.742041 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:32 crc kubenswrapper[4556]: I0218 09:05:32.742049 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:32 crc kubenswrapper[4556]: I0218 09:05:32.742058 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:32 crc kubenswrapper[4556]: I0218 09:05:32.742066 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:32Z","lastTransitionTime":"2026-02-18T09:05:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:32 crc kubenswrapper[4556]: I0218 09:05:32.843332 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:32 crc kubenswrapper[4556]: I0218 09:05:32.843385 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:32 crc kubenswrapper[4556]: I0218 09:05:32.843398 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:32 crc kubenswrapper[4556]: I0218 09:05:32.843416 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:32 crc kubenswrapper[4556]: I0218 09:05:32.843426 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:32Z","lastTransitionTime":"2026-02-18T09:05:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:32 crc kubenswrapper[4556]: I0218 09:05:32.945054 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:32 crc kubenswrapper[4556]: I0218 09:05:32.945084 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:32 crc kubenswrapper[4556]: I0218 09:05:32.945094 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:32 crc kubenswrapper[4556]: I0218 09:05:32.945105 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:32 crc kubenswrapper[4556]: I0218 09:05:32.945112 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:32Z","lastTransitionTime":"2026-02-18T09:05:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.046717 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.046737 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.046744 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.046755 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.046764 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:33Z","lastTransitionTime":"2026-02-18T09:05:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.148071 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.148095 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.148103 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.148112 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.148119 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:33Z","lastTransitionTime":"2026-02-18T09:05:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.249350 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.249374 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.249383 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.249392 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.249400 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:33Z","lastTransitionTime":"2026-02-18T09:05:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.281908 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.281999 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.281962 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:05:33 crc kubenswrapper[4556]: E0218 09:05:33.282191 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 09:05:33 crc kubenswrapper[4556]: E0218 09:05:33.282201 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 09:05:33 crc kubenswrapper[4556]: E0218 09:05:33.282241 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.292959 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:33Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.297183 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-18 10:31:50.258058634 +0000 UTC Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.299883 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bp8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"804f8152-54ae-4486-92e6-f9595c035820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb617e46b8522d73dabef4bbc660e41b737f6e6e62f7ce42dad722490a1041dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phr7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://104b8ff3c41d4304d414e58303c23792a7f0ba580f73a2a8678044741af71380\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phr7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bp8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:33Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.312753 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ae60f3-adab-4740-821a-1570059c7101\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f629262e27020512301a7b53ee836ee98e1e7cdee7a52b0ed506bbaaebd823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f59306c0d55e9d00714ce2ee2c1ed6f2ed33d8b21eecde490e86a93335a14f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c87bc671742a47de6c8e68aaf0c8645f9a8cae7b3d52264f36ae3827b3a8d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2678001ff826712da6716541210f018c74fd3d0218b388d2e7a6f16a1f4027\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6f37933281c927328ecfaf842d97ee8361726768f393726cb39517d2bdc9425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6a4eb8518698fa20f33ed04a7ede2aa3ebbf3c597f096a48a4044662b1b530d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7437ad40b24f9766d0446bfb24dcea0f6e9e6c4a8e1dd68de8f3894da1585de0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78a10dcc6e199f6e16c17c45ef56161c50b84474966822071721dec7f7440b25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:33Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.320846 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee8f6014-2544-4173-a94d-441cf42ba891\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77b0269625b7c6baab44a40373618f4ccfe9e1a028a1bc746b7d60c20ac4efaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac09c421b3101887bfb83107dbe26f2b5856d9b32a9cea92f6233fb6ae218cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce79cfe0242ecf403f989dc6b1c1b8be5ab54eee75ac413a8f9d03cef8bca0ed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://095ecf7e8763fa098b06eb9a8aac9f39451f7762bf20196e0803bbc841610e02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:33Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.328257 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:33Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.336709 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d653c21d3d7c875bed37ed6da700dc8c801b95c80d7e361121dfbbd3d05a922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5211d0cd5172d45b92b003e2bf97d7969a65ea448b0491f8f6ea67364925ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:33Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.345020 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6387f58543cf2601a53b0bbe9be3ec7816ba90d5ffdc4465594b37ccac0f6dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:33Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.351240 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.351264 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.351273 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.351282 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.351289 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:33Z","lastTransitionTime":"2026-02-18T09:05:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.351429 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cgwv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2f0e36c-cb29-4b06-bb22-6afd59466cab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpldv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpldv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:40Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cgwv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:33Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.359855 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"302eb215-a4c2-4324-836d-f37debfd620f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a99dc228d1c4cb8aeab2140ec5f37358b65684739f833fbde45822782430a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14f794063db9d33b82b79acc7134b5a2a3435e09d7cffe49d49251d1f7d20402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5eb3d3728da247903bece56bc2481547317e386e9e97b6c69926a3b55dc89b8e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b42d13c1767c29d22e52537ff45561a0fbf8f2069b2c3612e903e8c127e12485\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e865e6a50c03519ce22958ac4fcbb911fc140c330b6d7ffe5c9d7dd381e96\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T09:04:20Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0218 09:04:15.143085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 09:04:15.144650 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2160144045/tls.crt::/tmp/serving-cert-2160144045/tls.key\\\\\\\"\\\\nI0218 09:04:20.680799 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 09:04:20.682555 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 09:04:20.682573 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 09:04:20.682590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 09:04:20.682594 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 09:04:20.686822 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 09:04:20.686847 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686852 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 09:04:20.686856 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 09:04:20.686859 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 09:04:20.686861 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 09:04:20.686864 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 09:04:20.688257 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 09:04:20.690726 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://974e17ae7082631df2d1bcee60c8a5f142bae2659e9d200d537a4d650f401b4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:33Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.367693 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:33Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.376209 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qfksk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e226b189-cfa9-47fc-b94d-19b5cbe0859f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23f488cd416c6d49acf193e27beba9e9f1dcfb36430f9ae540742c2e2bd0f144\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc2cb92a7735d16b8e5918797e52e1e488d9261a0418ed5ebbb94c033ca9328a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T09:05:12Z\\\",\\\"message\\\":\\\"2026-02-18T09:04:27+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_3fa1df1f-8463-485c-a4c9-cd0c6dd4d227\\\\n2026-02-18T09:04:27+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_3fa1df1f-8463-485c-a4c9-cd0c6dd4d227 to /host/opt/cni/bin/\\\\n2026-02-18T09:04:27Z [verbose] multus-daemon started\\\\n2026-02-18T09:04:27Z [verbose] Readiness Indicator file check\\\\n2026-02-18T09:05:12Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:05:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2fx4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qfksk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:33Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.384184 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dac7f27-d3d1-4778-9e54-f273035a1d37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18da54fe071964d8aae8a977f5ee209c31f1c75e1041c44f8d755f58f3e39d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d41b85076e695bbc51a41fce200e0e47c099d88b4d825b4d133d33ed96237e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtvs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f76hs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:33Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.392773 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sgcq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9945b792-abbd-4103-a7b4-9496e1cc1b56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5dd402827182bc79283f1ea1c355b60c4ced53ffa9ce6781b0a0db3109359a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2de1113b5fd2fc38bcfa50c394664b02e8bb71378acfe1d888ab13a872dc2b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c1b0f76e042ec10b6dff0ef8f49694cacff6eea19a6da55a789b07942ac3cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51b60658dba3be5911c432499a93d923248720eb7cb8e3aafc55674b6821e93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac0d9d7b9bc867b32cf6cbeb804c08e67df757e4b22914e2da69a376511cf75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac0d9d7b9bc867b32cf6cbeb804c08e67df757e4b22914e2da69a376511cf75c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b807d53d620e44d3196f68492b9f9665593adac804bc7b91569796e5594c4ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b807d53d620e44d3196f68492b9f9665593adac804bc7b91569796e5594c4ebb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebc044bbd5b0c1e725e6ed46ec2e0015b254070b7518d20815539dcf8c6e4e45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ebc044bbd5b0c1e725e6ed46ec2e0015b254070b7518d20815539dcf8c6e4e45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx8j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sgcq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:33Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.402056 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4e7c95a-56e9-4e22-83dd-54962e1233b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52ea6a334e8cad154e46eb2f917cdac914557556bae88f922433a9545cb29eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28a600e8daaa2796b2f0cc6ce4112f8d41c611b5b6ca5a7eb0f260dab615caa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1fa733477ee6330835d52333d6e9588626b6fc7ae08ae450dd2e1821447e995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfbda9799e2d2bc16964e79b90fb7087f9e6893d867711612c63aaaa3dd2b141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfbda9799e2d2bc16964e79b90fb7087f9e6893d867711612c63aaaa3dd2b141\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:33Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.408557 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64e03bf0-d2fc-40ba-a468-4d9a2b43afc7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ac8fc80676085932e05da4ff18e892eebaec46d096c7c343b667864f6c5e3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6e0422278544bddfec087d7b0a94bedc55a7ac67ff280a36030528eb87fd72b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6e0422278544bddfec087d7b0a94bedc55a7ac67ff280a36030528eb87fd72b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:33Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.416517 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47341f31f94eaf5dd86efbe78ff62873e733b1f86ea895f281b6797770519459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:33Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.422549 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5vcgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d50c39aa-f956-441a-8abe-1d7247a7fd86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ddc5d42163d1fa86f1a59e224adac434c990224e839ef3cb12c6d7689e2b439\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4llw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5vcgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:33Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.434079 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09971522-31bc-4e10-8831-4fb62675e220\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9170bdcbaf1065995d03e6448e40bdced0741c716c9e81f37a72b015132f1161\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c526c0613f19e1a1a584b7129bd56001faf08d0b97fd77feeb69f764ed01f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cd6e758ee396c9000b1bdb6ded2176ec803fa9da6201cdd9b8146940a2c00cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b0412f3330999c3348ac726ac5a0c2d541571f4b13786ca24cb3109a44aa653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://318cb69e50df4c14df9bf5ef57d5c381421a9bdb72135c9b9620f55c0547b410\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0f15d331af9bc2ca3082c1573dd167c66773eb390593844e681f8d0b455d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2f16fe4bf2d659398ec141ba12d2dda32727810b03b96bfaf491602656027cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2f16fe4bf2d659398ec141ba12d2dda32727810b03b96bfaf491602656027cb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T09:05:16Z\\\",\\\"message\\\":\\\"d to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:16Z is after 2025-08-24T17:21:41Z]\\\\nI0218 09:05:16.892059 6632 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/package-server-manager-metrics]} name:Service_openshift-operator-lifecycle-manager/package-server-manager-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.110:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {f9232b32-e89f-4c8e-acc4-c6801b70dcb0}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:NB_Global Row:map[] Rows:[] Columns:[] Mutations:[{Column:nb_cfg Mutator:+= Value:1}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {6011affd-30a6-4be6-872d-e4cf1ca780cf}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0218 09:05:16.891870 6632 handler.go:208] Rem\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T09:05:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-b7fsj_openshift-ovn-kubernetes(09971522-31bc-4e10-8831-4fb62675e220)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aec6030df0a59ad8902d5c694d5cd30d249ca1e895155c57a458e359fa531b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T09:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcd67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7fsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:33Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.440352 4556 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jflp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689d3d82-4071-4fe5-b3e3-b5b4f511580a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T09:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d56d65964aaf12feaaa3e5fa3ca779bffdb13fcebfc9f027b563c7dd01f4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T09:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bm85s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T09:04:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jflp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T09:05:33Z is after 2025-08-24T17:21:41Z" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.452637 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.452668 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.452677 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.452688 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.452696 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:33Z","lastTransitionTime":"2026-02-18T09:05:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.554411 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.554431 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.554439 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.554449 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.554456 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:33Z","lastTransitionTime":"2026-02-18T09:05:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.655754 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.655777 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.655784 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.655795 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.655803 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:33Z","lastTransitionTime":"2026-02-18T09:05:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.757360 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.757383 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.757392 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.757404 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.757415 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:33Z","lastTransitionTime":"2026-02-18T09:05:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.859424 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.859453 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.859461 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.859475 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.859484 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:33Z","lastTransitionTime":"2026-02-18T09:05:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.960956 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.960984 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.960992 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.961003 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:33 crc kubenswrapper[4556]: I0218 09:05:33.961011 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:33Z","lastTransitionTime":"2026-02-18T09:05:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:34 crc kubenswrapper[4556]: I0218 09:05:34.062860 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:34 crc kubenswrapper[4556]: I0218 09:05:34.062903 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:34 crc kubenswrapper[4556]: I0218 09:05:34.062911 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:34 crc kubenswrapper[4556]: I0218 09:05:34.062921 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:34 crc kubenswrapper[4556]: I0218 09:05:34.062927 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:34Z","lastTransitionTime":"2026-02-18T09:05:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:34 crc kubenswrapper[4556]: I0218 09:05:34.164332 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:34 crc kubenswrapper[4556]: I0218 09:05:34.164364 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:34 crc kubenswrapper[4556]: I0218 09:05:34.164373 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:34 crc kubenswrapper[4556]: I0218 09:05:34.164387 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:34 crc kubenswrapper[4556]: I0218 09:05:34.164396 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:34Z","lastTransitionTime":"2026-02-18T09:05:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:34 crc kubenswrapper[4556]: I0218 09:05:34.265809 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:34 crc kubenswrapper[4556]: I0218 09:05:34.265838 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:34 crc kubenswrapper[4556]: I0218 09:05:34.265848 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:34 crc kubenswrapper[4556]: I0218 09:05:34.265859 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:34 crc kubenswrapper[4556]: I0218 09:05:34.265869 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:34Z","lastTransitionTime":"2026-02-18T09:05:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:34 crc kubenswrapper[4556]: I0218 09:05:34.281860 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cgwv9" Feb 18 09:05:34 crc kubenswrapper[4556]: E0218 09:05:34.281942 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cgwv9" podUID="d2f0e36c-cb29-4b06-bb22-6afd59466cab" Feb 18 09:05:34 crc kubenswrapper[4556]: I0218 09:05:34.298199 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 15:21:47.421885421 +0000 UTC Feb 18 09:05:34 crc kubenswrapper[4556]: I0218 09:05:34.368176 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:34 crc kubenswrapper[4556]: I0218 09:05:34.368208 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:34 crc kubenswrapper[4556]: I0218 09:05:34.368218 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:34 crc kubenswrapper[4556]: I0218 09:05:34.368231 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:34 crc kubenswrapper[4556]: I0218 09:05:34.368239 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:34Z","lastTransitionTime":"2026-02-18T09:05:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:34 crc kubenswrapper[4556]: I0218 09:05:34.470013 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:34 crc kubenswrapper[4556]: I0218 09:05:34.470038 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:34 crc kubenswrapper[4556]: I0218 09:05:34.470063 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:34 crc kubenswrapper[4556]: I0218 09:05:34.470072 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:34 crc kubenswrapper[4556]: I0218 09:05:34.470079 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:34Z","lastTransitionTime":"2026-02-18T09:05:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:34 crc kubenswrapper[4556]: I0218 09:05:34.571580 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:34 crc kubenswrapper[4556]: I0218 09:05:34.571607 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:34 crc kubenswrapper[4556]: I0218 09:05:34.571615 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:34 crc kubenswrapper[4556]: I0218 09:05:34.571625 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:34 crc kubenswrapper[4556]: I0218 09:05:34.571631 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:34Z","lastTransitionTime":"2026-02-18T09:05:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:34 crc kubenswrapper[4556]: I0218 09:05:34.673059 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:34 crc kubenswrapper[4556]: I0218 09:05:34.673104 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:34 crc kubenswrapper[4556]: I0218 09:05:34.673113 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:34 crc kubenswrapper[4556]: I0218 09:05:34.673123 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:34 crc kubenswrapper[4556]: I0218 09:05:34.673130 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:34Z","lastTransitionTime":"2026-02-18T09:05:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:34 crc kubenswrapper[4556]: I0218 09:05:34.775009 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:34 crc kubenswrapper[4556]: I0218 09:05:34.775046 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:34 crc kubenswrapper[4556]: I0218 09:05:34.775057 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:34 crc kubenswrapper[4556]: I0218 09:05:34.775070 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:34 crc kubenswrapper[4556]: I0218 09:05:34.775079 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:34Z","lastTransitionTime":"2026-02-18T09:05:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:34 crc kubenswrapper[4556]: I0218 09:05:34.877772 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:34 crc kubenswrapper[4556]: I0218 09:05:34.877810 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:34 crc kubenswrapper[4556]: I0218 09:05:34.877819 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:34 crc kubenswrapper[4556]: I0218 09:05:34.877832 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:34 crc kubenswrapper[4556]: I0218 09:05:34.877841 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:34Z","lastTransitionTime":"2026-02-18T09:05:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:34 crc kubenswrapper[4556]: I0218 09:05:34.979331 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:34 crc kubenswrapper[4556]: I0218 09:05:34.979361 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:34 crc kubenswrapper[4556]: I0218 09:05:34.979374 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:34 crc kubenswrapper[4556]: I0218 09:05:34.979386 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:34 crc kubenswrapper[4556]: I0218 09:05:34.979394 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:34Z","lastTransitionTime":"2026-02-18T09:05:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.080763 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.080793 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.080801 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.080811 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.080818 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:35Z","lastTransitionTime":"2026-02-18T09:05:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.182179 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.182209 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.182219 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.182230 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.182239 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:35Z","lastTransitionTime":"2026-02-18T09:05:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.281895 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:05:35 crc kubenswrapper[4556]: E0218 09:05:35.282091 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.282124 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.282177 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:05:35 crc kubenswrapper[4556]: E0218 09:05:35.282316 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 09:05:35 crc kubenswrapper[4556]: E0218 09:05:35.282636 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.283402 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.283427 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.283435 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.283446 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.283453 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:35Z","lastTransitionTime":"2026-02-18T09:05:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.298703 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-29 20:20:40.867149631 +0000 UTC Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.385447 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.385470 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.385478 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.385488 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.385495 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:35Z","lastTransitionTime":"2026-02-18T09:05:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.486732 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.486753 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.486766 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.486775 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.486782 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:35Z","lastTransitionTime":"2026-02-18T09:05:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.588289 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.588322 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.588330 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.588342 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.588352 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:35Z","lastTransitionTime":"2026-02-18T09:05:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.635566 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.635593 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.635601 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.635611 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.635618 4556 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T09:05:35Z","lastTransitionTime":"2026-02-18T09:05:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.663079 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-wp6jv"] Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.663402 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wp6jv" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.664656 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.664710 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.664776 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.665489 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.685486 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=74.685472226 podStartE2EDuration="1m14.685472226s" podCreationTimestamp="2026-02-18 09:04:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:05:35.676219508 +0000 UTC m=+92.693180488" watchObservedRunningTime="2026-02-18 09:05:35.685472226 +0000 UTC m=+92.702433206" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.701229 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podStartSLOduration=69.701216711 podStartE2EDuration="1m9.701216711s" podCreationTimestamp="2026-02-18 09:04:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:05:35.701131752 +0000 UTC m=+92.718092742" watchObservedRunningTime="2026-02-18 09:05:35.701216711 +0000 UTC m=+92.718177701" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.701509 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-qfksk" podStartSLOduration=69.701501757 podStartE2EDuration="1m9.701501757s" podCreationTimestamp="2026-02-18 09:04:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:05:35.694258075 +0000 UTC m=+92.711219055" watchObservedRunningTime="2026-02-18 09:05:35.701501757 +0000 UTC m=+92.718462737" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.711976 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-sgcq7" podStartSLOduration=69.711959819 podStartE2EDuration="1m9.711959819s" podCreationTimestamp="2026-02-18 09:04:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:05:35.711738553 +0000 UTC m=+92.728699533" watchObservedRunningTime="2026-02-18 09:05:35.711959819 +0000 UTC m=+92.728920879" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.725374 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=45.725360922 podStartE2EDuration="45.725360922s" podCreationTimestamp="2026-02-18 09:04:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:05:35.725013929 +0000 UTC m=+92.741974909" watchObservedRunningTime="2026-02-18 09:05:35.725360922 +0000 UTC m=+92.742321902" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.731618 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=19.731605355 podStartE2EDuration="19.731605355s" podCreationTimestamp="2026-02-18 09:05:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:05:35.731068056 +0000 UTC m=+92.748029035" watchObservedRunningTime="2026-02-18 09:05:35.731605355 +0000 UTC m=+92.748566336" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.748955 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-5vcgz" podStartSLOduration=69.748942314 podStartE2EDuration="1m9.748942314s" podCreationTimestamp="2026-02-18 09:04:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:05:35.748355181 +0000 UTC m=+92.765316160" watchObservedRunningTime="2026-02-18 09:05:35.748942314 +0000 UTC m=+92.765903294" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.750677 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/b62637f5-d5da-4c76-8f05-db6d9178c5e5-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-wp6jv\" (UID: \"b62637f5-d5da-4c76-8f05-db6d9178c5e5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wp6jv" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.750731 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/b62637f5-d5da-4c76-8f05-db6d9178c5e5-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-wp6jv\" (UID: \"b62637f5-d5da-4c76-8f05-db6d9178c5e5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wp6jv" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.750761 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b62637f5-d5da-4c76-8f05-db6d9178c5e5-service-ca\") pod \"cluster-version-operator-5c965bbfc6-wp6jv\" (UID: \"b62637f5-d5da-4c76-8f05-db6d9178c5e5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wp6jv" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.750821 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b62637f5-d5da-4c76-8f05-db6d9178c5e5-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-wp6jv\" (UID: \"b62637f5-d5da-4c76-8f05-db6d9178c5e5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wp6jv" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.750886 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b62637f5-d5da-4c76-8f05-db6d9178c5e5-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-wp6jv\" (UID: \"b62637f5-d5da-4c76-8f05-db6d9178c5e5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wp6jv" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.771283 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-2jflp" podStartSLOduration=69.771271503 podStartE2EDuration="1m9.771271503s" podCreationTimestamp="2026-02-18 09:04:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:05:35.770988692 +0000 UTC m=+92.787949672" watchObservedRunningTime="2026-02-18 09:05:35.771271503 +0000 UTC m=+92.788232483" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.787368 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bp8gg" podStartSLOduration=69.787352472 podStartE2EDuration="1m9.787352472s" podCreationTimestamp="2026-02-18 09:04:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:05:35.786066395 +0000 UTC m=+92.803027375" watchObservedRunningTime="2026-02-18 09:05:35.787352472 +0000 UTC m=+92.804313453" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.803478 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=71.803465751 podStartE2EDuration="1m11.803465751s" podCreationTimestamp="2026-02-18 09:04:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:05:35.802653575 +0000 UTC m=+92.819614554" watchObservedRunningTime="2026-02-18 09:05:35.803465751 +0000 UTC m=+92.820426731" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.812790 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=73.81277825 podStartE2EDuration="1m13.81277825s" podCreationTimestamp="2026-02-18 09:04:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:05:35.812005829 +0000 UTC m=+92.828966819" watchObservedRunningTime="2026-02-18 09:05:35.81277825 +0000 UTC m=+92.829739231" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.851787 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b62637f5-d5da-4c76-8f05-db6d9178c5e5-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-wp6jv\" (UID: \"b62637f5-d5da-4c76-8f05-db6d9178c5e5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wp6jv" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.851818 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b62637f5-d5da-4c76-8f05-db6d9178c5e5-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-wp6jv\" (UID: \"b62637f5-d5da-4c76-8f05-db6d9178c5e5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wp6jv" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.851851 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/b62637f5-d5da-4c76-8f05-db6d9178c5e5-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-wp6jv\" (UID: \"b62637f5-d5da-4c76-8f05-db6d9178c5e5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wp6jv" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.851881 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/b62637f5-d5da-4c76-8f05-db6d9178c5e5-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-wp6jv\" (UID: \"b62637f5-d5da-4c76-8f05-db6d9178c5e5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wp6jv" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.851907 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b62637f5-d5da-4c76-8f05-db6d9178c5e5-service-ca\") pod \"cluster-version-operator-5c965bbfc6-wp6jv\" (UID: \"b62637f5-d5da-4c76-8f05-db6d9178c5e5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wp6jv" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.851972 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/b62637f5-d5da-4c76-8f05-db6d9178c5e5-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-wp6jv\" (UID: \"b62637f5-d5da-4c76-8f05-db6d9178c5e5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wp6jv" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.852000 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/b62637f5-d5da-4c76-8f05-db6d9178c5e5-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-wp6jv\" (UID: \"b62637f5-d5da-4c76-8f05-db6d9178c5e5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wp6jv" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.852589 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b62637f5-d5da-4c76-8f05-db6d9178c5e5-service-ca\") pod \"cluster-version-operator-5c965bbfc6-wp6jv\" (UID: \"b62637f5-d5da-4c76-8f05-db6d9178c5e5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wp6jv" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.855806 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b62637f5-d5da-4c76-8f05-db6d9178c5e5-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-wp6jv\" (UID: \"b62637f5-d5da-4c76-8f05-db6d9178c5e5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wp6jv" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.864088 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b62637f5-d5da-4c76-8f05-db6d9178c5e5-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-wp6jv\" (UID: \"b62637f5-d5da-4c76-8f05-db6d9178c5e5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wp6jv" Feb 18 09:05:35 crc kubenswrapper[4556]: I0218 09:05:35.973719 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wp6jv" Feb 18 09:05:36 crc kubenswrapper[4556]: I0218 09:05:36.281684 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cgwv9" Feb 18 09:05:36 crc kubenswrapper[4556]: E0218 09:05:36.281956 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cgwv9" podUID="d2f0e36c-cb29-4b06-bb22-6afd59466cab" Feb 18 09:05:36 crc kubenswrapper[4556]: I0218 09:05:36.299197 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-27 09:55:14.308923999 +0000 UTC Feb 18 09:05:36 crc kubenswrapper[4556]: I0218 09:05:36.299253 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Feb 18 09:05:36 crc kubenswrapper[4556]: I0218 09:05:36.304124 4556 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Feb 18 09:05:36 crc kubenswrapper[4556]: I0218 09:05:36.636962 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wp6jv" event={"ID":"b62637f5-d5da-4c76-8f05-db6d9178c5e5","Type":"ContainerStarted","Data":"236d96cbfccad6fc98d05d77a12bde25346009d50b8229cbfeb8db121f2822ed"} Feb 18 09:05:36 crc kubenswrapper[4556]: I0218 09:05:36.637000 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wp6jv" event={"ID":"b62637f5-d5da-4c76-8f05-db6d9178c5e5","Type":"ContainerStarted","Data":"3ccbe93366b571c6d7fd6b3cb5487c5c67b83a1654d4afdda7cbc9d70ce80f3d"} Feb 18 09:05:36 crc kubenswrapper[4556]: I0218 09:05:36.646455 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wp6jv" podStartSLOduration=70.646441584 podStartE2EDuration="1m10.646441584s" podCreationTimestamp="2026-02-18 09:04:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:05:36.646269481 +0000 UTC m=+93.663230451" watchObservedRunningTime="2026-02-18 09:05:36.646441584 +0000 UTC m=+93.663402574" Feb 18 09:05:37 crc kubenswrapper[4556]: I0218 09:05:37.281553 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:05:37 crc kubenswrapper[4556]: E0218 09:05:37.281802 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 09:05:37 crc kubenswrapper[4556]: I0218 09:05:37.281583 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:05:37 crc kubenswrapper[4556]: E0218 09:05:37.281981 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 09:05:37 crc kubenswrapper[4556]: I0218 09:05:37.281558 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:05:37 crc kubenswrapper[4556]: E0218 09:05:37.282127 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 09:05:38 crc kubenswrapper[4556]: I0218 09:05:38.282390 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cgwv9" Feb 18 09:05:38 crc kubenswrapper[4556]: E0218 09:05:38.282989 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cgwv9" podUID="d2f0e36c-cb29-4b06-bb22-6afd59466cab" Feb 18 09:05:39 crc kubenswrapper[4556]: I0218 09:05:39.281780 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:05:39 crc kubenswrapper[4556]: I0218 09:05:39.281968 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:05:39 crc kubenswrapper[4556]: I0218 09:05:39.281824 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:05:39 crc kubenswrapper[4556]: E0218 09:05:39.282094 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 09:05:39 crc kubenswrapper[4556]: E0218 09:05:39.282195 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 09:05:39 crc kubenswrapper[4556]: E0218 09:05:39.282360 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 09:05:40 crc kubenswrapper[4556]: I0218 09:05:40.281765 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cgwv9" Feb 18 09:05:40 crc kubenswrapper[4556]: E0218 09:05:40.281856 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cgwv9" podUID="d2f0e36c-cb29-4b06-bb22-6afd59466cab" Feb 18 09:05:41 crc kubenswrapper[4556]: I0218 09:05:41.281754 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:05:41 crc kubenswrapper[4556]: I0218 09:05:41.281794 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:05:41 crc kubenswrapper[4556]: E0218 09:05:41.281830 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 09:05:41 crc kubenswrapper[4556]: E0218 09:05:41.281864 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 09:05:41 crc kubenswrapper[4556]: I0218 09:05:41.281928 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:05:41 crc kubenswrapper[4556]: E0218 09:05:41.282041 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 09:05:42 crc kubenswrapper[4556]: I0218 09:05:42.282534 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cgwv9" Feb 18 09:05:42 crc kubenswrapper[4556]: E0218 09:05:42.282625 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cgwv9" podUID="d2f0e36c-cb29-4b06-bb22-6afd59466cab" Feb 18 09:05:43 crc kubenswrapper[4556]: I0218 09:05:43.282133 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:05:43 crc kubenswrapper[4556]: E0218 09:05:43.282911 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 09:05:43 crc kubenswrapper[4556]: I0218 09:05:43.282944 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:05:43 crc kubenswrapper[4556]: I0218 09:05:43.282984 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:05:43 crc kubenswrapper[4556]: E0218 09:05:43.283043 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 09:05:43 crc kubenswrapper[4556]: E0218 09:05:43.283078 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 09:05:44 crc kubenswrapper[4556]: I0218 09:05:44.282058 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cgwv9" Feb 18 09:05:44 crc kubenswrapper[4556]: E0218 09:05:44.282190 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cgwv9" podUID="d2f0e36c-cb29-4b06-bb22-6afd59466cab" Feb 18 09:05:44 crc kubenswrapper[4556]: I0218 09:05:44.315972 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d2f0e36c-cb29-4b06-bb22-6afd59466cab-metrics-certs\") pod \"network-metrics-daemon-cgwv9\" (UID: \"d2f0e36c-cb29-4b06-bb22-6afd59466cab\") " pod="openshift-multus/network-metrics-daemon-cgwv9" Feb 18 09:05:44 crc kubenswrapper[4556]: E0218 09:05:44.316098 4556 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 18 09:05:44 crc kubenswrapper[4556]: E0218 09:05:44.316212 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d2f0e36c-cb29-4b06-bb22-6afd59466cab-metrics-certs podName:d2f0e36c-cb29-4b06-bb22-6afd59466cab nodeName:}" failed. No retries permitted until 2026-02-18 09:06:48.316192621 +0000 UTC m=+165.333153611 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d2f0e36c-cb29-4b06-bb22-6afd59466cab-metrics-certs") pod "network-metrics-daemon-cgwv9" (UID: "d2f0e36c-cb29-4b06-bb22-6afd59466cab") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 18 09:05:45 crc kubenswrapper[4556]: I0218 09:05:45.281859 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:05:45 crc kubenswrapper[4556]: E0218 09:05:45.282129 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 09:05:45 crc kubenswrapper[4556]: I0218 09:05:45.281886 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:05:45 crc kubenswrapper[4556]: I0218 09:05:45.281874 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:05:45 crc kubenswrapper[4556]: E0218 09:05:45.282615 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 09:05:45 crc kubenswrapper[4556]: E0218 09:05:45.282700 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 09:05:46 crc kubenswrapper[4556]: I0218 09:05:46.282291 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cgwv9" Feb 18 09:05:46 crc kubenswrapper[4556]: E0218 09:05:46.282385 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cgwv9" podUID="d2f0e36c-cb29-4b06-bb22-6afd59466cab" Feb 18 09:05:46 crc kubenswrapper[4556]: I0218 09:05:46.282773 4556 scope.go:117] "RemoveContainer" containerID="c2f16fe4bf2d659398ec141ba12d2dda32727810b03b96bfaf491602656027cb" Feb 18 09:05:46 crc kubenswrapper[4556]: E0218 09:05:46.282891 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-b7fsj_openshift-ovn-kubernetes(09971522-31bc-4e10-8831-4fb62675e220)\"" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" podUID="09971522-31bc-4e10-8831-4fb62675e220" Feb 18 09:05:47 crc kubenswrapper[4556]: I0218 09:05:47.282278 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:05:47 crc kubenswrapper[4556]: E0218 09:05:47.282399 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 09:05:47 crc kubenswrapper[4556]: I0218 09:05:47.282306 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:05:47 crc kubenswrapper[4556]: E0218 09:05:47.282482 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 09:05:47 crc kubenswrapper[4556]: I0218 09:05:47.282293 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:05:47 crc kubenswrapper[4556]: E0218 09:05:47.282552 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 09:05:48 crc kubenswrapper[4556]: I0218 09:05:48.281416 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cgwv9" Feb 18 09:05:48 crc kubenswrapper[4556]: E0218 09:05:48.281815 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cgwv9" podUID="d2f0e36c-cb29-4b06-bb22-6afd59466cab" Feb 18 09:05:49 crc kubenswrapper[4556]: I0218 09:05:49.281642 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:05:49 crc kubenswrapper[4556]: I0218 09:05:49.281690 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:05:49 crc kubenswrapper[4556]: E0218 09:05:49.281752 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 09:05:49 crc kubenswrapper[4556]: I0218 09:05:49.281772 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:05:49 crc kubenswrapper[4556]: E0218 09:05:49.281794 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 09:05:49 crc kubenswrapper[4556]: E0218 09:05:49.281833 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 09:05:50 crc kubenswrapper[4556]: I0218 09:05:50.282033 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cgwv9" Feb 18 09:05:50 crc kubenswrapper[4556]: E0218 09:05:50.282118 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cgwv9" podUID="d2f0e36c-cb29-4b06-bb22-6afd59466cab" Feb 18 09:05:51 crc kubenswrapper[4556]: I0218 09:05:51.282358 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:05:51 crc kubenswrapper[4556]: E0218 09:05:51.282439 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 09:05:51 crc kubenswrapper[4556]: I0218 09:05:51.282457 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:05:51 crc kubenswrapper[4556]: E0218 09:05:51.282525 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 09:05:51 crc kubenswrapper[4556]: I0218 09:05:51.282956 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:05:51 crc kubenswrapper[4556]: E0218 09:05:51.283089 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 09:05:52 crc kubenswrapper[4556]: I0218 09:05:52.281557 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cgwv9" Feb 18 09:05:52 crc kubenswrapper[4556]: E0218 09:05:52.281649 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cgwv9" podUID="d2f0e36c-cb29-4b06-bb22-6afd59466cab" Feb 18 09:05:53 crc kubenswrapper[4556]: I0218 09:05:53.282083 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:05:53 crc kubenswrapper[4556]: I0218 09:05:53.282268 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:05:53 crc kubenswrapper[4556]: E0218 09:05:53.283595 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 09:05:53 crc kubenswrapper[4556]: I0218 09:05:53.283618 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:05:53 crc kubenswrapper[4556]: E0218 09:05:53.283784 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 09:05:53 crc kubenswrapper[4556]: E0218 09:05:53.284018 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 09:05:54 crc kubenswrapper[4556]: I0218 09:05:54.281375 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cgwv9" Feb 18 09:05:54 crc kubenswrapper[4556]: E0218 09:05:54.281735 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cgwv9" podUID="d2f0e36c-cb29-4b06-bb22-6afd59466cab" Feb 18 09:05:55 crc kubenswrapper[4556]: I0218 09:05:55.281658 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:05:55 crc kubenswrapper[4556]: I0218 09:05:55.281667 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:05:55 crc kubenswrapper[4556]: I0218 09:05:55.281934 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:05:55 crc kubenswrapper[4556]: E0218 09:05:55.282558 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 09:05:55 crc kubenswrapper[4556]: E0218 09:05:55.282603 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 09:05:55 crc kubenswrapper[4556]: E0218 09:05:55.282317 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 09:05:56 crc kubenswrapper[4556]: I0218 09:05:56.282031 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cgwv9" Feb 18 09:05:56 crc kubenswrapper[4556]: E0218 09:05:56.282243 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cgwv9" podUID="d2f0e36c-cb29-4b06-bb22-6afd59466cab" Feb 18 09:05:57 crc kubenswrapper[4556]: I0218 09:05:57.282067 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:05:57 crc kubenswrapper[4556]: E0218 09:05:57.282217 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 09:05:57 crc kubenswrapper[4556]: I0218 09:05:57.282250 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:05:57 crc kubenswrapper[4556]: I0218 09:05:57.282307 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:05:57 crc kubenswrapper[4556]: E0218 09:05:57.282331 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 09:05:57 crc kubenswrapper[4556]: E0218 09:05:57.284458 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 09:05:57 crc kubenswrapper[4556]: I0218 09:05:57.288057 4556 scope.go:117] "RemoveContainer" containerID="c2f16fe4bf2d659398ec141ba12d2dda32727810b03b96bfaf491602656027cb" Feb 18 09:05:57 crc kubenswrapper[4556]: I0218 09:05:57.680825 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-b7fsj_09971522-31bc-4e10-8831-4fb62675e220/ovnkube-controller/3.log" Feb 18 09:05:57 crc kubenswrapper[4556]: I0218 09:05:57.682886 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" event={"ID":"09971522-31bc-4e10-8831-4fb62675e220","Type":"ContainerStarted","Data":"ad37b48f74a075b5562885db93abefd92df9f91f81a3b2ff685ee97a13b5e403"} Feb 18 09:05:57 crc kubenswrapper[4556]: I0218 09:05:57.683204 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:05:57 crc kubenswrapper[4556]: I0218 09:05:57.701667 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" podStartSLOduration=91.701655278 podStartE2EDuration="1m31.701655278s" podCreationTimestamp="2026-02-18 09:04:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:05:57.701105565 +0000 UTC m=+114.718066565" watchObservedRunningTime="2026-02-18 09:05:57.701655278 +0000 UTC m=+114.718616259" Feb 18 09:05:57 crc kubenswrapper[4556]: I0218 09:05:57.880769 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-cgwv9"] Feb 18 09:05:57 crc kubenswrapper[4556]: I0218 09:05:57.880875 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cgwv9" Feb 18 09:05:57 crc kubenswrapper[4556]: E0218 09:05:57.880955 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cgwv9" podUID="d2f0e36c-cb29-4b06-bb22-6afd59466cab" Feb 18 09:05:59 crc kubenswrapper[4556]: I0218 09:05:59.281826 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:05:59 crc kubenswrapper[4556]: I0218 09:05:59.281888 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:05:59 crc kubenswrapper[4556]: I0218 09:05:59.281948 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:05:59 crc kubenswrapper[4556]: E0218 09:05:59.282278 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 09:05:59 crc kubenswrapper[4556]: E0218 09:05:59.282137 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 09:05:59 crc kubenswrapper[4556]: I0218 09:05:59.281957 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cgwv9" Feb 18 09:05:59 crc kubenswrapper[4556]: E0218 09:05:59.282381 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 09:05:59 crc kubenswrapper[4556]: E0218 09:05:59.282462 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cgwv9" podUID="d2f0e36c-cb29-4b06-bb22-6afd59466cab" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.024367 4556 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.024481 4556 kubelet_node_status.go:538] "Fast updating node status as it just became ready" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.049909 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-26rzb"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.050229 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-lzjw5"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.050483 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-lzjw5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.050719 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-26rzb" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.050833 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-md5fm"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.051078 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-md5fm" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.052385 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-kxkx5"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.052647 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxkx5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.052876 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-j9ccg"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.053060 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-j9ccg" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.053721 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.053864 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.053923 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-5wlpw"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.054105 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-5wlpw" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.054348 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.054821 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4dfqf"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.055084 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4dfqf" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.055454 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n2cfq"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.055648 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n2cfq" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.056109 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r44z2"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.056403 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r44z2" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.056646 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mgwg5"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.056851 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mgwg5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.057293 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-55hcm"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.057518 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-55hcm" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.058114 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.058199 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-lsngn"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.058535 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lgh4n"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.058729 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lgh4n" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.058938 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-lsngn" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.059252 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-6d57b"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.059548 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-6d57b" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.060057 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-8gpk5"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.060320 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-8gpk5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.060687 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-sxb49"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.060979 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sxb49" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.061724 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-fmk6k"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.062123 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-fmk6k" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.064176 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.064194 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.064227 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.066585 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.066798 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.066828 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.067435 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.067602 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.067717 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.072499 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.072546 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.072499 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.072832 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.073195 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.073504 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.073502 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.074790 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.074953 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.075406 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.075490 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.076338 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.081185 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.081243 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.081254 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.081290 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.081396 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.081446 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.081470 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.081478 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.081511 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.081451 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.081484 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.081514 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.081458 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.081612 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.081620 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.081626 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.081649 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.081617 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.081690 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.081707 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.081739 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.081750 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.081773 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.081981 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.081997 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.082045 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.082060 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.082131 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.082189 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.082227 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.082281 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.082334 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.082368 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.082408 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.082440 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.082465 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.082471 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.082522 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.082527 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.082547 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.082597 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.082679 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.082808 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.082912 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.083205 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.083316 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.083394 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.083420 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.083474 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.083627 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.084493 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.084741 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-wz622"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.085116 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-lv2k7"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.085430 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-lv2k7" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.085660 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.085780 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-zhtqs"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.085970 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.086206 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.089367 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-dncx2"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.089686 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-dffhx"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.089920 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lkxrl"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.090130 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-vd7f7"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.090421 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-vd7f7" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.090623 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-zhtqs" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.090779 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dncx2" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.090945 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-dffhx" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.091107 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lkxrl" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.091300 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6ss98"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.091656 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-4vkc6"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.092035 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-4vkc6" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.092095 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6ss98" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.092934 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.092979 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.093457 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.093652 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.093722 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-mvvrj"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.093988 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-mvvrj" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.094208 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-vxdmn"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.094552 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-vxdmn" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.094570 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.095062 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vc7vq"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.095374 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-vc7vq" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.095492 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.095742 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.096193 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-pzbhd"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.096507 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-pzbhd" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.107075 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.107555 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.107610 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-hhxz6"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.107678 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.108236 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.108330 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.108366 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.108431 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.108442 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.108464 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.108702 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.108774 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.108831 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.109558 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29523420-qb78x"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.109745 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-hhxz6" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.109869 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.110232 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29523420-qb78x" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.111051 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-5smq4"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.112750 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.133604 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5smq4" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.135026 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.135201 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5chhg"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.135713 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5chhg" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.135829 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n9zlc"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.136204 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n9zlc" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.136236 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.136641 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.137863 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j5fdh"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.139784 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-z7mx2"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.139969 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j5fdh" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.142233 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-2sdct"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.142465 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-z7mx2" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.142785 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-26rzb"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.142806 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-lzjw5"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.142851 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2sdct" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.143833 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-md5fm"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.145793 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-zlp6j"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.146730 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-zlp6j" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.147307 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsbwl\" (UniqueName: \"kubernetes.io/projected/944ea0f1-a1d8-4912-b117-86e73f15b0fe-kube-api-access-jsbwl\") pod \"etcd-operator-b45778765-5wlpw\" (UID: \"944ea0f1-a1d8-4912-b117-86e73f15b0fe\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5wlpw" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.147329 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d1d69c14-a464-4f22-88ca-4e0cfe97111a-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-lgh4n\" (UID: \"d1d69c14-a464-4f22-88ca-4e0cfe97111a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lgh4n" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.147366 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d1d69c14-a464-4f22-88ca-4e0cfe97111a-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-lgh4n\" (UID: \"d1d69c14-a464-4f22-88ca-4e0cfe97111a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lgh4n" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.147393 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/679a536c-c470-4231-b3aa-bf7004cbcb4f-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-n2cfq\" (UID: \"679a536c-c470-4231-b3aa-bf7004cbcb4f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n2cfq" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.147408 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/679a536c-c470-4231-b3aa-bf7004cbcb4f-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-n2cfq\" (UID: \"679a536c-c470-4231-b3aa-bf7004cbcb4f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n2cfq" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.147425 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9a47b8c4-1288-4b4b-9d82-4c6c22cc65b3-serving-cert\") pod \"openshift-config-operator-7777fb866f-lzjw5\" (UID: \"9a47b8c4-1288-4b4b-9d82-4c6c22cc65b3\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-lzjw5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.147440 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0c42eda9-04e2-4035-afb6-3bde005af88a-webhook-cert\") pod \"packageserver-d55dfcdfc-55hcm\" (UID: \"0c42eda9-04e2-4035-afb6-3bde005af88a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-55hcm" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.147456 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/b1111d02-c210-4787-8dda-b118339cfc79-encryption-config\") pod \"apiserver-76f77b778f-lsngn\" (UID: \"b1111d02-c210-4787-8dda-b118339cfc79\") " pod="openshift-apiserver/apiserver-76f77b778f-lsngn" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.147471 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8129ffba-fc55-45de-945f-66071d25c4eb-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-kxkx5\" (UID: \"8129ffba-fc55-45de-945f-66071d25c4eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxkx5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.147483 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/8129ffba-fc55-45de-945f-66071d25c4eb-encryption-config\") pod \"apiserver-7bbb656c7d-kxkx5\" (UID: \"8129ffba-fc55-45de-945f-66071d25c4eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxkx5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.147496 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/9cf0e7d1-1645-47aa-add2-3a76bfc624a7-default-certificate\") pod \"router-default-5444994796-j9ccg\" (UID: \"9cf0e7d1-1645-47aa-add2-3a76bfc624a7\") " pod="openshift-ingress/router-default-5444994796-j9ccg" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.147510 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/c44b5f9b-46a0-4ca6-b6ac-456bcae851fa-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-r44z2\" (UID: \"c44b5f9b-46a0-4ca6-b6ac-456bcae851fa\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r44z2" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.147526 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8129ffba-fc55-45de-945f-66071d25c4eb-serving-cert\") pod \"apiserver-7bbb656c7d-kxkx5\" (UID: \"8129ffba-fc55-45de-945f-66071d25c4eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxkx5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.147541 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jg6sd\" (UniqueName: \"kubernetes.io/projected/b1111d02-c210-4787-8dda-b118339cfc79-kube-api-access-jg6sd\") pod \"apiserver-76f77b778f-lsngn\" (UID: \"b1111d02-c210-4787-8dda-b118339cfc79\") " pod="openshift-apiserver/apiserver-76f77b778f-lsngn" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.147560 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/d1d69c14-a464-4f22-88ca-4e0cfe97111a-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-lgh4n\" (UID: \"d1d69c14-a464-4f22-88ca-4e0cfe97111a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lgh4n" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.147574 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0c42eda9-04e2-4035-afb6-3bde005af88a-apiservice-cert\") pod \"packageserver-d55dfcdfc-55hcm\" (UID: \"0c42eda9-04e2-4035-afb6-3bde005af88a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-55hcm" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.147648 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/716bc347-cbf5-452c-a7c8-3eeff4e70870-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-4dfqf\" (UID: \"716bc347-cbf5-452c-a7c8-3eeff4e70870\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4dfqf" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.147664 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7ntm\" (UniqueName: \"kubernetes.io/projected/eb17d2da-96ad-44eb-a909-fcffbce9ecfb-kube-api-access-b7ntm\") pod \"catalog-operator-68c6474976-26rzb\" (UID: \"eb17d2da-96ad-44eb-a909-fcffbce9ecfb\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-26rzb" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.147697 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/9cf0e7d1-1645-47aa-add2-3a76bfc624a7-stats-auth\") pod \"router-default-5444994796-j9ccg\" (UID: \"9cf0e7d1-1645-47aa-add2-3a76bfc624a7\") " pod="openshift-ingress/router-default-5444994796-j9ccg" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.147726 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9cf0e7d1-1645-47aa-add2-3a76bfc624a7-metrics-certs\") pod \"router-default-5444994796-j9ccg\" (UID: \"9cf0e7d1-1645-47aa-add2-3a76bfc624a7\") " pod="openshift-ingress/router-default-5444994796-j9ccg" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.147748 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4xnt\" (UniqueName: \"kubernetes.io/projected/d1d69c14-a464-4f22-88ca-4e0cfe97111a-kube-api-access-x4xnt\") pod \"cluster-image-registry-operator-dc59b4c8b-lgh4n\" (UID: \"d1d69c14-a464-4f22-88ca-4e0cfe97111a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lgh4n" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.147766 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnz4s\" (UniqueName: \"kubernetes.io/projected/0c42eda9-04e2-4035-afb6-3bde005af88a-kube-api-access-fnz4s\") pod \"packageserver-d55dfcdfc-55hcm\" (UID: \"0c42eda9-04e2-4035-afb6-3bde005af88a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-55hcm" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.147781 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/8129ffba-fc55-45de-945f-66071d25c4eb-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-kxkx5\" (UID: \"8129ffba-fc55-45de-945f-66071d25c4eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxkx5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.147794 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/d3cd4918-7d0a-409d-9fe3-e168ce2e80fa-profile-collector-cert\") pod \"olm-operator-6b444d44fb-mgwg5\" (UID: \"d3cd4918-7d0a-409d-9fe3-e168ce2e80fa\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mgwg5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.147807 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/944ea0f1-a1d8-4912-b117-86e73f15b0fe-etcd-ca\") pod \"etcd-operator-b45778765-5wlpw\" (UID: \"944ea0f1-a1d8-4912-b117-86e73f15b0fe\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5wlpw" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.147845 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/944ea0f1-a1d8-4912-b117-86e73f15b0fe-config\") pod \"etcd-operator-b45778765-5wlpw\" (UID: \"944ea0f1-a1d8-4912-b117-86e73f15b0fe\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5wlpw" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.147871 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b1111d02-c210-4787-8dda-b118339cfc79-etcd-client\") pod \"apiserver-76f77b778f-lsngn\" (UID: \"b1111d02-c210-4787-8dda-b118339cfc79\") " pod="openshift-apiserver/apiserver-76f77b778f-lsngn" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.147887 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/0c42eda9-04e2-4035-afb6-3bde005af88a-tmpfs\") pod \"packageserver-d55dfcdfc-55hcm\" (UID: \"0c42eda9-04e2-4035-afb6-3bde005af88a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-55hcm" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.147925 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mh57\" (UniqueName: \"kubernetes.io/projected/6bfd85fe-faa2-48f8-9375-3a5a15214496-kube-api-access-5mh57\") pod \"machine-config-operator-74547568cd-md5fm\" (UID: \"6bfd85fe-faa2-48f8-9375-3a5a15214496\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-md5fm" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.147945 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b1111d02-c210-4787-8dda-b118339cfc79-serving-cert\") pod \"apiserver-76f77b778f-lsngn\" (UID: \"b1111d02-c210-4787-8dda-b118339cfc79\") " pod="openshift-apiserver/apiserver-76f77b778f-lsngn" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.147961 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/eb17d2da-96ad-44eb-a909-fcffbce9ecfb-srv-cert\") pod \"catalog-operator-68c6474976-26rzb\" (UID: \"eb17d2da-96ad-44eb-a909-fcffbce9ecfb\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-26rzb" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.147978 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzkkn\" (UniqueName: \"kubernetes.io/projected/716bc347-cbf5-452c-a7c8-3eeff4e70870-kube-api-access-jzkkn\") pod \"cluster-samples-operator-665b6dd947-4dfqf\" (UID: \"716bc347-cbf5-452c-a7c8-3eeff4e70870\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4dfqf" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.148033 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1111d02-c210-4787-8dda-b118339cfc79-config\") pod \"apiserver-76f77b778f-lsngn\" (UID: \"b1111d02-c210-4787-8dda-b118339cfc79\") " pod="openshift-apiserver/apiserver-76f77b778f-lsngn" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.148052 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/9a47b8c4-1288-4b4b-9d82-4c6c22cc65b3-available-featuregates\") pod \"openshift-config-operator-7777fb866f-lzjw5\" (UID: \"9a47b8c4-1288-4b4b-9d82-4c6c22cc65b3\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-lzjw5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.148074 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8129ffba-fc55-45de-945f-66071d25c4eb-audit-policies\") pod \"apiserver-7bbb656c7d-kxkx5\" (UID: \"8129ffba-fc55-45de-945f-66071d25c4eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxkx5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.148088 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6bfd85fe-faa2-48f8-9375-3a5a15214496-images\") pod \"machine-config-operator-74547568cd-md5fm\" (UID: \"6bfd85fe-faa2-48f8-9375-3a5a15214496\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-md5fm" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.148101 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-splsc\" (UniqueName: \"kubernetes.io/projected/9a47b8c4-1288-4b4b-9d82-4c6c22cc65b3-kube-api-access-splsc\") pod \"openshift-config-operator-7777fb866f-lzjw5\" (UID: \"9a47b8c4-1288-4b4b-9d82-4c6c22cc65b3\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-lzjw5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.148113 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/b1111d02-c210-4787-8dda-b118339cfc79-audit\") pod \"apiserver-76f77b778f-lsngn\" (UID: \"b1111d02-c210-4787-8dda-b118339cfc79\") " pod="openshift-apiserver/apiserver-76f77b778f-lsngn" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.148126 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9cf0e7d1-1645-47aa-add2-3a76bfc624a7-service-ca-bundle\") pod \"router-default-5444994796-j9ccg\" (UID: \"9cf0e7d1-1645-47aa-add2-3a76bfc624a7\") " pod="openshift-ingress/router-default-5444994796-j9ccg" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.148174 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6bfd85fe-faa2-48f8-9375-3a5a15214496-proxy-tls\") pod \"machine-config-operator-74547568cd-md5fm\" (UID: \"6bfd85fe-faa2-48f8-9375-3a5a15214496\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-md5fm" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.148188 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/944ea0f1-a1d8-4912-b117-86e73f15b0fe-etcd-service-ca\") pod \"etcd-operator-b45778765-5wlpw\" (UID: \"944ea0f1-a1d8-4912-b117-86e73f15b0fe\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5wlpw" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.148217 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/eb17d2da-96ad-44eb-a909-fcffbce9ecfb-profile-collector-cert\") pod \"catalog-operator-68c6474976-26rzb\" (UID: \"eb17d2da-96ad-44eb-a909-fcffbce9ecfb\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-26rzb" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.148268 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/d3cd4918-7d0a-409d-9fe3-e168ce2e80fa-srv-cert\") pod \"olm-operator-6b444d44fb-mgwg5\" (UID: \"d3cd4918-7d0a-409d-9fe3-e168ce2e80fa\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mgwg5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.148298 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/b1111d02-c210-4787-8dda-b118339cfc79-node-pullsecrets\") pod \"apiserver-76f77b778f-lsngn\" (UID: \"b1111d02-c210-4787-8dda-b118339cfc79\") " pod="openshift-apiserver/apiserver-76f77b778f-lsngn" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.148352 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qcpmk\" (UniqueName: \"kubernetes.io/projected/d3cd4918-7d0a-409d-9fe3-e168ce2e80fa-kube-api-access-qcpmk\") pod \"olm-operator-6b444d44fb-mgwg5\" (UID: \"d3cd4918-7d0a-409d-9fe3-e168ce2e80fa\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mgwg5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.148376 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/b1111d02-c210-4787-8dda-b118339cfc79-image-import-ca\") pod \"apiserver-76f77b778f-lsngn\" (UID: \"b1111d02-c210-4787-8dda-b118339cfc79\") " pod="openshift-apiserver/apiserver-76f77b778f-lsngn" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.148392 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8129ffba-fc55-45de-945f-66071d25c4eb-audit-dir\") pod \"apiserver-7bbb656c7d-kxkx5\" (UID: \"8129ffba-fc55-45de-945f-66071d25c4eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxkx5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.148410 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-55hcm"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.148419 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/b1111d02-c210-4787-8dda-b118339cfc79-etcd-serving-ca\") pod \"apiserver-76f77b778f-lsngn\" (UID: \"b1111d02-c210-4787-8dda-b118339cfc79\") " pod="openshift-apiserver/apiserver-76f77b778f-lsngn" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.148434 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/944ea0f1-a1d8-4912-b117-86e73f15b0fe-serving-cert\") pod \"etcd-operator-b45778765-5wlpw\" (UID: \"944ea0f1-a1d8-4912-b117-86e73f15b0fe\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5wlpw" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.148447 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvkz9\" (UniqueName: \"kubernetes.io/projected/8129ffba-fc55-45de-945f-66071d25c4eb-kube-api-access-zvkz9\") pod \"apiserver-7bbb656c7d-kxkx5\" (UID: \"8129ffba-fc55-45de-945f-66071d25c4eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxkx5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.148460 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2m56t\" (UniqueName: \"kubernetes.io/projected/9cf0e7d1-1645-47aa-add2-3a76bfc624a7-kube-api-access-2m56t\") pod \"router-default-5444994796-j9ccg\" (UID: \"9cf0e7d1-1645-47aa-add2-3a76bfc624a7\") " pod="openshift-ingress/router-default-5444994796-j9ccg" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.148473 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/679a536c-c470-4231-b3aa-bf7004cbcb4f-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-n2cfq\" (UID: \"679a536c-c470-4231-b3aa-bf7004cbcb4f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n2cfq" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.148497 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwr4x\" (UniqueName: \"kubernetes.io/projected/c44b5f9b-46a0-4ca6-b6ac-456bcae851fa-kube-api-access-lwr4x\") pod \"package-server-manager-789f6589d5-r44z2\" (UID: \"c44b5f9b-46a0-4ca6-b6ac-456bcae851fa\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r44z2" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.148542 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6bfd85fe-faa2-48f8-9375-3a5a15214496-auth-proxy-config\") pod \"machine-config-operator-74547568cd-md5fm\" (UID: \"6bfd85fe-faa2-48f8-9375-3a5a15214496\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-md5fm" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.148574 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8129ffba-fc55-45de-945f-66071d25c4eb-etcd-client\") pod \"apiserver-7bbb656c7d-kxkx5\" (UID: \"8129ffba-fc55-45de-945f-66071d25c4eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxkx5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.148600 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/944ea0f1-a1d8-4912-b117-86e73f15b0fe-etcd-client\") pod \"etcd-operator-b45778765-5wlpw\" (UID: \"944ea0f1-a1d8-4912-b117-86e73f15b0fe\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5wlpw" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.148634 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b1111d02-c210-4787-8dda-b118339cfc79-trusted-ca-bundle\") pod \"apiserver-76f77b778f-lsngn\" (UID: \"b1111d02-c210-4787-8dda-b118339cfc79\") " pod="openshift-apiserver/apiserver-76f77b778f-lsngn" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.148660 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b1111d02-c210-4787-8dda-b118339cfc79-audit-dir\") pod \"apiserver-76f77b778f-lsngn\" (UID: \"b1111d02-c210-4787-8dda-b118339cfc79\") " pod="openshift-apiserver/apiserver-76f77b778f-lsngn" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.149998 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-lsngn"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.152170 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.155017 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lgh4n"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.155924 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-6d57b"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.156925 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mgwg5"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.157568 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-kxkx5"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.158392 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n2cfq"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.159439 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-lv2k7"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.160003 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-wz622"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.162700 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-pzbhd"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.162720 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4dfqf"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.162730 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6ss98"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.164067 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-vd7f7"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.164923 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-mvvrj"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.165493 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-5wlpw"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.167452 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-zhtqs"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.169216 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r44z2"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.170303 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-fmk6k"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.171824 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-sxb49"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.172268 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lkxrl"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.174394 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.175219 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-8gpk5"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.176347 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5chhg"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.177556 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-vxdmn"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.177834 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29523420-qb78x"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.179384 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-hhxz6"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.180583 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n9zlc"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.182648 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-4vkc6"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.183731 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j5fdh"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.184584 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vc7vq"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.185597 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-dffhx"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.191281 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-5smq4"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.192650 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-zlp6j"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.193321 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.193655 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-z7mx2"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.195374 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-2sdct"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.196662 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-qw4hv"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.197471 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-qw4hv" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.198407 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-qw4hv"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.211952 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.231967 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.249233 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6bfd85fe-faa2-48f8-9375-3a5a15214496-proxy-tls\") pod \"machine-config-operator-74547568cd-md5fm\" (UID: \"6bfd85fe-faa2-48f8-9375-3a5a15214496\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-md5fm" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.249397 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/944ea0f1-a1d8-4912-b117-86e73f15b0fe-etcd-service-ca\") pod \"etcd-operator-b45778765-5wlpw\" (UID: \"944ea0f1-a1d8-4912-b117-86e73f15b0fe\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5wlpw" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.250061 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hq65b\" (UniqueName: \"kubernetes.io/projected/a303aa13-ef7e-4a65-a5d8-9763d2f0b4ea-kube-api-access-hq65b\") pod \"dns-operator-744455d44c-hhxz6\" (UID: \"a303aa13-ef7e-4a65-a5d8-9763d2f0b4ea\") " pod="openshift-dns-operator/dns-operator-744455d44c-hhxz6" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.250184 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/eb17d2da-96ad-44eb-a909-fcffbce9ecfb-profile-collector-cert\") pod \"catalog-operator-68c6474976-26rzb\" (UID: \"eb17d2da-96ad-44eb-a909-fcffbce9ecfb\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-26rzb" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.250271 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/d3cd4918-7d0a-409d-9fe3-e168ce2e80fa-srv-cert\") pod \"olm-operator-6b444d44fb-mgwg5\" (UID: \"d3cd4918-7d0a-409d-9fe3-e168ce2e80fa\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mgwg5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.250333 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/b1111d02-c210-4787-8dda-b118339cfc79-node-pullsecrets\") pod \"apiserver-76f77b778f-lsngn\" (UID: \"b1111d02-c210-4787-8dda-b118339cfc79\") " pod="openshift-apiserver/apiserver-76f77b778f-lsngn" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.250408 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/944ea0f1-a1d8-4912-b117-86e73f15b0fe-etcd-service-ca\") pod \"etcd-operator-b45778765-5wlpw\" (UID: \"944ea0f1-a1d8-4912-b117-86e73f15b0fe\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5wlpw" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.250417 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qcpmk\" (UniqueName: \"kubernetes.io/projected/d3cd4918-7d0a-409d-9fe3-e168ce2e80fa-kube-api-access-qcpmk\") pod \"olm-operator-6b444d44fb-mgwg5\" (UID: \"d3cd4918-7d0a-409d-9fe3-e168ce2e80fa\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mgwg5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.250459 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/75ba12a5-6494-4d1d-a2d8-b2674a55b3ca-serving-cert\") pod \"console-operator-58897d9998-lv2k7\" (UID: \"75ba12a5-6494-4d1d-a2d8-b2674a55b3ca\") " pod="openshift-console-operator/console-operator-58897d9998-lv2k7" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.250475 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a303aa13-ef7e-4a65-a5d8-9763d2f0b4ea-metrics-tls\") pod \"dns-operator-744455d44c-hhxz6\" (UID: \"a303aa13-ef7e-4a65-a5d8-9763d2f0b4ea\") " pod="openshift-dns-operator/dns-operator-744455d44c-hhxz6" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.250494 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/b1111d02-c210-4787-8dda-b118339cfc79-image-import-ca\") pod \"apiserver-76f77b778f-lsngn\" (UID: \"b1111d02-c210-4787-8dda-b118339cfc79\") " pod="openshift-apiserver/apiserver-76f77b778f-lsngn" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.250509 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5bc2daed-4513-47bb-ad46-52cc8c14f02b-config\") pod \"kube-controller-manager-operator-78b949d7b-5chhg\" (UID: \"5bc2daed-4513-47bb-ad46-52cc8c14f02b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5chhg" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.250527 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8129ffba-fc55-45de-945f-66071d25c4eb-audit-dir\") pod \"apiserver-7bbb656c7d-kxkx5\" (UID: \"8129ffba-fc55-45de-945f-66071d25c4eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxkx5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.250544 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/b1111d02-c210-4787-8dda-b118339cfc79-etcd-serving-ca\") pod \"apiserver-76f77b778f-lsngn\" (UID: \"b1111d02-c210-4787-8dda-b118339cfc79\") " pod="openshift-apiserver/apiserver-76f77b778f-lsngn" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.250853 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/944ea0f1-a1d8-4912-b117-86e73f15b0fe-serving-cert\") pod \"etcd-operator-b45778765-5wlpw\" (UID: \"944ea0f1-a1d8-4912-b117-86e73f15b0fe\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5wlpw" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.250873 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6f9fe7cf-4384-4a95-abbb-20b487e7bc6c-client-ca\") pod \"route-controller-manager-6576b87f9c-2sdct\" (UID: \"6f9fe7cf-4384-4a95-abbb-20b487e7bc6c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2sdct" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.250890 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvkz9\" (UniqueName: \"kubernetes.io/projected/8129ffba-fc55-45de-945f-66071d25c4eb-kube-api-access-zvkz9\") pod \"apiserver-7bbb656c7d-kxkx5\" (UID: \"8129ffba-fc55-45de-945f-66071d25c4eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxkx5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.250905 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2m56t\" (UniqueName: \"kubernetes.io/projected/9cf0e7d1-1645-47aa-add2-3a76bfc624a7-kube-api-access-2m56t\") pod \"router-default-5444994796-j9ccg\" (UID: \"9cf0e7d1-1645-47aa-add2-3a76bfc624a7\") " pod="openshift-ingress/router-default-5444994796-j9ccg" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.250920 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/679a536c-c470-4231-b3aa-bf7004cbcb4f-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-n2cfq\" (UID: \"679a536c-c470-4231-b3aa-bf7004cbcb4f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n2cfq" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.250955 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwr4x\" (UniqueName: \"kubernetes.io/projected/c44b5f9b-46a0-4ca6-b6ac-456bcae851fa-kube-api-access-lwr4x\") pod \"package-server-manager-789f6589d5-r44z2\" (UID: \"c44b5f9b-46a0-4ca6-b6ac-456bcae851fa\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r44z2" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.250972 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6bfd85fe-faa2-48f8-9375-3a5a15214496-auth-proxy-config\") pod \"machine-config-operator-74547568cd-md5fm\" (UID: \"6bfd85fe-faa2-48f8-9375-3a5a15214496\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-md5fm" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.250988 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8129ffba-fc55-45de-945f-66071d25c4eb-etcd-client\") pod \"apiserver-7bbb656c7d-kxkx5\" (UID: \"8129ffba-fc55-45de-945f-66071d25c4eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxkx5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.251184 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/b1111d02-c210-4787-8dda-b118339cfc79-node-pullsecrets\") pod \"apiserver-76f77b778f-lsngn\" (UID: \"b1111d02-c210-4787-8dda-b118339cfc79\") " pod="openshift-apiserver/apiserver-76f77b778f-lsngn" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.251706 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8129ffba-fc55-45de-945f-66071d25c4eb-audit-dir\") pod \"apiserver-7bbb656c7d-kxkx5\" (UID: \"8129ffba-fc55-45de-945f-66071d25c4eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxkx5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.251968 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/b1111d02-c210-4787-8dda-b118339cfc79-image-import-ca\") pod \"apiserver-76f77b778f-lsngn\" (UID: \"b1111d02-c210-4787-8dda-b118339cfc79\") " pod="openshift-apiserver/apiserver-76f77b778f-lsngn" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.252041 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/944ea0f1-a1d8-4912-b117-86e73f15b0fe-etcd-client\") pod \"etcd-operator-b45778765-5wlpw\" (UID: \"944ea0f1-a1d8-4912-b117-86e73f15b0fe\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5wlpw" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.252278 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/679a536c-c470-4231-b3aa-bf7004cbcb4f-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-n2cfq\" (UID: \"679a536c-c470-4231-b3aa-bf7004cbcb4f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n2cfq" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.253067 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/b1111d02-c210-4787-8dda-b118339cfc79-etcd-serving-ca\") pod \"apiserver-76f77b778f-lsngn\" (UID: \"b1111d02-c210-4787-8dda-b118339cfc79\") " pod="openshift-apiserver/apiserver-76f77b778f-lsngn" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.253398 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8926df7-dcbc-480d-bded-d4941bcc0d91-config\") pod \"openshift-apiserver-operator-796bbdcf4f-fmk6k\" (UID: \"e8926df7-dcbc-480d-bded-d4941bcc0d91\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-fmk6k" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.253430 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a457911e-1ca9-4f57-8286-b13f51549de7-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-4vkc6\" (UID: \"a457911e-1ca9-4f57-8286-b13f51549de7\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-4vkc6" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.253469 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f34c2f1a-5d5b-4719-82e3-d7534e16f7e6-console-serving-cert\") pod \"console-f9d7485db-8gpk5\" (UID: \"f34c2f1a-5d5b-4719-82e3-d7534e16f7e6\") " pod="openshift-console/console-f9d7485db-8gpk5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.253488 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b1111d02-c210-4787-8dda-b118339cfc79-trusted-ca-bundle\") pod \"apiserver-76f77b778f-lsngn\" (UID: \"b1111d02-c210-4787-8dda-b118339cfc79\") " pod="openshift-apiserver/apiserver-76f77b778f-lsngn" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.253504 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b1111d02-c210-4787-8dda-b118339cfc79-audit-dir\") pod \"apiserver-76f77b778f-lsngn\" (UID: \"b1111d02-c210-4787-8dda-b118339cfc79\") " pod="openshift-apiserver/apiserver-76f77b778f-lsngn" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.253582 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djx6h\" (UniqueName: \"kubernetes.io/projected/b3ca6b25-cb73-47a9-867c-8d1f6f628077-kube-api-access-djx6h\") pod \"control-plane-machine-set-operator-78cbb6b69f-j5fdh\" (UID: \"b3ca6b25-cb73-47a9-867c-8d1f6f628077\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j5fdh" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.253603 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a7a96d74-e090-45ac-a088-46cda659a8e7-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-6d57b\" (UID: \"a7a96d74-e090-45ac-a088-46cda659a8e7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-6d57b" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.253619 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f34c2f1a-5d5b-4719-82e3-d7534e16f7e6-console-config\") pod \"console-f9d7485db-8gpk5\" (UID: \"f34c2f1a-5d5b-4719-82e3-d7534e16f7e6\") " pod="openshift-console/console-f9d7485db-8gpk5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.253637 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsbwl\" (UniqueName: \"kubernetes.io/projected/944ea0f1-a1d8-4912-b117-86e73f15b0fe-kube-api-access-jsbwl\") pod \"etcd-operator-b45778765-5wlpw\" (UID: \"944ea0f1-a1d8-4912-b117-86e73f15b0fe\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5wlpw" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.253654 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d1d69c14-a464-4f22-88ca-4e0cfe97111a-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-lgh4n\" (UID: \"d1d69c14-a464-4f22-88ca-4e0cfe97111a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lgh4n" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.253658 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b1111d02-c210-4787-8dda-b118339cfc79-audit-dir\") pod \"apiserver-76f77b778f-lsngn\" (UID: \"b1111d02-c210-4787-8dda-b118339cfc79\") " pod="openshift-apiserver/apiserver-76f77b778f-lsngn" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.253668 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d1d69c14-a464-4f22-88ca-4e0cfe97111a-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-lgh4n\" (UID: \"d1d69c14-a464-4f22-88ca-4e0cfe97111a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lgh4n" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.253717 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/679a536c-c470-4231-b3aa-bf7004cbcb4f-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-n2cfq\" (UID: \"679a536c-c470-4231-b3aa-bf7004cbcb4f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n2cfq" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.253738 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/679a536c-c470-4231-b3aa-bf7004cbcb4f-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-n2cfq\" (UID: \"679a536c-c470-4231-b3aa-bf7004cbcb4f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n2cfq" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.253753 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9a47b8c4-1288-4b4b-9d82-4c6c22cc65b3-serving-cert\") pod \"openshift-config-operator-7777fb866f-lzjw5\" (UID: \"9a47b8c4-1288-4b4b-9d82-4c6c22cc65b3\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-lzjw5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.253601 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6bfd85fe-faa2-48f8-9375-3a5a15214496-auth-proxy-config\") pod \"machine-config-operator-74547568cd-md5fm\" (UID: \"6bfd85fe-faa2-48f8-9375-3a5a15214496\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-md5fm" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.253769 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0c42eda9-04e2-4035-afb6-3bde005af88a-webhook-cert\") pod \"packageserver-d55dfcdfc-55hcm\" (UID: \"0c42eda9-04e2-4035-afb6-3bde005af88a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-55hcm" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.253779 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.253785 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/b1111d02-c210-4787-8dda-b118339cfc79-encryption-config\") pod \"apiserver-76f77b778f-lsngn\" (UID: \"b1111d02-c210-4787-8dda-b118339cfc79\") " pod="openshift-apiserver/apiserver-76f77b778f-lsngn" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.254569 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8129ffba-fc55-45de-945f-66071d25c4eb-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-kxkx5\" (UID: \"8129ffba-fc55-45de-945f-66071d25c4eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxkx5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.254589 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/8129ffba-fc55-45de-945f-66071d25c4eb-encryption-config\") pod \"apiserver-7bbb656c7d-kxkx5\" (UID: \"8129ffba-fc55-45de-945f-66071d25c4eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxkx5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.254604 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/9cf0e7d1-1645-47aa-add2-3a76bfc624a7-default-certificate\") pod \"router-default-5444994796-j9ccg\" (UID: \"9cf0e7d1-1645-47aa-add2-3a76bfc624a7\") " pod="openshift-ingress/router-default-5444994796-j9ccg" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.254623 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/c44b5f9b-46a0-4ca6-b6ac-456bcae851fa-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-r44z2\" (UID: \"c44b5f9b-46a0-4ca6-b6ac-456bcae851fa\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r44z2" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.254640 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6f9fe7cf-4384-4a95-abbb-20b487e7bc6c-serving-cert\") pod \"route-controller-manager-6576b87f9c-2sdct\" (UID: \"6f9fe7cf-4384-4a95-abbb-20b487e7bc6c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2sdct" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.254657 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8129ffba-fc55-45de-945f-66071d25c4eb-serving-cert\") pod \"apiserver-7bbb656c7d-kxkx5\" (UID: \"8129ffba-fc55-45de-945f-66071d25c4eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxkx5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.254672 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jg6sd\" (UniqueName: \"kubernetes.io/projected/b1111d02-c210-4787-8dda-b118339cfc79-kube-api-access-jg6sd\") pod \"apiserver-76f77b778f-lsngn\" (UID: \"b1111d02-c210-4787-8dda-b118339cfc79\") " pod="openshift-apiserver/apiserver-76f77b778f-lsngn" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.254689 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5bc2daed-4513-47bb-ad46-52cc8c14f02b-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-5chhg\" (UID: \"5bc2daed-4513-47bb-ad46-52cc8c14f02b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5chhg" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.254702 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75ba12a5-6494-4d1d-a2d8-b2674a55b3ca-config\") pod \"console-operator-58897d9998-lv2k7\" (UID: \"75ba12a5-6494-4d1d-a2d8-b2674a55b3ca\") " pod="openshift-console-operator/console-operator-58897d9998-lv2k7" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.254726 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/d1d69c14-a464-4f22-88ca-4e0cfe97111a-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-lgh4n\" (UID: \"d1d69c14-a464-4f22-88ca-4e0cfe97111a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lgh4n" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.254741 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0c42eda9-04e2-4035-afb6-3bde005af88a-apiservice-cert\") pod \"packageserver-d55dfcdfc-55hcm\" (UID: \"0c42eda9-04e2-4035-afb6-3bde005af88a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-55hcm" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.254757 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/716bc347-cbf5-452c-a7c8-3eeff4e70870-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-4dfqf\" (UID: \"716bc347-cbf5-452c-a7c8-3eeff4e70870\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4dfqf" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.254773 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7ntm\" (UniqueName: \"kubernetes.io/projected/eb17d2da-96ad-44eb-a909-fcffbce9ecfb-kube-api-access-b7ntm\") pod \"catalog-operator-68c6474976-26rzb\" (UID: \"eb17d2da-96ad-44eb-a909-fcffbce9ecfb\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-26rzb" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.254787 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/9cf0e7d1-1645-47aa-add2-3a76bfc624a7-stats-auth\") pod \"router-default-5444994796-j9ccg\" (UID: \"9cf0e7d1-1645-47aa-add2-3a76bfc624a7\") " pod="openshift-ingress/router-default-5444994796-j9ccg" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.254800 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9cf0e7d1-1645-47aa-add2-3a76bfc624a7-metrics-certs\") pod \"router-default-5444994796-j9ccg\" (UID: \"9cf0e7d1-1645-47aa-add2-3a76bfc624a7\") " pod="openshift-ingress/router-default-5444994796-j9ccg" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.254815 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4xnt\" (UniqueName: \"kubernetes.io/projected/d1d69c14-a464-4f22-88ca-4e0cfe97111a-kube-api-access-x4xnt\") pod \"cluster-image-registry-operator-dc59b4c8b-lgh4n\" (UID: \"d1d69c14-a464-4f22-88ca-4e0cfe97111a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lgh4n" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.254831 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnz4s\" (UniqueName: \"kubernetes.io/projected/0c42eda9-04e2-4035-afb6-3bde005af88a-kube-api-access-fnz4s\") pod \"packageserver-d55dfcdfc-55hcm\" (UID: \"0c42eda9-04e2-4035-afb6-3bde005af88a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-55hcm" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.254845 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/8129ffba-fc55-45de-945f-66071d25c4eb-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-kxkx5\" (UID: \"8129ffba-fc55-45de-945f-66071d25c4eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxkx5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.254858 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/d3cd4918-7d0a-409d-9fe3-e168ce2e80fa-profile-collector-cert\") pod \"olm-operator-6b444d44fb-mgwg5\" (UID: \"d3cd4918-7d0a-409d-9fe3-e168ce2e80fa\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mgwg5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.254871 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/944ea0f1-a1d8-4912-b117-86e73f15b0fe-etcd-ca\") pod \"etcd-operator-b45778765-5wlpw\" (UID: \"944ea0f1-a1d8-4912-b117-86e73f15b0fe\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5wlpw" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.254877 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/944ea0f1-a1d8-4912-b117-86e73f15b0fe-serving-cert\") pod \"etcd-operator-b45778765-5wlpw\" (UID: \"944ea0f1-a1d8-4912-b117-86e73f15b0fe\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5wlpw" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.254890 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-946hm\" (UniqueName: \"kubernetes.io/projected/a7a96d74-e090-45ac-a088-46cda659a8e7-kube-api-access-946hm\") pod \"multus-admission-controller-857f4d67dd-6d57b\" (UID: \"a7a96d74-e090-45ac-a088-46cda659a8e7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-6d57b" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.254904 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftlsd\" (UniqueName: \"kubernetes.io/projected/f34c2f1a-5d5b-4719-82e3-d7534e16f7e6-kube-api-access-ftlsd\") pod \"console-f9d7485db-8gpk5\" (UID: \"f34c2f1a-5d5b-4719-82e3-d7534e16f7e6\") " pod="openshift-console/console-f9d7485db-8gpk5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.254919 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/944ea0f1-a1d8-4912-b117-86e73f15b0fe-config\") pod \"etcd-operator-b45778765-5wlpw\" (UID: \"944ea0f1-a1d8-4912-b117-86e73f15b0fe\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5wlpw" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.254935 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b1111d02-c210-4787-8dda-b118339cfc79-etcd-client\") pod \"apiserver-76f77b778f-lsngn\" (UID: \"b1111d02-c210-4787-8dda-b118339cfc79\") " pod="openshift-apiserver/apiserver-76f77b778f-lsngn" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.254950 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/0c42eda9-04e2-4035-afb6-3bde005af88a-tmpfs\") pod \"packageserver-d55dfcdfc-55hcm\" (UID: \"0c42eda9-04e2-4035-afb6-3bde005af88a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-55hcm" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.254965 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2wz8\" (UniqueName: \"kubernetes.io/projected/e8926df7-dcbc-480d-bded-d4941bcc0d91-kube-api-access-s2wz8\") pod \"openshift-apiserver-operator-796bbdcf4f-fmk6k\" (UID: \"e8926df7-dcbc-480d-bded-d4941bcc0d91\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-fmk6k" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.254981 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5mh57\" (UniqueName: \"kubernetes.io/projected/6bfd85fe-faa2-48f8-9375-3a5a15214496-kube-api-access-5mh57\") pod \"machine-config-operator-74547568cd-md5fm\" (UID: \"6bfd85fe-faa2-48f8-9375-3a5a15214496\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-md5fm" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.254995 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5bc2daed-4513-47bb-ad46-52cc8c14f02b-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-5chhg\" (UID: \"5bc2daed-4513-47bb-ad46-52cc8c14f02b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5chhg" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.255009 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b1111d02-c210-4787-8dda-b118339cfc79-serving-cert\") pod \"apiserver-76f77b778f-lsngn\" (UID: \"b1111d02-c210-4787-8dda-b118339cfc79\") " pod="openshift-apiserver/apiserver-76f77b778f-lsngn" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.255023 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/eb17d2da-96ad-44eb-a909-fcffbce9ecfb-srv-cert\") pod \"catalog-operator-68c6474976-26rzb\" (UID: \"eb17d2da-96ad-44eb-a909-fcffbce9ecfb\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-26rzb" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.255038 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzkkn\" (UniqueName: \"kubernetes.io/projected/716bc347-cbf5-452c-a7c8-3eeff4e70870-kube-api-access-jzkkn\") pod \"cluster-samples-operator-665b6dd947-4dfqf\" (UID: \"716bc347-cbf5-452c-a7c8-3eeff4e70870\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4dfqf" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.255051 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e8926df7-dcbc-480d-bded-d4941bcc0d91-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-fmk6k\" (UID: \"e8926df7-dcbc-480d-bded-d4941bcc0d91\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-fmk6k" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.255065 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f9fe7cf-4384-4a95-abbb-20b487e7bc6c-config\") pod \"route-controller-manager-6576b87f9c-2sdct\" (UID: \"6f9fe7cf-4384-4a95-abbb-20b487e7bc6c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2sdct" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.255088 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/75ba12a5-6494-4d1d-a2d8-b2674a55b3ca-trusted-ca\") pod \"console-operator-58897d9998-lv2k7\" (UID: \"75ba12a5-6494-4d1d-a2d8-b2674a55b3ca\") " pod="openshift-console-operator/console-operator-58897d9998-lv2k7" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.255105 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59tbs\" (UniqueName: \"kubernetes.io/projected/a457911e-1ca9-4f57-8286-b13f51549de7-kube-api-access-59tbs\") pod \"machine-config-controller-84d6567774-4vkc6\" (UID: \"a457911e-1ca9-4f57-8286-b13f51549de7\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-4vkc6" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.255122 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1111d02-c210-4787-8dda-b118339cfc79-config\") pod \"apiserver-76f77b778f-lsngn\" (UID: \"b1111d02-c210-4787-8dda-b118339cfc79\") " pod="openshift-apiserver/apiserver-76f77b778f-lsngn" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.255168 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/9a47b8c4-1288-4b4b-9d82-4c6c22cc65b3-available-featuregates\") pod \"openshift-config-operator-7777fb866f-lzjw5\" (UID: \"9a47b8c4-1288-4b4b-9d82-4c6c22cc65b3\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-lzjw5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.255185 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/b3ca6b25-cb73-47a9-867c-8d1f6f628077-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-j5fdh\" (UID: \"b3ca6b25-cb73-47a9-867c-8d1f6f628077\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j5fdh" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.255204 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a457911e-1ca9-4f57-8286-b13f51549de7-proxy-tls\") pod \"machine-config-controller-84d6567774-4vkc6\" (UID: \"a457911e-1ca9-4f57-8286-b13f51549de7\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-4vkc6" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.255219 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f34c2f1a-5d5b-4719-82e3-d7534e16f7e6-console-oauth-config\") pod \"console-f9d7485db-8gpk5\" (UID: \"f34c2f1a-5d5b-4719-82e3-d7534e16f7e6\") " pod="openshift-console/console-f9d7485db-8gpk5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.255236 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8129ffba-fc55-45de-945f-66071d25c4eb-audit-policies\") pod \"apiserver-7bbb656c7d-kxkx5\" (UID: \"8129ffba-fc55-45de-945f-66071d25c4eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxkx5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.255250 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6bfd85fe-faa2-48f8-9375-3a5a15214496-images\") pod \"machine-config-operator-74547568cd-md5fm\" (UID: \"6bfd85fe-faa2-48f8-9375-3a5a15214496\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-md5fm" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.255264 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f34c2f1a-5d5b-4719-82e3-d7534e16f7e6-service-ca\") pod \"console-f9d7485db-8gpk5\" (UID: \"f34c2f1a-5d5b-4719-82e3-d7534e16f7e6\") " pod="openshift-console/console-f9d7485db-8gpk5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.255278 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f34c2f1a-5d5b-4719-82e3-d7534e16f7e6-trusted-ca-bundle\") pod \"console-f9d7485db-8gpk5\" (UID: \"f34c2f1a-5d5b-4719-82e3-d7534e16f7e6\") " pod="openshift-console/console-f9d7485db-8gpk5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.255292 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f34c2f1a-5d5b-4719-82e3-d7534e16f7e6-oauth-serving-cert\") pod \"console-f9d7485db-8gpk5\" (UID: \"f34c2f1a-5d5b-4719-82e3-d7534e16f7e6\") " pod="openshift-console/console-f9d7485db-8gpk5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.255307 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-splsc\" (UniqueName: \"kubernetes.io/projected/9a47b8c4-1288-4b4b-9d82-4c6c22cc65b3-kube-api-access-splsc\") pod \"openshift-config-operator-7777fb866f-lzjw5\" (UID: \"9a47b8c4-1288-4b4b-9d82-4c6c22cc65b3\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-lzjw5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.255321 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2dfp\" (UniqueName: \"kubernetes.io/projected/75ba12a5-6494-4d1d-a2d8-b2674a55b3ca-kube-api-access-q2dfp\") pod \"console-operator-58897d9998-lv2k7\" (UID: \"75ba12a5-6494-4d1d-a2d8-b2674a55b3ca\") " pod="openshift-console-operator/console-operator-58897d9998-lv2k7" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.255337 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/b1111d02-c210-4787-8dda-b118339cfc79-audit\") pod \"apiserver-76f77b778f-lsngn\" (UID: \"b1111d02-c210-4787-8dda-b118339cfc79\") " pod="openshift-apiserver/apiserver-76f77b778f-lsngn" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.255352 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9cf0e7d1-1645-47aa-add2-3a76bfc624a7-service-ca-bundle\") pod \"router-default-5444994796-j9ccg\" (UID: \"9cf0e7d1-1645-47aa-add2-3a76bfc624a7\") " pod="openshift-ingress/router-default-5444994796-j9ccg" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.255367 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktxmv\" (UniqueName: \"kubernetes.io/projected/6f9fe7cf-4384-4a95-abbb-20b487e7bc6c-kube-api-access-ktxmv\") pod \"route-controller-manager-6576b87f9c-2sdct\" (UID: \"6f9fe7cf-4384-4a95-abbb-20b487e7bc6c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2sdct" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.255905 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6bfd85fe-faa2-48f8-9375-3a5a15214496-proxy-tls\") pod \"machine-config-operator-74547568cd-md5fm\" (UID: \"6bfd85fe-faa2-48f8-9375-3a5a15214496\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-md5fm" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.255943 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d1d69c14-a464-4f22-88ca-4e0cfe97111a-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-lgh4n\" (UID: \"d1d69c14-a464-4f22-88ca-4e0cfe97111a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lgh4n" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.255952 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-kvbqn"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.256421 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-gw4gv"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.256626 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b1111d02-c210-4787-8dda-b118339cfc79-trusted-ca-bundle\") pod \"apiserver-76f77b778f-lsngn\" (UID: \"b1111d02-c210-4787-8dda-b118339cfc79\") " pod="openshift-apiserver/apiserver-76f77b778f-lsngn" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.256782 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-gw4gv" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.256997 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-kvbqn" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.257407 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/716bc347-cbf5-452c-a7c8-3eeff4e70870-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-4dfqf\" (UID: \"716bc347-cbf5-452c-a7c8-3eeff4e70870\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4dfqf" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.257998 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1111d02-c210-4787-8dda-b118339cfc79-config\") pod \"apiserver-76f77b778f-lsngn\" (UID: \"b1111d02-c210-4787-8dda-b118339cfc79\") " pod="openshift-apiserver/apiserver-76f77b778f-lsngn" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.258664 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/9a47b8c4-1288-4b4b-9d82-4c6c22cc65b3-available-featuregates\") pod \"openshift-config-operator-7777fb866f-lzjw5\" (UID: \"9a47b8c4-1288-4b4b-9d82-4c6c22cc65b3\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-lzjw5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.259177 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8129ffba-fc55-45de-945f-66071d25c4eb-audit-policies\") pod \"apiserver-7bbb656c7d-kxkx5\" (UID: \"8129ffba-fc55-45de-945f-66071d25c4eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxkx5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.259972 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6bfd85fe-faa2-48f8-9375-3a5a15214496-images\") pod \"machine-config-operator-74547568cd-md5fm\" (UID: \"6bfd85fe-faa2-48f8-9375-3a5a15214496\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-md5fm" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.260206 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/944ea0f1-a1d8-4912-b117-86e73f15b0fe-config\") pod \"etcd-operator-b45778765-5wlpw\" (UID: \"944ea0f1-a1d8-4912-b117-86e73f15b0fe\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5wlpw" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.260292 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/8129ffba-fc55-45de-945f-66071d25c4eb-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-kxkx5\" (UID: \"8129ffba-fc55-45de-945f-66071d25c4eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxkx5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.260908 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/b1111d02-c210-4787-8dda-b118339cfc79-audit\") pod \"apiserver-76f77b778f-lsngn\" (UID: \"b1111d02-c210-4787-8dda-b118339cfc79\") " pod="openshift-apiserver/apiserver-76f77b778f-lsngn" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.261071 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-kvbqn"] Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.261077 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/0c42eda9-04e2-4035-afb6-3bde005af88a-tmpfs\") pod \"packageserver-d55dfcdfc-55hcm\" (UID: \"0c42eda9-04e2-4035-afb6-3bde005af88a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-55hcm" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.261255 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9cf0e7d1-1645-47aa-add2-3a76bfc624a7-service-ca-bundle\") pod \"router-default-5444994796-j9ccg\" (UID: \"9cf0e7d1-1645-47aa-add2-3a76bfc624a7\") " pod="openshift-ingress/router-default-5444994796-j9ccg" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.261659 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8129ffba-fc55-45de-945f-66071d25c4eb-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-kxkx5\" (UID: \"8129ffba-fc55-45de-945f-66071d25c4eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxkx5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.261702 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/8129ffba-fc55-45de-945f-66071d25c4eb-encryption-config\") pod \"apiserver-7bbb656c7d-kxkx5\" (UID: \"8129ffba-fc55-45de-945f-66071d25c4eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxkx5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.262062 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/679a536c-c470-4231-b3aa-bf7004cbcb4f-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-n2cfq\" (UID: \"679a536c-c470-4231-b3aa-bf7004cbcb4f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n2cfq" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.262298 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/9cf0e7d1-1645-47aa-add2-3a76bfc624a7-default-certificate\") pod \"router-default-5444994796-j9ccg\" (UID: \"9cf0e7d1-1645-47aa-add2-3a76bfc624a7\") " pod="openshift-ingress/router-default-5444994796-j9ccg" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.262677 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/944ea0f1-a1d8-4912-b117-86e73f15b0fe-etcd-ca\") pod \"etcd-operator-b45778765-5wlpw\" (UID: \"944ea0f1-a1d8-4912-b117-86e73f15b0fe\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5wlpw" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.263111 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/c44b5f9b-46a0-4ca6-b6ac-456bcae851fa-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-r44z2\" (UID: \"c44b5f9b-46a0-4ca6-b6ac-456bcae851fa\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r44z2" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.263247 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0c42eda9-04e2-4035-afb6-3bde005af88a-apiservice-cert\") pod \"packageserver-d55dfcdfc-55hcm\" (UID: \"0c42eda9-04e2-4035-afb6-3bde005af88a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-55hcm" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.263312 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8129ffba-fc55-45de-945f-66071d25c4eb-etcd-client\") pod \"apiserver-7bbb656c7d-kxkx5\" (UID: \"8129ffba-fc55-45de-945f-66071d25c4eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxkx5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.264417 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/944ea0f1-a1d8-4912-b117-86e73f15b0fe-etcd-client\") pod \"etcd-operator-b45778765-5wlpw\" (UID: \"944ea0f1-a1d8-4912-b117-86e73f15b0fe\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5wlpw" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.264805 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/d1d69c14-a464-4f22-88ca-4e0cfe97111a-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-lgh4n\" (UID: \"d1d69c14-a464-4f22-88ca-4e0cfe97111a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lgh4n" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.264918 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b1111d02-c210-4787-8dda-b118339cfc79-serving-cert\") pod \"apiserver-76f77b778f-lsngn\" (UID: \"b1111d02-c210-4787-8dda-b118339cfc79\") " pod="openshift-apiserver/apiserver-76f77b778f-lsngn" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.265435 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/eb17d2da-96ad-44eb-a909-fcffbce9ecfb-profile-collector-cert\") pod \"catalog-operator-68c6474976-26rzb\" (UID: \"eb17d2da-96ad-44eb-a909-fcffbce9ecfb\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-26rzb" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.265647 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0c42eda9-04e2-4035-afb6-3bde005af88a-webhook-cert\") pod \"packageserver-d55dfcdfc-55hcm\" (UID: \"0c42eda9-04e2-4035-afb6-3bde005af88a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-55hcm" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.265702 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b1111d02-c210-4787-8dda-b118339cfc79-etcd-client\") pod \"apiserver-76f77b778f-lsngn\" (UID: \"b1111d02-c210-4787-8dda-b118339cfc79\") " pod="openshift-apiserver/apiserver-76f77b778f-lsngn" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.265715 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/d3cd4918-7d0a-409d-9fe3-e168ce2e80fa-srv-cert\") pod \"olm-operator-6b444d44fb-mgwg5\" (UID: \"d3cd4918-7d0a-409d-9fe3-e168ce2e80fa\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mgwg5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.266098 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/eb17d2da-96ad-44eb-a909-fcffbce9ecfb-srv-cert\") pod \"catalog-operator-68c6474976-26rzb\" (UID: \"eb17d2da-96ad-44eb-a909-fcffbce9ecfb\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-26rzb" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.266100 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/d3cd4918-7d0a-409d-9fe3-e168ce2e80fa-profile-collector-cert\") pod \"olm-operator-6b444d44fb-mgwg5\" (UID: \"d3cd4918-7d0a-409d-9fe3-e168ce2e80fa\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mgwg5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.266251 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9cf0e7d1-1645-47aa-add2-3a76bfc624a7-metrics-certs\") pod \"router-default-5444994796-j9ccg\" (UID: \"9cf0e7d1-1645-47aa-add2-3a76bfc624a7\") " pod="openshift-ingress/router-default-5444994796-j9ccg" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.267261 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9a47b8c4-1288-4b4b-9d82-4c6c22cc65b3-serving-cert\") pod \"openshift-config-operator-7777fb866f-lzjw5\" (UID: \"9a47b8c4-1288-4b4b-9d82-4c6c22cc65b3\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-lzjw5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.267721 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8129ffba-fc55-45de-945f-66071d25c4eb-serving-cert\") pod \"apiserver-7bbb656c7d-kxkx5\" (UID: \"8129ffba-fc55-45de-945f-66071d25c4eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxkx5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.267793 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/9cf0e7d1-1645-47aa-add2-3a76bfc624a7-stats-auth\") pod \"router-default-5444994796-j9ccg\" (UID: \"9cf0e7d1-1645-47aa-add2-3a76bfc624a7\") " pod="openshift-ingress/router-default-5444994796-j9ccg" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.269448 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/b1111d02-c210-4787-8dda-b118339cfc79-encryption-config\") pod \"apiserver-76f77b778f-lsngn\" (UID: \"b1111d02-c210-4787-8dda-b118339cfc79\") " pod="openshift-apiserver/apiserver-76f77b778f-lsngn" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.272353 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.292319 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.312478 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.332443 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.351892 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.356299 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f34c2f1a-5d5b-4719-82e3-d7534e16f7e6-service-ca\") pod \"console-f9d7485db-8gpk5\" (UID: \"f34c2f1a-5d5b-4719-82e3-d7534e16f7e6\") " pod="openshift-console/console-f9d7485db-8gpk5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.356329 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f34c2f1a-5d5b-4719-82e3-d7534e16f7e6-trusted-ca-bundle\") pod \"console-f9d7485db-8gpk5\" (UID: \"f34c2f1a-5d5b-4719-82e3-d7534e16f7e6\") " pod="openshift-console/console-f9d7485db-8gpk5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.356347 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f34c2f1a-5d5b-4719-82e3-d7534e16f7e6-oauth-serving-cert\") pod \"console-f9d7485db-8gpk5\" (UID: \"f34c2f1a-5d5b-4719-82e3-d7534e16f7e6\") " pod="openshift-console/console-f9d7485db-8gpk5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.356368 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2dfp\" (UniqueName: \"kubernetes.io/projected/75ba12a5-6494-4d1d-a2d8-b2674a55b3ca-kube-api-access-q2dfp\") pod \"console-operator-58897d9998-lv2k7\" (UID: \"75ba12a5-6494-4d1d-a2d8-b2674a55b3ca\") " pod="openshift-console-operator/console-operator-58897d9998-lv2k7" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.356386 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktxmv\" (UniqueName: \"kubernetes.io/projected/6f9fe7cf-4384-4a95-abbb-20b487e7bc6c-kube-api-access-ktxmv\") pod \"route-controller-manager-6576b87f9c-2sdct\" (UID: \"6f9fe7cf-4384-4a95-abbb-20b487e7bc6c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2sdct" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.356410 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hq65b\" (UniqueName: \"kubernetes.io/projected/a303aa13-ef7e-4a65-a5d8-9763d2f0b4ea-kube-api-access-hq65b\") pod \"dns-operator-744455d44c-hhxz6\" (UID: \"a303aa13-ef7e-4a65-a5d8-9763d2f0b4ea\") " pod="openshift-dns-operator/dns-operator-744455d44c-hhxz6" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.356432 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a303aa13-ef7e-4a65-a5d8-9763d2f0b4ea-metrics-tls\") pod \"dns-operator-744455d44c-hhxz6\" (UID: \"a303aa13-ef7e-4a65-a5d8-9763d2f0b4ea\") " pod="openshift-dns-operator/dns-operator-744455d44c-hhxz6" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.356450 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/75ba12a5-6494-4d1d-a2d8-b2674a55b3ca-serving-cert\") pod \"console-operator-58897d9998-lv2k7\" (UID: \"75ba12a5-6494-4d1d-a2d8-b2674a55b3ca\") " pod="openshift-console-operator/console-operator-58897d9998-lv2k7" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.356466 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5bc2daed-4513-47bb-ad46-52cc8c14f02b-config\") pod \"kube-controller-manager-operator-78b949d7b-5chhg\" (UID: \"5bc2daed-4513-47bb-ad46-52cc8c14f02b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5chhg" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.356493 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6f9fe7cf-4384-4a95-abbb-20b487e7bc6c-client-ca\") pod \"route-controller-manager-6576b87f9c-2sdct\" (UID: \"6f9fe7cf-4384-4a95-abbb-20b487e7bc6c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2sdct" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.356512 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8926df7-dcbc-480d-bded-d4941bcc0d91-config\") pod \"openshift-apiserver-operator-796bbdcf4f-fmk6k\" (UID: \"e8926df7-dcbc-480d-bded-d4941bcc0d91\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-fmk6k" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.356526 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a457911e-1ca9-4f57-8286-b13f51549de7-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-4vkc6\" (UID: \"a457911e-1ca9-4f57-8286-b13f51549de7\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-4vkc6" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.356538 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f34c2f1a-5d5b-4719-82e3-d7534e16f7e6-console-serving-cert\") pod \"console-f9d7485db-8gpk5\" (UID: \"f34c2f1a-5d5b-4719-82e3-d7534e16f7e6\") " pod="openshift-console/console-f9d7485db-8gpk5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.356554 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djx6h\" (UniqueName: \"kubernetes.io/projected/b3ca6b25-cb73-47a9-867c-8d1f6f628077-kube-api-access-djx6h\") pod \"control-plane-machine-set-operator-78cbb6b69f-j5fdh\" (UID: \"b3ca6b25-cb73-47a9-867c-8d1f6f628077\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j5fdh" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.356579 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a7a96d74-e090-45ac-a088-46cda659a8e7-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-6d57b\" (UID: \"a7a96d74-e090-45ac-a088-46cda659a8e7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-6d57b" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.356592 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f34c2f1a-5d5b-4719-82e3-d7534e16f7e6-console-config\") pod \"console-f9d7485db-8gpk5\" (UID: \"f34c2f1a-5d5b-4719-82e3-d7534e16f7e6\") " pod="openshift-console/console-f9d7485db-8gpk5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.356633 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6f9fe7cf-4384-4a95-abbb-20b487e7bc6c-serving-cert\") pod \"route-controller-manager-6576b87f9c-2sdct\" (UID: \"6f9fe7cf-4384-4a95-abbb-20b487e7bc6c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2sdct" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.356656 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5bc2daed-4513-47bb-ad46-52cc8c14f02b-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-5chhg\" (UID: \"5bc2daed-4513-47bb-ad46-52cc8c14f02b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5chhg" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.356671 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75ba12a5-6494-4d1d-a2d8-b2674a55b3ca-config\") pod \"console-operator-58897d9998-lv2k7\" (UID: \"75ba12a5-6494-4d1d-a2d8-b2674a55b3ca\") " pod="openshift-console-operator/console-operator-58897d9998-lv2k7" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.356702 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-946hm\" (UniqueName: \"kubernetes.io/projected/a7a96d74-e090-45ac-a088-46cda659a8e7-kube-api-access-946hm\") pod \"multus-admission-controller-857f4d67dd-6d57b\" (UID: \"a7a96d74-e090-45ac-a088-46cda659a8e7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-6d57b" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.356717 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftlsd\" (UniqueName: \"kubernetes.io/projected/f34c2f1a-5d5b-4719-82e3-d7534e16f7e6-kube-api-access-ftlsd\") pod \"console-f9d7485db-8gpk5\" (UID: \"f34c2f1a-5d5b-4719-82e3-d7534e16f7e6\") " pod="openshift-console/console-f9d7485db-8gpk5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.356733 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2wz8\" (UniqueName: \"kubernetes.io/projected/e8926df7-dcbc-480d-bded-d4941bcc0d91-kube-api-access-s2wz8\") pod \"openshift-apiserver-operator-796bbdcf4f-fmk6k\" (UID: \"e8926df7-dcbc-480d-bded-d4941bcc0d91\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-fmk6k" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.356753 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5bc2daed-4513-47bb-ad46-52cc8c14f02b-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-5chhg\" (UID: \"5bc2daed-4513-47bb-ad46-52cc8c14f02b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5chhg" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.356771 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e8926df7-dcbc-480d-bded-d4941bcc0d91-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-fmk6k\" (UID: \"e8926df7-dcbc-480d-bded-d4941bcc0d91\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-fmk6k" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.356784 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f9fe7cf-4384-4a95-abbb-20b487e7bc6c-config\") pod \"route-controller-manager-6576b87f9c-2sdct\" (UID: \"6f9fe7cf-4384-4a95-abbb-20b487e7bc6c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2sdct" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.356804 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/75ba12a5-6494-4d1d-a2d8-b2674a55b3ca-trusted-ca\") pod \"console-operator-58897d9998-lv2k7\" (UID: \"75ba12a5-6494-4d1d-a2d8-b2674a55b3ca\") " pod="openshift-console-operator/console-operator-58897d9998-lv2k7" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.356820 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59tbs\" (UniqueName: \"kubernetes.io/projected/a457911e-1ca9-4f57-8286-b13f51549de7-kube-api-access-59tbs\") pod \"machine-config-controller-84d6567774-4vkc6\" (UID: \"a457911e-1ca9-4f57-8286-b13f51549de7\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-4vkc6" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.356835 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/b3ca6b25-cb73-47a9-867c-8d1f6f628077-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-j5fdh\" (UID: \"b3ca6b25-cb73-47a9-867c-8d1f6f628077\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j5fdh" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.356860 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f34c2f1a-5d5b-4719-82e3-d7534e16f7e6-console-oauth-config\") pod \"console-f9d7485db-8gpk5\" (UID: \"f34c2f1a-5d5b-4719-82e3-d7534e16f7e6\") " pod="openshift-console/console-f9d7485db-8gpk5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.356878 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a457911e-1ca9-4f57-8286-b13f51549de7-proxy-tls\") pod \"machine-config-controller-84d6567774-4vkc6\" (UID: \"a457911e-1ca9-4f57-8286-b13f51549de7\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-4vkc6" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.356977 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f34c2f1a-5d5b-4719-82e3-d7534e16f7e6-service-ca\") pod \"console-f9d7485db-8gpk5\" (UID: \"f34c2f1a-5d5b-4719-82e3-d7534e16f7e6\") " pod="openshift-console/console-f9d7485db-8gpk5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.357215 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f34c2f1a-5d5b-4719-82e3-d7534e16f7e6-trusted-ca-bundle\") pod \"console-f9d7485db-8gpk5\" (UID: \"f34c2f1a-5d5b-4719-82e3-d7534e16f7e6\") " pod="openshift-console/console-f9d7485db-8gpk5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.357697 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75ba12a5-6494-4d1d-a2d8-b2674a55b3ca-config\") pod \"console-operator-58897d9998-lv2k7\" (UID: \"75ba12a5-6494-4d1d-a2d8-b2674a55b3ca\") " pod="openshift-console-operator/console-operator-58897d9998-lv2k7" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.357722 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f34c2f1a-5d5b-4719-82e3-d7534e16f7e6-console-config\") pod \"console-f9d7485db-8gpk5\" (UID: \"f34c2f1a-5d5b-4719-82e3-d7534e16f7e6\") " pod="openshift-console/console-f9d7485db-8gpk5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.357872 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/75ba12a5-6494-4d1d-a2d8-b2674a55b3ca-trusted-ca\") pod \"console-operator-58897d9998-lv2k7\" (UID: \"75ba12a5-6494-4d1d-a2d8-b2674a55b3ca\") " pod="openshift-console-operator/console-operator-58897d9998-lv2k7" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.357957 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f34c2f1a-5d5b-4719-82e3-d7534e16f7e6-oauth-serving-cert\") pod \"console-f9d7485db-8gpk5\" (UID: \"f34c2f1a-5d5b-4719-82e3-d7534e16f7e6\") " pod="openshift-console/console-f9d7485db-8gpk5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.358811 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8926df7-dcbc-480d-bded-d4941bcc0d91-config\") pod \"openshift-apiserver-operator-796bbdcf4f-fmk6k\" (UID: \"e8926df7-dcbc-480d-bded-d4941bcc0d91\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-fmk6k" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.359125 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a457911e-1ca9-4f57-8286-b13f51549de7-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-4vkc6\" (UID: \"a457911e-1ca9-4f57-8286-b13f51549de7\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-4vkc6" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.360098 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f34c2f1a-5d5b-4719-82e3-d7534e16f7e6-console-oauth-config\") pod \"console-f9d7485db-8gpk5\" (UID: \"f34c2f1a-5d5b-4719-82e3-d7534e16f7e6\") " pod="openshift-console/console-f9d7485db-8gpk5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.360347 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a457911e-1ca9-4f57-8286-b13f51549de7-proxy-tls\") pod \"machine-config-controller-84d6567774-4vkc6\" (UID: \"a457911e-1ca9-4f57-8286-b13f51549de7\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-4vkc6" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.360414 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f34c2f1a-5d5b-4719-82e3-d7534e16f7e6-console-serving-cert\") pod \"console-f9d7485db-8gpk5\" (UID: \"f34c2f1a-5d5b-4719-82e3-d7534e16f7e6\") " pod="openshift-console/console-f9d7485db-8gpk5" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.360463 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e8926df7-dcbc-480d-bded-d4941bcc0d91-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-fmk6k\" (UID: \"e8926df7-dcbc-480d-bded-d4941bcc0d91\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-fmk6k" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.360777 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/75ba12a5-6494-4d1d-a2d8-b2674a55b3ca-serving-cert\") pod \"console-operator-58897d9998-lv2k7\" (UID: \"75ba12a5-6494-4d1d-a2d8-b2674a55b3ca\") " pod="openshift-console-operator/console-operator-58897d9998-lv2k7" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.361087 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a7a96d74-e090-45ac-a088-46cda659a8e7-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-6d57b\" (UID: \"a7a96d74-e090-45ac-a088-46cda659a8e7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-6d57b" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.372774 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.392547 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.412882 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.432059 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.452533 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.471946 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.492353 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.517605 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.531992 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.551897 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.572190 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.592820 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.612587 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.632643 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.652201 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.672236 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.692561 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.712027 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.732372 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.752143 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.781801 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.795835 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.812335 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.832166 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.852077 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.876887 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.892163 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.911753 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.932542 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.956748 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.971912 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Feb 18 09:06:00 crc kubenswrapper[4556]: I0218 09:06:00.992621 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.012752 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.032249 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.052025 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.072717 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.092570 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.111356 4556 request.go:700] Waited for 1.004043851s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-service-ca-operator/configmaps?fieldSelector=metadata.name%3Dservice-ca-operator-config&limit=500&resourceVersion=0 Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.112035 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.151861 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.172714 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.192749 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.212683 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.221181 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a303aa13-ef7e-4a65-a5d8-9763d2f0b4ea-metrics-tls\") pod \"dns-operator-744455d44c-hhxz6\" (UID: \"a303aa13-ef7e-4a65-a5d8-9763d2f0b4ea\") " pod="openshift-dns-operator/dns-operator-744455d44c-hhxz6" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.232172 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.252716 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.271986 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.281582 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.281626 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.281592 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.281730 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cgwv9" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.292125 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.311952 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.337042 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.353225 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Feb 18 09:06:01 crc kubenswrapper[4556]: E0218 09:06:01.357148 4556 secret.go:188] Couldn't get secret openshift-route-controller-manager/serving-cert: failed to sync secret cache: timed out waiting for the condition Feb 18 09:06:01 crc kubenswrapper[4556]: E0218 09:06:01.357265 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6f9fe7cf-4384-4a95-abbb-20b487e7bc6c-serving-cert podName:6f9fe7cf-4384-4a95-abbb-20b487e7bc6c nodeName:}" failed. No retries permitted until 2026-02-18 09:06:01.857251207 +0000 UTC m=+118.874212187 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/6f9fe7cf-4384-4a95-abbb-20b487e7bc6c-serving-cert") pod "route-controller-manager-6576b87f9c-2sdct" (UID: "6f9fe7cf-4384-4a95-abbb-20b487e7bc6c") : failed to sync secret cache: timed out waiting for the condition Feb 18 09:06:01 crc kubenswrapper[4556]: E0218 09:06:01.357192 4556 secret.go:188] Couldn't get secret openshift-kube-controller-manager-operator/kube-controller-manager-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Feb 18 09:06:01 crc kubenswrapper[4556]: E0218 09:06:01.357553 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5bc2daed-4513-47bb-ad46-52cc8c14f02b-serving-cert podName:5bc2daed-4513-47bb-ad46-52cc8c14f02b nodeName:}" failed. No retries permitted until 2026-02-18 09:06:01.857543165 +0000 UTC m=+118.874504146 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/5bc2daed-4513-47bb-ad46-52cc8c14f02b-serving-cert") pod "kube-controller-manager-operator-78b949d7b-5chhg" (UID: "5bc2daed-4513-47bb-ad46-52cc8c14f02b") : failed to sync secret cache: timed out waiting for the condition Feb 18 09:06:01 crc kubenswrapper[4556]: E0218 09:06:01.358289 4556 configmap.go:193] Couldn't get configMap openshift-route-controller-manager/config: failed to sync configmap cache: timed out waiting for the condition Feb 18 09:06:01 crc kubenswrapper[4556]: E0218 09:06:01.358355 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6f9fe7cf-4384-4a95-abbb-20b487e7bc6c-config podName:6f9fe7cf-4384-4a95-abbb-20b487e7bc6c nodeName:}" failed. No retries permitted until 2026-02-18 09:06:01.858339682 +0000 UTC m=+118.875300662 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/6f9fe7cf-4384-4a95-abbb-20b487e7bc6c-config") pod "route-controller-manager-6576b87f9c-2sdct" (UID: "6f9fe7cf-4384-4a95-abbb-20b487e7bc6c") : failed to sync configmap cache: timed out waiting for the condition Feb 18 09:06:01 crc kubenswrapper[4556]: E0218 09:06:01.358399 4556 configmap.go:193] Couldn't get configMap openshift-route-controller-manager/client-ca: failed to sync configmap cache: timed out waiting for the condition Feb 18 09:06:01 crc kubenswrapper[4556]: E0218 09:06:01.358433 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6f9fe7cf-4384-4a95-abbb-20b487e7bc6c-client-ca podName:6f9fe7cf-4384-4a95-abbb-20b487e7bc6c nodeName:}" failed. No retries permitted until 2026-02-18 09:06:01.858426505 +0000 UTC m=+118.875387485 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "client-ca" (UniqueName: "kubernetes.io/configmap/6f9fe7cf-4384-4a95-abbb-20b487e7bc6c-client-ca") pod "route-controller-manager-6576b87f9c-2sdct" (UID: "6f9fe7cf-4384-4a95-abbb-20b487e7bc6c") : failed to sync configmap cache: timed out waiting for the condition Feb 18 09:06:01 crc kubenswrapper[4556]: E0218 09:06:01.358451 4556 configmap.go:193] Couldn't get configMap openshift-kube-controller-manager-operator/kube-controller-manager-operator-config: failed to sync configmap cache: timed out waiting for the condition Feb 18 09:06:01 crc kubenswrapper[4556]: E0218 09:06:01.358497 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5bc2daed-4513-47bb-ad46-52cc8c14f02b-config podName:5bc2daed-4513-47bb-ad46-52cc8c14f02b nodeName:}" failed. No retries permitted until 2026-02-18 09:06:01.85848767 +0000 UTC m=+118.875448650 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/5bc2daed-4513-47bb-ad46-52cc8c14f02b-config") pod "kube-controller-manager-operator-78b949d7b-5chhg" (UID: "5bc2daed-4513-47bb-ad46-52cc8c14f02b") : failed to sync configmap cache: timed out waiting for the condition Feb 18 09:06:01 crc kubenswrapper[4556]: E0218 09:06:01.358404 4556 secret.go:188] Couldn't get secret openshift-machine-api/control-plane-machine-set-operator-tls: failed to sync secret cache: timed out waiting for the condition Feb 18 09:06:01 crc kubenswrapper[4556]: E0218 09:06:01.358609 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b3ca6b25-cb73-47a9-867c-8d1f6f628077-control-plane-machine-set-operator-tls podName:b3ca6b25-cb73-47a9-867c-8d1f6f628077 nodeName:}" failed. No retries permitted until 2026-02-18 09:06:01.858600723 +0000 UTC m=+118.875561703 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "control-plane-machine-set-operator-tls" (UniqueName: "kubernetes.io/secret/b3ca6b25-cb73-47a9-867c-8d1f6f628077-control-plane-machine-set-operator-tls") pod "control-plane-machine-set-operator-78cbb6b69f-j5fdh" (UID: "b3ca6b25-cb73-47a9-867c-8d1f6f628077") : failed to sync secret cache: timed out waiting for the condition Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.372435 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.392365 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.412346 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.432541 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.451966 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.471979 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.492226 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.512654 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.532090 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.552468 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.572838 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.592459 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.611888 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.633253 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.652684 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.672299 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.692801 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.712407 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.732599 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.751972 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.772387 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.792583 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.812206 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.832753 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.852280 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.872752 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.875343 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6f9fe7cf-4384-4a95-abbb-20b487e7bc6c-serving-cert\") pod \"route-controller-manager-6576b87f9c-2sdct\" (UID: \"6f9fe7cf-4384-4a95-abbb-20b487e7bc6c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2sdct" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.875382 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5bc2daed-4513-47bb-ad46-52cc8c14f02b-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-5chhg\" (UID: \"5bc2daed-4513-47bb-ad46-52cc8c14f02b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5chhg" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.875448 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f9fe7cf-4384-4a95-abbb-20b487e7bc6c-config\") pod \"route-controller-manager-6576b87f9c-2sdct\" (UID: \"6f9fe7cf-4384-4a95-abbb-20b487e7bc6c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2sdct" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.875487 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/b3ca6b25-cb73-47a9-867c-8d1f6f628077-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-j5fdh\" (UID: \"b3ca6b25-cb73-47a9-867c-8d1f6f628077\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j5fdh" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.875558 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5bc2daed-4513-47bb-ad46-52cc8c14f02b-config\") pod \"kube-controller-manager-operator-78b949d7b-5chhg\" (UID: \"5bc2daed-4513-47bb-ad46-52cc8c14f02b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5chhg" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.875590 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6f9fe7cf-4384-4a95-abbb-20b487e7bc6c-client-ca\") pod \"route-controller-manager-6576b87f9c-2sdct\" (UID: \"6f9fe7cf-4384-4a95-abbb-20b487e7bc6c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2sdct" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.876392 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5bc2daed-4513-47bb-ad46-52cc8c14f02b-config\") pod \"kube-controller-manager-operator-78b949d7b-5chhg\" (UID: \"5bc2daed-4513-47bb-ad46-52cc8c14f02b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5chhg" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.876574 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6f9fe7cf-4384-4a95-abbb-20b487e7bc6c-client-ca\") pod \"route-controller-manager-6576b87f9c-2sdct\" (UID: \"6f9fe7cf-4384-4a95-abbb-20b487e7bc6c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2sdct" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.876715 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f9fe7cf-4384-4a95-abbb-20b487e7bc6c-config\") pod \"route-controller-manager-6576b87f9c-2sdct\" (UID: \"6f9fe7cf-4384-4a95-abbb-20b487e7bc6c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2sdct" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.878780 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5bc2daed-4513-47bb-ad46-52cc8c14f02b-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-5chhg\" (UID: \"5bc2daed-4513-47bb-ad46-52cc8c14f02b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5chhg" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.878899 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/b3ca6b25-cb73-47a9-867c-8d1f6f628077-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-j5fdh\" (UID: \"b3ca6b25-cb73-47a9-867c-8d1f6f628077\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j5fdh" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.879769 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6f9fe7cf-4384-4a95-abbb-20b487e7bc6c-serving-cert\") pod \"route-controller-manager-6576b87f9c-2sdct\" (UID: \"6f9fe7cf-4384-4a95-abbb-20b487e7bc6c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2sdct" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.912774 4556 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.932585 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.951935 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Feb 18 09:06:01 crc kubenswrapper[4556]: I0218 09:06:01.983734 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvkz9\" (UniqueName: \"kubernetes.io/projected/8129ffba-fc55-45de-945f-66071d25c4eb-kube-api-access-zvkz9\") pod \"apiserver-7bbb656c7d-kxkx5\" (UID: \"8129ffba-fc55-45de-945f-66071d25c4eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxkx5" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.003938 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2m56t\" (UniqueName: \"kubernetes.io/projected/9cf0e7d1-1645-47aa-add2-3a76bfc624a7-kube-api-access-2m56t\") pod \"router-default-5444994796-j9ccg\" (UID: \"9cf0e7d1-1645-47aa-add2-3a76bfc624a7\") " pod="openshift-ingress/router-default-5444994796-j9ccg" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.033516 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwr4x\" (UniqueName: \"kubernetes.io/projected/c44b5f9b-46a0-4ca6-b6ac-456bcae851fa-kube-api-access-lwr4x\") pod \"package-server-manager-789f6589d5-r44z2\" (UID: \"c44b5f9b-46a0-4ca6-b6ac-456bcae851fa\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r44z2" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.050300 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qcpmk\" (UniqueName: \"kubernetes.io/projected/d3cd4918-7d0a-409d-9fe3-e168ce2e80fa-kube-api-access-qcpmk\") pod \"olm-operator-6b444d44fb-mgwg5\" (UID: \"d3cd4918-7d0a-409d-9fe3-e168ce2e80fa\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mgwg5" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.062726 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d1d69c14-a464-4f22-88ca-4e0cfe97111a-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-lgh4n\" (UID: \"d1d69c14-a464-4f22-88ca-4e0cfe97111a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lgh4n" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.083379 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsbwl\" (UniqueName: \"kubernetes.io/projected/944ea0f1-a1d8-4912-b117-86e73f15b0fe-kube-api-access-jsbwl\") pod \"etcd-operator-b45778765-5wlpw\" (UID: \"944ea0f1-a1d8-4912-b117-86e73f15b0fe\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5wlpw" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.102691 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/679a536c-c470-4231-b3aa-bf7004cbcb4f-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-n2cfq\" (UID: \"679a536c-c470-4231-b3aa-bf7004cbcb4f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n2cfq" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.111785 4556 request.go:700] Waited for 1.855060183s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-cluster-samples-operator/serviceaccounts/cluster-samples-operator/token Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.123431 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzkkn\" (UniqueName: \"kubernetes.io/projected/716bc347-cbf5-452c-a7c8-3eeff4e70870-kube-api-access-jzkkn\") pod \"cluster-samples-operator-665b6dd947-4dfqf\" (UID: \"716bc347-cbf5-452c-a7c8-3eeff4e70870\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4dfqf" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.143081 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7ntm\" (UniqueName: \"kubernetes.io/projected/eb17d2da-96ad-44eb-a909-fcffbce9ecfb-kube-api-access-b7ntm\") pod \"catalog-operator-68c6474976-26rzb\" (UID: \"eb17d2da-96ad-44eb-a909-fcffbce9ecfb\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-26rzb" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.152561 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.172180 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.174149 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-26rzb" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.188737 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.192478 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.207275 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxkx5" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.212869 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.234619 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.237470 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-j9ccg" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.244480 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-5wlpw" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.249708 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4dfqf" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.253500 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.258284 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n2cfq" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.264303 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r44z2" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.269621 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mgwg5" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.272535 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.310089 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-splsc\" (UniqueName: \"kubernetes.io/projected/9a47b8c4-1288-4b4b-9d82-4c6c22cc65b3-kube-api-access-splsc\") pod \"openshift-config-operator-7777fb866f-lzjw5\" (UID: \"9a47b8c4-1288-4b4b-9d82-4c6c22cc65b3\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-lzjw5" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.329061 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4xnt\" (UniqueName: \"kubernetes.io/projected/d1d69c14-a464-4f22-88ca-4e0cfe97111a-kube-api-access-x4xnt\") pod \"cluster-image-registry-operator-dc59b4c8b-lgh4n\" (UID: \"d1d69c14-a464-4f22-88ca-4e0cfe97111a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lgh4n" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.347926 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnz4s\" (UniqueName: \"kubernetes.io/projected/0c42eda9-04e2-4035-afb6-3bde005af88a-kube-api-access-fnz4s\") pod \"packageserver-d55dfcdfc-55hcm\" (UID: \"0c42eda9-04e2-4035-afb6-3bde005af88a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-55hcm" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.361746 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-26rzb"] Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.365835 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5mh57\" (UniqueName: \"kubernetes.io/projected/6bfd85fe-faa2-48f8-9375-3a5a15214496-kube-api-access-5mh57\") pod \"machine-config-operator-74547568cd-md5fm\" (UID: \"6bfd85fe-faa2-48f8-9375-3a5a15214496\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-md5fm" Feb 18 09:06:02 crc kubenswrapper[4556]: W0218 09:06:02.377240 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeb17d2da_96ad_44eb_a909_fcffbce9ecfb.slice/crio-dbcb25d916b09a80f998dc724b8e69509d0bcd6d7e91f126e333c3db3eb79cb0 WatchSource:0}: Error finding container dbcb25d916b09a80f998dc724b8e69509d0bcd6d7e91f126e333c3db3eb79cb0: Status 404 returned error can't find the container with id dbcb25d916b09a80f998dc724b8e69509d0bcd6d7e91f126e333c3db3eb79cb0 Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.390426 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-kxkx5"] Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.392067 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jg6sd\" (UniqueName: \"kubernetes.io/projected/b1111d02-c210-4787-8dda-b118339cfc79-kube-api-access-jg6sd\") pod \"apiserver-76f77b778f-lsngn\" (UID: \"b1111d02-c210-4787-8dda-b118339cfc79\") " pod="openshift-apiserver/apiserver-76f77b778f-lsngn" Feb 18 09:06:02 crc kubenswrapper[4556]: W0218 09:06:02.403660 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8129ffba_fc55_45de_945f_66071d25c4eb.slice/crio-6ecd65c1ba5cb66a9d184b07c458673e96b11de991021983508ea6e8fca40183 WatchSource:0}: Error finding container 6ecd65c1ba5cb66a9d184b07c458673e96b11de991021983508ea6e8fca40183: Status 404 returned error can't find the container with id 6ecd65c1ba5cb66a9d184b07c458673e96b11de991021983508ea6e8fca40183 Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.407137 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5bc2daed-4513-47bb-ad46-52cc8c14f02b-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-5chhg\" (UID: \"5bc2daed-4513-47bb-ad46-52cc8c14f02b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5chhg" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.426857 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-946hm\" (UniqueName: \"kubernetes.io/projected/a7a96d74-e090-45ac-a088-46cda659a8e7-kube-api-access-946hm\") pod \"multus-admission-controller-857f4d67dd-6d57b\" (UID: \"a7a96d74-e090-45ac-a088-46cda659a8e7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-6d57b" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.436994 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5chhg" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.438416 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-5wlpw"] Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.445415 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftlsd\" (UniqueName: \"kubernetes.io/projected/f34c2f1a-5d5b-4719-82e3-d7534e16f7e6-kube-api-access-ftlsd\") pod \"console-f9d7485db-8gpk5\" (UID: \"f34c2f1a-5d5b-4719-82e3-d7534e16f7e6\") " pod="openshift-console/console-f9d7485db-8gpk5" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.462868 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-lzjw5" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.463036 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n2cfq"] Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.470718 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2wz8\" (UniqueName: \"kubernetes.io/projected/e8926df7-dcbc-480d-bded-d4941bcc0d91-kube-api-access-s2wz8\") pod \"openshift-apiserver-operator-796bbdcf4f-fmk6k\" (UID: \"e8926df7-dcbc-480d-bded-d4941bcc0d91\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-fmk6k" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.485684 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r44z2"] Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.487084 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59tbs\" (UniqueName: \"kubernetes.io/projected/a457911e-1ca9-4f57-8286-b13f51549de7-kube-api-access-59tbs\") pod \"machine-config-controller-84d6567774-4vkc6\" (UID: \"a457911e-1ca9-4f57-8286-b13f51549de7\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-4vkc6" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.489935 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-md5fm" Feb 18 09:06:02 crc kubenswrapper[4556]: W0218 09:06:02.500929 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc44b5f9b_46a0_4ca6_b6ac_456bcae851fa.slice/crio-bfd120741be75047095917068dc0cf83bfa8ac7d176faf53f3732e6406854b65 WatchSource:0}: Error finding container bfd120741be75047095917068dc0cf83bfa8ac7d176faf53f3732e6406854b65: Status 404 returned error can't find the container with id bfd120741be75047095917068dc0cf83bfa8ac7d176faf53f3732e6406854b65 Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.505527 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2dfp\" (UniqueName: \"kubernetes.io/projected/75ba12a5-6494-4d1d-a2d8-b2674a55b3ca-kube-api-access-q2dfp\") pod \"console-operator-58897d9998-lv2k7\" (UID: \"75ba12a5-6494-4d1d-a2d8-b2674a55b3ca\") " pod="openshift-console-operator/console-operator-58897d9998-lv2k7" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.517408 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mgwg5"] Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.522735 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4dfqf"] Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.526333 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktxmv\" (UniqueName: \"kubernetes.io/projected/6f9fe7cf-4384-4a95-abbb-20b487e7bc6c-kube-api-access-ktxmv\") pod \"route-controller-manager-6576b87f9c-2sdct\" (UID: \"6f9fe7cf-4384-4a95-abbb-20b487e7bc6c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2sdct" Feb 18 09:06:02 crc kubenswrapper[4556]: W0218 09:06:02.549300 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd3cd4918_7d0a_409d_9fe3_e168ce2e80fa.slice/crio-060bdacb6a900fb10128babba3efcafc63b0d7eba4b24c32491671b824bea015 WatchSource:0}: Error finding container 060bdacb6a900fb10128babba3efcafc63b0d7eba4b24c32491671b824bea015: Status 404 returned error can't find the container with id 060bdacb6a900fb10128babba3efcafc63b0d7eba4b24c32491671b824bea015 Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.549765 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hq65b\" (UniqueName: \"kubernetes.io/projected/a303aa13-ef7e-4a65-a5d8-9763d2f0b4ea-kube-api-access-hq65b\") pod \"dns-operator-744455d44c-hhxz6\" (UID: \"a303aa13-ef7e-4a65-a5d8-9763d2f0b4ea\") " pod="openshift-dns-operator/dns-operator-744455d44c-hhxz6" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.570319 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djx6h\" (UniqueName: \"kubernetes.io/projected/b3ca6b25-cb73-47a9-867c-8d1f6f628077-kube-api-access-djx6h\") pod \"control-plane-machine-set-operator-78cbb6b69f-j5fdh\" (UID: \"b3ca6b25-cb73-47a9-867c-8d1f6f628077\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j5fdh" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.579590 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-55hcm" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.585618 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lgh4n" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.595398 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.595827 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-lsngn" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.597242 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-6d57b" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.605291 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-8gpk5" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.612845 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-fmk6k" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.613093 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.619200 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-lv2k7" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.633577 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.633789 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-lzjw5"] Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.649087 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5chhg"] Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.653142 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.673105 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Feb 18 09:06:02 crc kubenswrapper[4556]: W0218 09:06:02.681949 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9a47b8c4_1288_4b4b_9d82_4c6c22cc65b3.slice/crio-342f5457fab202ac162ceacd5a2678a46b7262d66faa07422f423a2f04c0df4d WatchSource:0}: Error finding container 342f5457fab202ac162ceacd5a2678a46b7262d66faa07422f423a2f04c0df4d: Status 404 returned error can't find the container with id 342f5457fab202ac162ceacd5a2678a46b7262d66faa07422f423a2f04c0df4d Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.692949 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-4vkc6" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.694282 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.705241 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-md5fm"] Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.719805 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n2cfq" event={"ID":"679a536c-c470-4231-b3aa-bf7004cbcb4f","Type":"ContainerStarted","Data":"07b5c9a9c8fabffc9cb4f7766500bade2b5669e658666bfdafc6c61292439b72"} Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.721818 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-hhxz6" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.722332 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxkx5" event={"ID":"8129ffba-fc55-45de-945f-66071d25c4eb","Type":"ContainerStarted","Data":"6ecd65c1ba5cb66a9d184b07c458673e96b11de991021983508ea6e8fca40183"} Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.729963 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4dfqf" event={"ID":"716bc347-cbf5-452c-a7c8-3eeff4e70870","Type":"ContainerStarted","Data":"8c1e59b150031dba51493a997838fdfb232d7d9ec98fe4015e9fe86092d74c1a"} Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.735415 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mgwg5" event={"ID":"d3cd4918-7d0a-409d-9fe3-e168ce2e80fa","Type":"ContainerStarted","Data":"060bdacb6a900fb10128babba3efcafc63b0d7eba4b24c32491671b824bea015"} Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.740958 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r44z2" event={"ID":"c44b5f9b-46a0-4ca6-b6ac-456bcae851fa","Type":"ContainerStarted","Data":"2db271a79643d14d99618d2605d7ecc235979f717dea281ad4b5f1fa96055bb0"} Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.740983 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r44z2" event={"ID":"c44b5f9b-46a0-4ca6-b6ac-456bcae851fa","Type":"ContainerStarted","Data":"bfd120741be75047095917068dc0cf83bfa8ac7d176faf53f3732e6406854b65"} Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.742567 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-5wlpw" event={"ID":"944ea0f1-a1d8-4912-b117-86e73f15b0fe","Type":"ContainerStarted","Data":"1a5596de825a5eee99e3db8d74b5ee8594e2e871566e7b9b3c4292d7378d7a97"} Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.746435 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-j9ccg" event={"ID":"9cf0e7d1-1645-47aa-add2-3a76bfc624a7","Type":"ContainerStarted","Data":"9a44a657cad9d3c26c7188c605f782cf718b91e079f116934dfea3d49d514244"} Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.746460 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-j9ccg" event={"ID":"9cf0e7d1-1645-47aa-add2-3a76bfc624a7","Type":"ContainerStarted","Data":"6d369ffee039443cf0083b16593fae625bf80db58ce55ee20c2fc14c1b905cd3"} Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.748231 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j5fdh" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.758685 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2sdct" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.759893 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-26rzb" event={"ID":"eb17d2da-96ad-44eb-a909-fcffbce9ecfb","Type":"ContainerStarted","Data":"815539cbbd77d181fda5939be990173ed233f8ae9b92a7edd7c7d6a0ea8f74dd"} Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.759942 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-26rzb" event={"ID":"eb17d2da-96ad-44eb-a909-fcffbce9ecfb","Type":"ContainerStarted","Data":"dbcb25d916b09a80f998dc724b8e69509d0bcd6d7e91f126e333c3db3eb79cb0"} Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.760063 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-26rzb" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.763224 4556 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-26rzb container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.763265 4556 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-26rzb" podUID="eb17d2da-96ad-44eb-a909-fcffbce9ecfb" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.768832 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-lzjw5" event={"ID":"9a47b8c4-1288-4b4b-9d82-4c6c22cc65b3","Type":"ContainerStarted","Data":"342f5457fab202ac162ceacd5a2678a46b7262d66faa07422f423a2f04c0df4d"} Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.769947 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5chhg" event={"ID":"5bc2daed-4513-47bb-ad46-52cc8c14f02b","Type":"ContainerStarted","Data":"5e2dbe5fce3c280662fdbd67763d2389b32e6fb62ff108a8a8c768c3d1076118"} Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.789629 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-vxdmn\" (UID: \"9db33ada-017a-4281-85c1-0c1530bc3bbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-vxdmn" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.789659 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1aa03714-6c33-42a8-a427-04a7ce839f21-bound-sa-token\") pod \"ingress-operator-5b745b69d9-5smq4\" (UID: \"1aa03714-6c33-42a8-a427-04a7ce839f21\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5smq4" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.789687 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-vxdmn\" (UID: \"9db33ada-017a-4281-85c1-0c1530bc3bbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-vxdmn" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.789708 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-vxdmn\" (UID: \"9db33ada-017a-4281-85c1-0c1530bc3bbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-vxdmn" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.789733 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7fb132fe-af75-4b1a-828f-5b1dcd074fd6-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-vc7vq\" (UID: \"7fb132fe-af75-4b1a-828f-5b1dcd074fd6\") " pod="openshift-marketplace/marketplace-operator-79b997595-vc7vq" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.789778 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff9016b7-a27a-4774-a7a2-498eee516d28-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-lkxrl\" (UID: \"ff9016b7-a27a-4774-a7a2-498eee516d28\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lkxrl" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.789796 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-677n8\" (UniqueName: \"kubernetes.io/projected/ff9016b7-a27a-4774-a7a2-498eee516d28-kube-api-access-677n8\") pod \"openshift-controller-manager-operator-756b6f6bc6-lkxrl\" (UID: \"ff9016b7-a27a-4774-a7a2-498eee516d28\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lkxrl" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.789813 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctf6x\" (UniqueName: \"kubernetes.io/projected/fd45fd0a-9812-4ff9-8bf7-91fda0bb0df3-kube-api-access-ctf6x\") pod \"migrator-59844c95c7-sxb49\" (UID: \"fd45fd0a-9812-4ff9-8bf7-91fda0bb0df3\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sxb49" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.789895 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/db10e059-4e0c-417d-b457-31fcd8734cc9-metrics-tls\") pod \"dns-default-zlp6j\" (UID: \"db10e059-4e0c-417d-b457-31fcd8734cc9\") " pod="openshift-dns/dns-default-zlp6j" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.789912 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/585ea2bd-474c-41e6-aa73-e7a62d02eb1c-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-n9zlc\" (UID: \"585ea2bd-474c-41e6-aa73-e7a62d02eb1c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n9zlc" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.789927 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g58cn\" (UniqueName: \"kubernetes.io/projected/87338513-3a60-432f-8edf-02695687e1e2-kube-api-access-g58cn\") pod \"downloads-7954f5f757-mvvrj\" (UID: \"87338513-3a60-432f-8edf-02695687e1e2\") " pod="openshift-console/downloads-7954f5f757-mvvrj" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.789970 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/709f253c-93c3-4aaf-b3db-630775991aeb-registry-tls\") pod \"image-registry-697d97f7c8-wz622\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.789984 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9023a93b-33ee-447e-8278-7b4d752d7a53-serving-cert\") pod \"controller-manager-879f6c89f-vd7f7\" (UID: \"9023a93b-33ee-447e-8278-7b4d752d7a53\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vd7f7" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.790004 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9db33ada-017a-4281-85c1-0c1530bc3bbc-audit-dir\") pod \"oauth-openshift-558db77b4-vxdmn\" (UID: \"9db33ada-017a-4281-85c1-0c1530bc3bbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-vxdmn" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.790020 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c567dda7-a353-4ff5-81fa-487517a51971-config\") pod \"service-ca-operator-777779d784-pzbhd\" (UID: \"c567dda7-a353-4ff5-81fa-487517a51971\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-pzbhd" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.790034 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d17a601-499d-4d9b-945a-557378ea7400-config\") pod \"kube-apiserver-operator-766d6c64bb-6ss98\" (UID: \"8d17a601-499d-4d9b-945a-557378ea7400\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6ss98" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.790063 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/709f253c-93c3-4aaf-b3db-630775991aeb-trusted-ca\") pod \"image-registry-697d97f7c8-wz622\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.790079 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgkkb\" (UniqueName: \"kubernetes.io/projected/a737bcdf-9a16-4220-842e-75a130efb35d-kube-api-access-lgkkb\") pod \"service-ca-9c57cc56f-z7mx2\" (UID: \"a737bcdf-9a16-4220-842e-75a130efb35d\") " pod="openshift-service-ca/service-ca-9c57cc56f-z7mx2" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.790093 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/db10e059-4e0c-417d-b457-31fcd8734cc9-config-volume\") pod \"dns-default-zlp6j\" (UID: \"db10e059-4e0c-417d-b457-31fcd8734cc9\") " pod="openshift-dns/dns-default-zlp6j" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.790110 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9023a93b-33ee-447e-8278-7b4d752d7a53-config\") pod \"controller-manager-879f6c89f-vd7f7\" (UID: \"9023a93b-33ee-447e-8278-7b4d752d7a53\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vd7f7" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.790132 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlvvv\" (UniqueName: \"kubernetes.io/projected/9023a93b-33ee-447e-8278-7b4d752d7a53-kube-api-access-jlvvv\") pod \"controller-manager-879f6c89f-vd7f7\" (UID: \"9023a93b-33ee-447e-8278-7b4d752d7a53\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vd7f7" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.790179 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lz2ml\" (UniqueName: \"kubernetes.io/projected/b5f35d95-f6b7-4068-8036-dffcbf955272-kube-api-access-lz2ml\") pod \"machine-api-operator-5694c8668f-zhtqs\" (UID: \"b5f35d95-f6b7-4068-8036-dffcbf955272\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zhtqs" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.791800 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1aa03714-6c33-42a8-a427-04a7ce839f21-trusted-ca\") pod \"ingress-operator-5b745b69d9-5smq4\" (UID: \"1aa03714-6c33-42a8-a427-04a7ce839f21\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5smq4" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.792486 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1aa03714-6c33-42a8-a427-04a7ce839f21-metrics-tls\") pod \"ingress-operator-5b745b69d9-5smq4\" (UID: \"1aa03714-6c33-42a8-a427-04a7ce839f21\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5smq4" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.792514 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/288bd6a4-25b6-483f-a1e7-7b07da5d63b6-machine-approver-tls\") pod \"machine-approver-56656f9798-dncx2\" (UID: \"288bd6a4-25b6-483f-a1e7-7b07da5d63b6\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dncx2" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.792546 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ac316465-87df-4f89-99d3-7633309d3158-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-dffhx\" (UID: \"ac316465-87df-4f89-99d3-7633309d3158\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dffhx" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.792566 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-vxdmn\" (UID: \"9db33ada-017a-4281-85c1-0c1530bc3bbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-vxdmn" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.792583 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5f35d95-f6b7-4068-8036-dffcbf955272-config\") pod \"machine-api-operator-5694c8668f-zhtqs\" (UID: \"b5f35d95-f6b7-4068-8036-dffcbf955272\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zhtqs" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.797977 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/585ea2bd-474c-41e6-aa73-e7a62d02eb1c-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-n9zlc\" (UID: \"585ea2bd-474c-41e6-aa73-e7a62d02eb1c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n9zlc" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.798002 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gtkk\" (UniqueName: \"kubernetes.io/projected/0aa2e566-fd95-498b-8df5-b63afb290572-kube-api-access-2gtkk\") pod \"collect-profiles-29523420-qb78x\" (UID: \"0aa2e566-fd95-498b-8df5-b63afb290572\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523420-qb78x" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.798021 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8d17a601-499d-4d9b-945a-557378ea7400-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-6ss98\" (UID: \"8d17a601-499d-4d9b-945a-557378ea7400\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6ss98" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.798035 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ac316465-87df-4f89-99d3-7633309d3158-serving-cert\") pod \"authentication-operator-69f744f599-dffhx\" (UID: \"ac316465-87df-4f89-99d3-7633309d3158\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dffhx" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.798092 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/a737bcdf-9a16-4220-842e-75a130efb35d-signing-cabundle\") pod \"service-ca-9c57cc56f-z7mx2\" (UID: \"a737bcdf-9a16-4220-842e-75a130efb35d\") " pod="openshift-service-ca/service-ca-9c57cc56f-z7mx2" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.798182 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wz622\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.798205 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ac316465-87df-4f89-99d3-7633309d3158-service-ca-bundle\") pod \"authentication-operator-69f744f599-dffhx\" (UID: \"ac316465-87df-4f89-99d3-7633309d3158\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dffhx" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.798232 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-vxdmn\" (UID: \"9db33ada-017a-4281-85c1-0c1530bc3bbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-vxdmn" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.798248 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-vxdmn\" (UID: \"9db33ada-017a-4281-85c1-0c1530bc3bbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-vxdmn" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.798274 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/709f253c-93c3-4aaf-b3db-630775991aeb-ca-trust-extracted\") pod \"image-registry-697d97f7c8-wz622\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.798290 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/b5f35d95-f6b7-4068-8036-dffcbf955272-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-zhtqs\" (UID: \"b5f35d95-f6b7-4068-8036-dffcbf955272\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zhtqs" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.798303 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/a737bcdf-9a16-4220-842e-75a130efb35d-signing-key\") pod \"service-ca-9c57cc56f-z7mx2\" (UID: \"a737bcdf-9a16-4220-842e-75a130efb35d\") " pod="openshift-service-ca/service-ca-9c57cc56f-z7mx2" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.798318 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzpr7\" (UniqueName: \"kubernetes.io/projected/ac316465-87df-4f89-99d3-7633309d3158-kube-api-access-lzpr7\") pod \"authentication-operator-69f744f599-dffhx\" (UID: \"ac316465-87df-4f89-99d3-7633309d3158\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dffhx" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.798332 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/288bd6a4-25b6-483f-a1e7-7b07da5d63b6-auth-proxy-config\") pod \"machine-approver-56656f9798-dncx2\" (UID: \"288bd6a4-25b6-483f-a1e7-7b07da5d63b6\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dncx2" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.798357 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-vxdmn\" (UID: \"9db33ada-017a-4281-85c1-0c1530bc3bbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-vxdmn" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.798381 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4pnl\" (UniqueName: \"kubernetes.io/projected/7fb132fe-af75-4b1a-828f-5b1dcd074fd6-kube-api-access-h4pnl\") pod \"marketplace-operator-79b997595-vc7vq\" (UID: \"7fb132fe-af75-4b1a-828f-5b1dcd074fd6\") " pod="openshift-marketplace/marketplace-operator-79b997595-vc7vq" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.798409 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/709f253c-93c3-4aaf-b3db-630775991aeb-bound-sa-token\") pod \"image-registry-697d97f7c8-wz622\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.798424 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac316465-87df-4f89-99d3-7633309d3158-config\") pod \"authentication-operator-69f744f599-dffhx\" (UID: \"ac316465-87df-4f89-99d3-7633309d3158\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dffhx" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.798439 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntd7h\" (UniqueName: \"kubernetes.io/projected/1aa03714-6c33-42a8-a427-04a7ce839f21-kube-api-access-ntd7h\") pod \"ingress-operator-5b745b69d9-5smq4\" (UID: \"1aa03714-6c33-42a8-a427-04a7ce839f21\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5smq4" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.798453 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ljnf\" (UniqueName: \"kubernetes.io/projected/585ea2bd-474c-41e6-aa73-e7a62d02eb1c-kube-api-access-9ljnf\") pod \"kube-storage-version-migrator-operator-b67b599dd-n9zlc\" (UID: \"585ea2bd-474c-41e6-aa73-e7a62d02eb1c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n9zlc" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.798469 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c567dda7-a353-4ff5-81fa-487517a51971-serving-cert\") pod \"service-ca-operator-777779d784-pzbhd\" (UID: \"c567dda7-a353-4ff5-81fa-487517a51971\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-pzbhd" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.798483 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ff9016b7-a27a-4774-a7a2-498eee516d28-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-lkxrl\" (UID: \"ff9016b7-a27a-4774-a7a2-498eee516d28\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lkxrl" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.798496 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/288bd6a4-25b6-483f-a1e7-7b07da5d63b6-config\") pod \"machine-approver-56656f9798-dncx2\" (UID: \"288bd6a4-25b6-483f-a1e7-7b07da5d63b6\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dncx2" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.798519 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-vxdmn\" (UID: \"9db33ada-017a-4281-85c1-0c1530bc3bbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-vxdmn" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.798532 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/709f253c-93c3-4aaf-b3db-630775991aeb-registry-certificates\") pod \"image-registry-697d97f7c8-wz622\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.798547 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7fb132fe-af75-4b1a-828f-5b1dcd074fd6-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-vc7vq\" (UID: \"7fb132fe-af75-4b1a-828f-5b1dcd074fd6\") " pod="openshift-marketplace/marketplace-operator-79b997595-vc7vq" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.798578 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9db33ada-017a-4281-85c1-0c1530bc3bbc-audit-policies\") pod \"oauth-openshift-558db77b4-vxdmn\" (UID: \"9db33ada-017a-4281-85c1-0c1530bc3bbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-vxdmn" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.798593 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8d17a601-499d-4d9b-945a-557378ea7400-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-6ss98\" (UID: \"8d17a601-499d-4d9b-945a-557378ea7400\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6ss98" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.798633 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-vxdmn\" (UID: \"9db33ada-017a-4281-85c1-0c1530bc3bbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-vxdmn" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.798652 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-vxdmn\" (UID: \"9db33ada-017a-4281-85c1-0c1530bc3bbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-vxdmn" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.798667 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0aa2e566-fd95-498b-8df5-b63afb290572-config-volume\") pod \"collect-profiles-29523420-qb78x\" (UID: \"0aa2e566-fd95-498b-8df5-b63afb290572\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523420-qb78x" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.798682 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/709f253c-93c3-4aaf-b3db-630775991aeb-installation-pull-secrets\") pod \"image-registry-697d97f7c8-wz622\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.798695 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9023a93b-33ee-447e-8278-7b4d752d7a53-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-vd7f7\" (UID: \"9023a93b-33ee-447e-8278-7b4d752d7a53\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vd7f7" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.798736 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kn29l\" (UniqueName: \"kubernetes.io/projected/9db33ada-017a-4281-85c1-0c1530bc3bbc-kube-api-access-kn29l\") pod \"oauth-openshift-558db77b4-vxdmn\" (UID: \"9db33ada-017a-4281-85c1-0c1530bc3bbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-vxdmn" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.798763 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9023a93b-33ee-447e-8278-7b4d752d7a53-client-ca\") pod \"controller-manager-879f6c89f-vd7f7\" (UID: \"9023a93b-33ee-447e-8278-7b4d752d7a53\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vd7f7" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.798793 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vrqj\" (UniqueName: \"kubernetes.io/projected/709f253c-93c3-4aaf-b3db-630775991aeb-kube-api-access-2vrqj\") pod \"image-registry-697d97f7c8-wz622\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.798810 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jtpw9\" (UniqueName: \"kubernetes.io/projected/db10e059-4e0c-417d-b457-31fcd8734cc9-kube-api-access-jtpw9\") pod \"dns-default-zlp6j\" (UID: \"db10e059-4e0c-417d-b457-31fcd8734cc9\") " pod="openshift-dns/dns-default-zlp6j" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.798827 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-vxdmn\" (UID: \"9db33ada-017a-4281-85c1-0c1530bc3bbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-vxdmn" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.798840 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jr5d8\" (UniqueName: \"kubernetes.io/projected/288bd6a4-25b6-483f-a1e7-7b07da5d63b6-kube-api-access-jr5d8\") pod \"machine-approver-56656f9798-dncx2\" (UID: \"288bd6a4-25b6-483f-a1e7-7b07da5d63b6\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dncx2" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.798855 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0aa2e566-fd95-498b-8df5-b63afb290572-secret-volume\") pod \"collect-profiles-29523420-qb78x\" (UID: \"0aa2e566-fd95-498b-8df5-b63afb290572\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523420-qb78x" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.798868 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/b5f35d95-f6b7-4068-8036-dffcbf955272-images\") pod \"machine-api-operator-5694c8668f-zhtqs\" (UID: \"b5f35d95-f6b7-4068-8036-dffcbf955272\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zhtqs" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.798883 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbxbm\" (UniqueName: \"kubernetes.io/projected/c567dda7-a353-4ff5-81fa-487517a51971-kube-api-access-rbxbm\") pod \"service-ca-operator-777779d784-pzbhd\" (UID: \"c567dda7-a353-4ff5-81fa-487517a51971\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-pzbhd" Feb 18 09:06:02 crc kubenswrapper[4556]: E0218 09:06:02.800989 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 09:06:03.300973071 +0000 UTC m=+120.317934051 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wz622" (UID: "709f253c-93c3-4aaf-b3db-630775991aeb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.900938 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.901267 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgkkb\" (UniqueName: \"kubernetes.io/projected/a737bcdf-9a16-4220-842e-75a130efb35d-kube-api-access-lgkkb\") pod \"service-ca-9c57cc56f-z7mx2\" (UID: \"a737bcdf-9a16-4220-842e-75a130efb35d\") " pod="openshift-service-ca/service-ca-9c57cc56f-z7mx2" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.901297 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/db10e059-4e0c-417d-b457-31fcd8734cc9-config-volume\") pod \"dns-default-zlp6j\" (UID: \"db10e059-4e0c-417d-b457-31fcd8734cc9\") " pod="openshift-dns/dns-default-zlp6j" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.901321 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9023a93b-33ee-447e-8278-7b4d752d7a53-config\") pod \"controller-manager-879f6c89f-vd7f7\" (UID: \"9023a93b-33ee-447e-8278-7b4d752d7a53\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vd7f7" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.901339 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlvvv\" (UniqueName: \"kubernetes.io/projected/9023a93b-33ee-447e-8278-7b4d752d7a53-kube-api-access-jlvvv\") pod \"controller-manager-879f6c89f-vd7f7\" (UID: \"9023a93b-33ee-447e-8278-7b4d752d7a53\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vd7f7" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.901375 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/709f253c-93c3-4aaf-b3db-630775991aeb-trusted-ca\") pod \"image-registry-697d97f7c8-wz622\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.901423 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/ccb1af48-7979-4f75-983e-7455ac471966-csi-data-dir\") pod \"csi-hostpathplugin-qw4hv\" (UID: \"ccb1af48-7979-4f75-983e-7455ac471966\") " pod="hostpath-provisioner/csi-hostpathplugin-qw4hv" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.901445 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lz2ml\" (UniqueName: \"kubernetes.io/projected/b5f35d95-f6b7-4068-8036-dffcbf955272-kube-api-access-lz2ml\") pod \"machine-api-operator-5694c8668f-zhtqs\" (UID: \"b5f35d95-f6b7-4068-8036-dffcbf955272\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zhtqs" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.901474 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1aa03714-6c33-42a8-a427-04a7ce839f21-trusted-ca\") pod \"ingress-operator-5b745b69d9-5smq4\" (UID: \"1aa03714-6c33-42a8-a427-04a7ce839f21\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5smq4" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.901495 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1aa03714-6c33-42a8-a427-04a7ce839f21-metrics-tls\") pod \"ingress-operator-5b745b69d9-5smq4\" (UID: \"1aa03714-6c33-42a8-a427-04a7ce839f21\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5smq4" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.901515 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/288bd6a4-25b6-483f-a1e7-7b07da5d63b6-machine-approver-tls\") pod \"machine-approver-56656f9798-dncx2\" (UID: \"288bd6a4-25b6-483f-a1e7-7b07da5d63b6\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dncx2" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.901551 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ac316465-87df-4f89-99d3-7633309d3158-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-dffhx\" (UID: \"ac316465-87df-4f89-99d3-7633309d3158\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dffhx" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.901570 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-vxdmn\" (UID: \"9db33ada-017a-4281-85c1-0c1530bc3bbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-vxdmn" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.901603 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5f35d95-f6b7-4068-8036-dffcbf955272-config\") pod \"machine-api-operator-5694c8668f-zhtqs\" (UID: \"b5f35d95-f6b7-4068-8036-dffcbf955272\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zhtqs" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.901622 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/585ea2bd-474c-41e6-aa73-e7a62d02eb1c-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-n9zlc\" (UID: \"585ea2bd-474c-41e6-aa73-e7a62d02eb1c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n9zlc" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.901643 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gtkk\" (UniqueName: \"kubernetes.io/projected/0aa2e566-fd95-498b-8df5-b63afb290572-kube-api-access-2gtkk\") pod \"collect-profiles-29523420-qb78x\" (UID: \"0aa2e566-fd95-498b-8df5-b63afb290572\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523420-qb78x" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.901662 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7fld\" (UniqueName: \"kubernetes.io/projected/6ea95a80-9834-4f76-a250-6c6394c0df87-kube-api-access-r7fld\") pod \"ingress-canary-kvbqn\" (UID: \"6ea95a80-9834-4f76-a250-6c6394c0df87\") " pod="openshift-ingress-canary/ingress-canary-kvbqn" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.901695 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/ccb1af48-7979-4f75-983e-7455ac471966-plugins-dir\") pod \"csi-hostpathplugin-qw4hv\" (UID: \"ccb1af48-7979-4f75-983e-7455ac471966\") " pod="hostpath-provisioner/csi-hostpathplugin-qw4hv" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.901726 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8d17a601-499d-4d9b-945a-557378ea7400-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-6ss98\" (UID: \"8d17a601-499d-4d9b-945a-557378ea7400\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6ss98" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.901744 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ac316465-87df-4f89-99d3-7633309d3158-serving-cert\") pod \"authentication-operator-69f744f599-dffhx\" (UID: \"ac316465-87df-4f89-99d3-7633309d3158\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dffhx" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.901766 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/a737bcdf-9a16-4220-842e-75a130efb35d-signing-cabundle\") pod \"service-ca-9c57cc56f-z7mx2\" (UID: \"a737bcdf-9a16-4220-842e-75a130efb35d\") " pod="openshift-service-ca/service-ca-9c57cc56f-z7mx2" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.901791 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ac316465-87df-4f89-99d3-7633309d3158-service-ca-bundle\") pod \"authentication-operator-69f744f599-dffhx\" (UID: \"ac316465-87df-4f89-99d3-7633309d3158\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dffhx" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.901846 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-vxdmn\" (UID: \"9db33ada-017a-4281-85c1-0c1530bc3bbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-vxdmn" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.901864 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-vxdmn\" (UID: \"9db33ada-017a-4281-85c1-0c1530bc3bbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-vxdmn" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.901884 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/ccb1af48-7979-4f75-983e-7455ac471966-socket-dir\") pod \"csi-hostpathplugin-qw4hv\" (UID: \"ccb1af48-7979-4f75-983e-7455ac471966\") " pod="hostpath-provisioner/csi-hostpathplugin-qw4hv" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.901924 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/709f253c-93c3-4aaf-b3db-630775991aeb-ca-trust-extracted\") pod \"image-registry-697d97f7c8-wz622\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.901941 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/b5f35d95-f6b7-4068-8036-dffcbf955272-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-zhtqs\" (UID: \"b5f35d95-f6b7-4068-8036-dffcbf955272\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zhtqs" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.901962 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-854xf\" (UniqueName: \"kubernetes.io/projected/1946a181-5c5e-49f4-980b-57135e0d8540-kube-api-access-854xf\") pod \"machine-config-server-gw4gv\" (UID: \"1946a181-5c5e-49f4-980b-57135e0d8540\") " pod="openshift-machine-config-operator/machine-config-server-gw4gv" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.901982 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzpr7\" (UniqueName: \"kubernetes.io/projected/ac316465-87df-4f89-99d3-7633309d3158-kube-api-access-lzpr7\") pod \"authentication-operator-69f744f599-dffhx\" (UID: \"ac316465-87df-4f89-99d3-7633309d3158\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dffhx" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.902000 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/288bd6a4-25b6-483f-a1e7-7b07da5d63b6-auth-proxy-config\") pod \"machine-approver-56656f9798-dncx2\" (UID: \"288bd6a4-25b6-483f-a1e7-7b07da5d63b6\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dncx2" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.902017 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/ccb1af48-7979-4f75-983e-7455ac471966-mountpoint-dir\") pod \"csi-hostpathplugin-qw4hv\" (UID: \"ccb1af48-7979-4f75-983e-7455ac471966\") " pod="hostpath-provisioner/csi-hostpathplugin-qw4hv" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.902037 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/a737bcdf-9a16-4220-842e-75a130efb35d-signing-key\") pod \"service-ca-9c57cc56f-z7mx2\" (UID: \"a737bcdf-9a16-4220-842e-75a130efb35d\") " pod="openshift-service-ca/service-ca-9c57cc56f-z7mx2" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.902065 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-vxdmn\" (UID: \"9db33ada-017a-4281-85c1-0c1530bc3bbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-vxdmn" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.902099 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4pnl\" (UniqueName: \"kubernetes.io/projected/7fb132fe-af75-4b1a-828f-5b1dcd074fd6-kube-api-access-h4pnl\") pod \"marketplace-operator-79b997595-vc7vq\" (UID: \"7fb132fe-af75-4b1a-828f-5b1dcd074fd6\") " pod="openshift-marketplace/marketplace-operator-79b997595-vc7vq" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.902129 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/709f253c-93c3-4aaf-b3db-630775991aeb-bound-sa-token\") pod \"image-registry-697d97f7c8-wz622\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.902190 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac316465-87df-4f89-99d3-7633309d3158-config\") pod \"authentication-operator-69f744f599-dffhx\" (UID: \"ac316465-87df-4f89-99d3-7633309d3158\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dffhx" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.902209 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntd7h\" (UniqueName: \"kubernetes.io/projected/1aa03714-6c33-42a8-a427-04a7ce839f21-kube-api-access-ntd7h\") pod \"ingress-operator-5b745b69d9-5smq4\" (UID: \"1aa03714-6c33-42a8-a427-04a7ce839f21\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5smq4" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.902240 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ljnf\" (UniqueName: \"kubernetes.io/projected/585ea2bd-474c-41e6-aa73-e7a62d02eb1c-kube-api-access-9ljnf\") pod \"kube-storage-version-migrator-operator-b67b599dd-n9zlc\" (UID: \"585ea2bd-474c-41e6-aa73-e7a62d02eb1c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n9zlc" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.902260 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c567dda7-a353-4ff5-81fa-487517a51971-serving-cert\") pod \"service-ca-operator-777779d784-pzbhd\" (UID: \"c567dda7-a353-4ff5-81fa-487517a51971\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-pzbhd" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.902282 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ff9016b7-a27a-4774-a7a2-498eee516d28-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-lkxrl\" (UID: \"ff9016b7-a27a-4774-a7a2-498eee516d28\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lkxrl" Feb 18 09:06:02 crc kubenswrapper[4556]: E0218 09:06:02.902332 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 09:06:03.402312651 +0000 UTC m=+120.419273632 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.902380 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/288bd6a4-25b6-483f-a1e7-7b07da5d63b6-config\") pod \"machine-approver-56656f9798-dncx2\" (UID: \"288bd6a4-25b6-483f-a1e7-7b07da5d63b6\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dncx2" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.902410 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-vxdmn\" (UID: \"9db33ada-017a-4281-85c1-0c1530bc3bbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-vxdmn" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.902438 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7fb132fe-af75-4b1a-828f-5b1dcd074fd6-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-vc7vq\" (UID: \"7fb132fe-af75-4b1a-828f-5b1dcd074fd6\") " pod="openshift-marketplace/marketplace-operator-79b997595-vc7vq" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.903328 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/288bd6a4-25b6-483f-a1e7-7b07da5d63b6-auth-proxy-config\") pod \"machine-approver-56656f9798-dncx2\" (UID: \"288bd6a4-25b6-483f-a1e7-7b07da5d63b6\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dncx2" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.904333 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac316465-87df-4f89-99d3-7633309d3158-config\") pod \"authentication-operator-69f744f599-dffhx\" (UID: \"ac316465-87df-4f89-99d3-7633309d3158\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dffhx" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.904532 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/709f253c-93c3-4aaf-b3db-630775991aeb-registry-certificates\") pod \"image-registry-697d97f7c8-wz622\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.904559 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9db33ada-017a-4281-85c1-0c1530bc3bbc-audit-policies\") pod \"oauth-openshift-558db77b4-vxdmn\" (UID: \"9db33ada-017a-4281-85c1-0c1530bc3bbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-vxdmn" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.904576 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8d17a601-499d-4d9b-945a-557378ea7400-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-6ss98\" (UID: \"8d17a601-499d-4d9b-945a-557378ea7400\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6ss98" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.904717 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/db10e059-4e0c-417d-b457-31fcd8734cc9-config-volume\") pod \"dns-default-zlp6j\" (UID: \"db10e059-4e0c-417d-b457-31fcd8734cc9\") " pod="openshift-dns/dns-default-zlp6j" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.904773 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/709f253c-93c3-4aaf-b3db-630775991aeb-ca-trust-extracted\") pod \"image-registry-697d97f7c8-wz622\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.905177 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/709f253c-93c3-4aaf-b3db-630775991aeb-trusted-ca\") pod \"image-registry-697d97f7c8-wz622\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.905500 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1aa03714-6c33-42a8-a427-04a7ce839f21-trusted-ca\") pod \"ingress-operator-5b745b69d9-5smq4\" (UID: \"1aa03714-6c33-42a8-a427-04a7ce839f21\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5smq4" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.905922 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ac316465-87df-4f89-99d3-7633309d3158-service-ca-bundle\") pod \"authentication-operator-69f744f599-dffhx\" (UID: \"ac316465-87df-4f89-99d3-7633309d3158\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dffhx" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.905942 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7fb132fe-af75-4b1a-828f-5b1dcd074fd6-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-vc7vq\" (UID: \"7fb132fe-af75-4b1a-828f-5b1dcd074fd6\") " pod="openshift-marketplace/marketplace-operator-79b997595-vc7vq" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.906210 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-vxdmn\" (UID: \"9db33ada-017a-4281-85c1-0c1530bc3bbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-vxdmn" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.907566 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9db33ada-017a-4281-85c1-0c1530bc3bbc-audit-policies\") pod \"oauth-openshift-558db77b4-vxdmn\" (UID: \"9db33ada-017a-4281-85c1-0c1530bc3bbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-vxdmn" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.908516 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-vxdmn\" (UID: \"9db33ada-017a-4281-85c1-0c1530bc3bbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-vxdmn" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.908757 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/a737bcdf-9a16-4220-842e-75a130efb35d-signing-cabundle\") pod \"service-ca-9c57cc56f-z7mx2\" (UID: \"a737bcdf-9a16-4220-842e-75a130efb35d\") " pod="openshift-service-ca/service-ca-9c57cc56f-z7mx2" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.909350 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/1946a181-5c5e-49f4-980b-57135e0d8540-certs\") pod \"machine-config-server-gw4gv\" (UID: \"1946a181-5c5e-49f4-980b-57135e0d8540\") " pod="openshift-machine-config-operator/machine-config-server-gw4gv" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.909385 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0aa2e566-fd95-498b-8df5-b63afb290572-config-volume\") pod \"collect-profiles-29523420-qb78x\" (UID: \"0aa2e566-fd95-498b-8df5-b63afb290572\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523420-qb78x" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.909403 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/ccb1af48-7979-4f75-983e-7455ac471966-registration-dir\") pod \"csi-hostpathplugin-qw4hv\" (UID: \"ccb1af48-7979-4f75-983e-7455ac471966\") " pod="hostpath-provisioner/csi-hostpathplugin-qw4hv" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.909443 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-vxdmn\" (UID: \"9db33ada-017a-4281-85c1-0c1530bc3bbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-vxdmn" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.909481 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-vxdmn\" (UID: \"9db33ada-017a-4281-85c1-0c1530bc3bbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-vxdmn" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.909539 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/709f253c-93c3-4aaf-b3db-630775991aeb-installation-pull-secrets\") pod \"image-registry-697d97f7c8-wz622\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.909570 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9023a93b-33ee-447e-8278-7b4d752d7a53-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-vd7f7\" (UID: \"9023a93b-33ee-447e-8278-7b4d752d7a53\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vd7f7" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.910232 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5f35d95-f6b7-4068-8036-dffcbf955272-config\") pod \"machine-api-operator-5694c8668f-zhtqs\" (UID: \"b5f35d95-f6b7-4068-8036-dffcbf955272\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zhtqs" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.910275 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ac316465-87df-4f89-99d3-7633309d3158-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-dffhx\" (UID: \"ac316465-87df-4f89-99d3-7633309d3158\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dffhx" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.910313 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/709f253c-93c3-4aaf-b3db-630775991aeb-registry-certificates\") pod \"image-registry-697d97f7c8-wz622\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.910535 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9023a93b-33ee-447e-8278-7b4d752d7a53-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-vd7f7\" (UID: \"9023a93b-33ee-447e-8278-7b4d752d7a53\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vd7f7" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.910964 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-vxdmn\" (UID: \"9db33ada-017a-4281-85c1-0c1530bc3bbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-vxdmn" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.911831 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0aa2e566-fd95-498b-8df5-b63afb290572-config-volume\") pod \"collect-profiles-29523420-qb78x\" (UID: \"0aa2e566-fd95-498b-8df5-b63afb290572\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523420-qb78x" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.911880 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/288bd6a4-25b6-483f-a1e7-7b07da5d63b6-config\") pod \"machine-approver-56656f9798-dncx2\" (UID: \"288bd6a4-25b6-483f-a1e7-7b07da5d63b6\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dncx2" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.912059 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ff9016b7-a27a-4774-a7a2-498eee516d28-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-lkxrl\" (UID: \"ff9016b7-a27a-4774-a7a2-498eee516d28\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lkxrl" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.912382 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/a737bcdf-9a16-4220-842e-75a130efb35d-signing-key\") pod \"service-ca-9c57cc56f-z7mx2\" (UID: \"a737bcdf-9a16-4220-842e-75a130efb35d\") " pod="openshift-service-ca/service-ca-9c57cc56f-z7mx2" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.912252 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/585ea2bd-474c-41e6-aa73-e7a62d02eb1c-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-n9zlc\" (UID: \"585ea2bd-474c-41e6-aa73-e7a62d02eb1c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n9zlc" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.914552 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-vxdmn\" (UID: \"9db33ada-017a-4281-85c1-0c1530bc3bbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-vxdmn" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.914572 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ac316465-87df-4f89-99d3-7633309d3158-serving-cert\") pod \"authentication-operator-69f744f599-dffhx\" (UID: \"ac316465-87df-4f89-99d3-7633309d3158\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dffhx" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.914589 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/b5f35d95-f6b7-4068-8036-dffcbf955272-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-zhtqs\" (UID: \"b5f35d95-f6b7-4068-8036-dffcbf955272\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zhtqs" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.914875 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-vxdmn\" (UID: \"9db33ada-017a-4281-85c1-0c1530bc3bbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-vxdmn" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.914882 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-vxdmn\" (UID: \"9db33ada-017a-4281-85c1-0c1530bc3bbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-vxdmn" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.915227 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rlw5\" (UniqueName: \"kubernetes.io/projected/ccb1af48-7979-4f75-983e-7455ac471966-kube-api-access-6rlw5\") pod \"csi-hostpathplugin-qw4hv\" (UID: \"ccb1af48-7979-4f75-983e-7455ac471966\") " pod="hostpath-provisioner/csi-hostpathplugin-qw4hv" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.915239 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/709f253c-93c3-4aaf-b3db-630775991aeb-installation-pull-secrets\") pod \"image-registry-697d97f7c8-wz622\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.915302 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kn29l\" (UniqueName: \"kubernetes.io/projected/9db33ada-017a-4281-85c1-0c1530bc3bbc-kube-api-access-kn29l\") pod \"oauth-openshift-558db77b4-vxdmn\" (UID: \"9db33ada-017a-4281-85c1-0c1530bc3bbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-vxdmn" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.915456 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9023a93b-33ee-447e-8278-7b4d752d7a53-client-ca\") pod \"controller-manager-879f6c89f-vd7f7\" (UID: \"9023a93b-33ee-447e-8278-7b4d752d7a53\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vd7f7" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.915493 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6ea95a80-9834-4f76-a250-6c6394c0df87-cert\") pod \"ingress-canary-kvbqn\" (UID: \"6ea95a80-9834-4f76-a250-6c6394c0df87\") " pod="openshift-ingress-canary/ingress-canary-kvbqn" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.915613 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vrqj\" (UniqueName: \"kubernetes.io/projected/709f253c-93c3-4aaf-b3db-630775991aeb-kube-api-access-2vrqj\") pod \"image-registry-697d97f7c8-wz622\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.915785 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jtpw9\" (UniqueName: \"kubernetes.io/projected/db10e059-4e0c-417d-b457-31fcd8734cc9-kube-api-access-jtpw9\") pod \"dns-default-zlp6j\" (UID: \"db10e059-4e0c-417d-b457-31fcd8734cc9\") " pod="openshift-dns/dns-default-zlp6j" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.915813 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-vxdmn\" (UID: \"9db33ada-017a-4281-85c1-0c1530bc3bbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-vxdmn" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.915833 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jr5d8\" (UniqueName: \"kubernetes.io/projected/288bd6a4-25b6-483f-a1e7-7b07da5d63b6-kube-api-access-jr5d8\") pod \"machine-approver-56656f9798-dncx2\" (UID: \"288bd6a4-25b6-483f-a1e7-7b07da5d63b6\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dncx2" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.915925 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0aa2e566-fd95-498b-8df5-b63afb290572-secret-volume\") pod \"collect-profiles-29523420-qb78x\" (UID: \"0aa2e566-fd95-498b-8df5-b63afb290572\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523420-qb78x" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.915976 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/b5f35d95-f6b7-4068-8036-dffcbf955272-images\") pod \"machine-api-operator-5694c8668f-zhtqs\" (UID: \"b5f35d95-f6b7-4068-8036-dffcbf955272\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zhtqs" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.915995 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbxbm\" (UniqueName: \"kubernetes.io/projected/c567dda7-a353-4ff5-81fa-487517a51971-kube-api-access-rbxbm\") pod \"service-ca-operator-777779d784-pzbhd\" (UID: \"c567dda7-a353-4ff5-81fa-487517a51971\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-pzbhd" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.916184 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/1946a181-5c5e-49f4-980b-57135e0d8540-node-bootstrap-token\") pod \"machine-config-server-gw4gv\" (UID: \"1946a181-5c5e-49f4-980b-57135e0d8540\") " pod="openshift-machine-config-operator/machine-config-server-gw4gv" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.916314 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-vxdmn\" (UID: \"9db33ada-017a-4281-85c1-0c1530bc3bbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-vxdmn" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.916393 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1aa03714-6c33-42a8-a427-04a7ce839f21-bound-sa-token\") pod \"ingress-operator-5b745b69d9-5smq4\" (UID: \"1aa03714-6c33-42a8-a427-04a7ce839f21\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5smq4" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.917850 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9023a93b-33ee-447e-8278-7b4d752d7a53-config\") pod \"controller-manager-879f6c89f-vd7f7\" (UID: \"9023a93b-33ee-447e-8278-7b4d752d7a53\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vd7f7" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.918177 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-vxdmn\" (UID: \"9db33ada-017a-4281-85c1-0c1530bc3bbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-vxdmn" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.918247 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/b5f35d95-f6b7-4068-8036-dffcbf955272-images\") pod \"machine-api-operator-5694c8668f-zhtqs\" (UID: \"b5f35d95-f6b7-4068-8036-dffcbf955272\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zhtqs" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.918901 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9023a93b-33ee-447e-8278-7b4d752d7a53-client-ca\") pod \"controller-manager-879f6c89f-vd7f7\" (UID: \"9023a93b-33ee-447e-8278-7b4d752d7a53\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vd7f7" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.920493 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c567dda7-a353-4ff5-81fa-487517a51971-serving-cert\") pod \"service-ca-operator-777779d784-pzbhd\" (UID: \"c567dda7-a353-4ff5-81fa-487517a51971\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-pzbhd" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.921957 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-vxdmn\" (UID: \"9db33ada-017a-4281-85c1-0c1530bc3bbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-vxdmn" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.922067 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-vxdmn\" (UID: \"9db33ada-017a-4281-85c1-0c1530bc3bbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-vxdmn" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.922257 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7fb132fe-af75-4b1a-828f-5b1dcd074fd6-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-vc7vq\" (UID: \"7fb132fe-af75-4b1a-828f-5b1dcd074fd6\") " pod="openshift-marketplace/marketplace-operator-79b997595-vc7vq" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.922365 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff9016b7-a27a-4774-a7a2-498eee516d28-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-lkxrl\" (UID: \"ff9016b7-a27a-4774-a7a2-498eee516d28\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lkxrl" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.922431 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-677n8\" (UniqueName: \"kubernetes.io/projected/ff9016b7-a27a-4774-a7a2-498eee516d28-kube-api-access-677n8\") pod \"openshift-controller-manager-operator-756b6f6bc6-lkxrl\" (UID: \"ff9016b7-a27a-4774-a7a2-498eee516d28\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lkxrl" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.922540 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctf6x\" (UniqueName: \"kubernetes.io/projected/fd45fd0a-9812-4ff9-8bf7-91fda0bb0df3-kube-api-access-ctf6x\") pod \"migrator-59844c95c7-sxb49\" (UID: \"fd45fd0a-9812-4ff9-8bf7-91fda0bb0df3\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sxb49" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.922616 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0aa2e566-fd95-498b-8df5-b63afb290572-secret-volume\") pod \"collect-profiles-29523420-qb78x\" (UID: \"0aa2e566-fd95-498b-8df5-b63afb290572\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523420-qb78x" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.922378 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1aa03714-6c33-42a8-a427-04a7ce839f21-metrics-tls\") pod \"ingress-operator-5b745b69d9-5smq4\" (UID: \"1aa03714-6c33-42a8-a427-04a7ce839f21\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5smq4" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.922264 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-vxdmn\" (UID: \"9db33ada-017a-4281-85c1-0c1530bc3bbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-vxdmn" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.923058 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/db10e059-4e0c-417d-b457-31fcd8734cc9-metrics-tls\") pod \"dns-default-zlp6j\" (UID: \"db10e059-4e0c-417d-b457-31fcd8734cc9\") " pod="openshift-dns/dns-default-zlp6j" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.923309 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/585ea2bd-474c-41e6-aa73-e7a62d02eb1c-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-n9zlc\" (UID: \"585ea2bd-474c-41e6-aa73-e7a62d02eb1c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n9zlc" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.923859 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/288bd6a4-25b6-483f-a1e7-7b07da5d63b6-machine-approver-tls\") pod \"machine-approver-56656f9798-dncx2\" (UID: \"288bd6a4-25b6-483f-a1e7-7b07da5d63b6\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dncx2" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.924380 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g58cn\" (UniqueName: \"kubernetes.io/projected/87338513-3a60-432f-8edf-02695687e1e2-kube-api-access-g58cn\") pod \"downloads-7954f5f757-mvvrj\" (UID: \"87338513-3a60-432f-8edf-02695687e1e2\") " pod="openshift-console/downloads-7954f5f757-mvvrj" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.924449 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/709f253c-93c3-4aaf-b3db-630775991aeb-registry-tls\") pod \"image-registry-697d97f7c8-wz622\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.924452 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff9016b7-a27a-4774-a7a2-498eee516d28-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-lkxrl\" (UID: \"ff9016b7-a27a-4774-a7a2-498eee516d28\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lkxrl" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.925167 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8d17a601-499d-4d9b-945a-557378ea7400-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-6ss98\" (UID: \"8d17a601-499d-4d9b-945a-557378ea7400\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6ss98" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.926076 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9023a93b-33ee-447e-8278-7b4d752d7a53-serving-cert\") pod \"controller-manager-879f6c89f-vd7f7\" (UID: \"9023a93b-33ee-447e-8278-7b4d752d7a53\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vd7f7" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.926188 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9db33ada-017a-4281-85c1-0c1530bc3bbc-audit-dir\") pod \"oauth-openshift-558db77b4-vxdmn\" (UID: \"9db33ada-017a-4281-85c1-0c1530bc3bbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-vxdmn" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.926213 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c567dda7-a353-4ff5-81fa-487517a51971-config\") pod \"service-ca-operator-777779d784-pzbhd\" (UID: \"c567dda7-a353-4ff5-81fa-487517a51971\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-pzbhd" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.926232 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d17a601-499d-4d9b-945a-557378ea7400-config\") pod \"kube-apiserver-operator-766d6c64bb-6ss98\" (UID: \"8d17a601-499d-4d9b-945a-557378ea7400\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6ss98" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.927277 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d17a601-499d-4d9b-945a-557378ea7400-config\") pod \"kube-apiserver-operator-766d6c64bb-6ss98\" (UID: \"8d17a601-499d-4d9b-945a-557378ea7400\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6ss98" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.927326 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9db33ada-017a-4281-85c1-0c1530bc3bbc-audit-dir\") pod \"oauth-openshift-558db77b4-vxdmn\" (UID: \"9db33ada-017a-4281-85c1-0c1530bc3bbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-vxdmn" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.927768 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/709f253c-93c3-4aaf-b3db-630775991aeb-registry-tls\") pod \"image-registry-697d97f7c8-wz622\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.929504 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c567dda7-a353-4ff5-81fa-487517a51971-config\") pod \"service-ca-operator-777779d784-pzbhd\" (UID: \"c567dda7-a353-4ff5-81fa-487517a51971\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-pzbhd" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.931005 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-vxdmn\" (UID: \"9db33ada-017a-4281-85c1-0c1530bc3bbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-vxdmn" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.931294 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-vxdmn\" (UID: \"9db33ada-017a-4281-85c1-0c1530bc3bbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-vxdmn" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.931662 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7fb132fe-af75-4b1a-828f-5b1dcd074fd6-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-vc7vq\" (UID: \"7fb132fe-af75-4b1a-828f-5b1dcd074fd6\") " pod="openshift-marketplace/marketplace-operator-79b997595-vc7vq" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.931731 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/585ea2bd-474c-41e6-aa73-e7a62d02eb1c-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-n9zlc\" (UID: \"585ea2bd-474c-41e6-aa73-e7a62d02eb1c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n9zlc" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.932865 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-vxdmn\" (UID: \"9db33ada-017a-4281-85c1-0c1530bc3bbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-vxdmn" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.938683 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9023a93b-33ee-447e-8278-7b4d752d7a53-serving-cert\") pod \"controller-manager-879f6c89f-vd7f7\" (UID: \"9023a93b-33ee-447e-8278-7b4d752d7a53\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vd7f7" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.942463 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/db10e059-4e0c-417d-b457-31fcd8734cc9-metrics-tls\") pod \"dns-default-zlp6j\" (UID: \"db10e059-4e0c-417d-b457-31fcd8734cc9\") " pod="openshift-dns/dns-default-zlp6j" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.966463 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gtkk\" (UniqueName: \"kubernetes.io/projected/0aa2e566-fd95-498b-8df5-b63afb290572-kube-api-access-2gtkk\") pod \"collect-profiles-29523420-qb78x\" (UID: \"0aa2e566-fd95-498b-8df5-b63afb290572\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523420-qb78x" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.981437 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4pnl\" (UniqueName: \"kubernetes.io/projected/7fb132fe-af75-4b1a-828f-5b1dcd074fd6-kube-api-access-h4pnl\") pod \"marketplace-operator-79b997595-vc7vq\" (UID: \"7fb132fe-af75-4b1a-828f-5b1dcd074fd6\") " pod="openshift-marketplace/marketplace-operator-79b997595-vc7vq" Feb 18 09:06:02 crc kubenswrapper[4556]: I0218 09:06:02.990287 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/709f253c-93c3-4aaf-b3db-630775991aeb-bound-sa-token\") pod \"image-registry-697d97f7c8-wz622\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.008920 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lgh4n"] Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.010137 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-vc7vq" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.010507 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ljnf\" (UniqueName: \"kubernetes.io/projected/585ea2bd-474c-41e6-aa73-e7a62d02eb1c-kube-api-access-9ljnf\") pod \"kube-storage-version-migrator-operator-b67b599dd-n9zlc\" (UID: \"585ea2bd-474c-41e6-aa73-e7a62d02eb1c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n9zlc" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.026356 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-55hcm"] Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.027201 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6ea95a80-9834-4f76-a250-6c6394c0df87-cert\") pod \"ingress-canary-kvbqn\" (UID: \"6ea95a80-9834-4f76-a250-6c6394c0df87\") " pod="openshift-ingress-canary/ingress-canary-kvbqn" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.027261 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/1946a181-5c5e-49f4-980b-57135e0d8540-node-bootstrap-token\") pod \"machine-config-server-gw4gv\" (UID: \"1946a181-5c5e-49f4-980b-57135e0d8540\") " pod="openshift-machine-config-operator/machine-config-server-gw4gv" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.027343 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29523420-qb78x" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.027355 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/ccb1af48-7979-4f75-983e-7455ac471966-csi-data-dir\") pod \"csi-hostpathplugin-qw4hv\" (UID: \"ccb1af48-7979-4f75-983e-7455ac471966\") " pod="hostpath-provisioner/csi-hostpathplugin-qw4hv" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.027381 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgkkb\" (UniqueName: \"kubernetes.io/projected/a737bcdf-9a16-4220-842e-75a130efb35d-kube-api-access-lgkkb\") pod \"service-ca-9c57cc56f-z7mx2\" (UID: \"a737bcdf-9a16-4220-842e-75a130efb35d\") " pod="openshift-service-ca/service-ca-9c57cc56f-z7mx2" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.027385 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7fld\" (UniqueName: \"kubernetes.io/projected/6ea95a80-9834-4f76-a250-6c6394c0df87-kube-api-access-r7fld\") pod \"ingress-canary-kvbqn\" (UID: \"6ea95a80-9834-4f76-a250-6c6394c0df87\") " pod="openshift-ingress-canary/ingress-canary-kvbqn" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.027416 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/ccb1af48-7979-4f75-983e-7455ac471966-plugins-dir\") pod \"csi-hostpathplugin-qw4hv\" (UID: \"ccb1af48-7979-4f75-983e-7455ac471966\") " pod="hostpath-provisioner/csi-hostpathplugin-qw4hv" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.027447 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wz622\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.027464 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/ccb1af48-7979-4f75-983e-7455ac471966-socket-dir\") pod \"csi-hostpathplugin-qw4hv\" (UID: \"ccb1af48-7979-4f75-983e-7455ac471966\") " pod="hostpath-provisioner/csi-hostpathplugin-qw4hv" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.027498 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-854xf\" (UniqueName: \"kubernetes.io/projected/1946a181-5c5e-49f4-980b-57135e0d8540-kube-api-access-854xf\") pod \"machine-config-server-gw4gv\" (UID: \"1946a181-5c5e-49f4-980b-57135e0d8540\") " pod="openshift-machine-config-operator/machine-config-server-gw4gv" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.027517 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/ccb1af48-7979-4f75-983e-7455ac471966-mountpoint-dir\") pod \"csi-hostpathplugin-qw4hv\" (UID: \"ccb1af48-7979-4f75-983e-7455ac471966\") " pod="hostpath-provisioner/csi-hostpathplugin-qw4hv" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.027574 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/1946a181-5c5e-49f4-980b-57135e0d8540-certs\") pod \"machine-config-server-gw4gv\" (UID: \"1946a181-5c5e-49f4-980b-57135e0d8540\") " pod="openshift-machine-config-operator/machine-config-server-gw4gv" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.027588 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/ccb1af48-7979-4f75-983e-7455ac471966-registration-dir\") pod \"csi-hostpathplugin-qw4hv\" (UID: \"ccb1af48-7979-4f75-983e-7455ac471966\") " pod="hostpath-provisioner/csi-hostpathplugin-qw4hv" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.027606 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rlw5\" (UniqueName: \"kubernetes.io/projected/ccb1af48-7979-4f75-983e-7455ac471966-kube-api-access-6rlw5\") pod \"csi-hostpathplugin-qw4hv\" (UID: \"ccb1af48-7979-4f75-983e-7455ac471966\") " pod="hostpath-provisioner/csi-hostpathplugin-qw4hv" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.027824 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/ccb1af48-7979-4f75-983e-7455ac471966-csi-data-dir\") pod \"csi-hostpathplugin-qw4hv\" (UID: \"ccb1af48-7979-4f75-983e-7455ac471966\") " pod="hostpath-provisioner/csi-hostpathplugin-qw4hv" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.027931 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/ccb1af48-7979-4f75-983e-7455ac471966-mountpoint-dir\") pod \"csi-hostpathplugin-qw4hv\" (UID: \"ccb1af48-7979-4f75-983e-7455ac471966\") " pod="hostpath-provisioner/csi-hostpathplugin-qw4hv" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.028114 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/ccb1af48-7979-4f75-983e-7455ac471966-plugins-dir\") pod \"csi-hostpathplugin-qw4hv\" (UID: \"ccb1af48-7979-4f75-983e-7455ac471966\") " pod="hostpath-provisioner/csi-hostpathplugin-qw4hv" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.028174 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/ccb1af48-7979-4f75-983e-7455ac471966-socket-dir\") pod \"csi-hostpathplugin-qw4hv\" (UID: \"ccb1af48-7979-4f75-983e-7455ac471966\") " pod="hostpath-provisioner/csi-hostpathplugin-qw4hv" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.028188 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/ccb1af48-7979-4f75-983e-7455ac471966-registration-dir\") pod \"csi-hostpathplugin-qw4hv\" (UID: \"ccb1af48-7979-4f75-983e-7455ac471966\") " pod="hostpath-provisioner/csi-hostpathplugin-qw4hv" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.030753 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/1946a181-5c5e-49f4-980b-57135e0d8540-certs\") pod \"machine-config-server-gw4gv\" (UID: \"1946a181-5c5e-49f4-980b-57135e0d8540\") " pod="openshift-machine-config-operator/machine-config-server-gw4gv" Feb 18 09:06:03 crc kubenswrapper[4556]: E0218 09:06:03.032561 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 09:06:03.532537902 +0000 UTC m=+120.549498882 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wz622" (UID: "709f253c-93c3-4aaf-b3db-630775991aeb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.042084 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n9zlc" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.044727 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/1946a181-5c5e-49f4-980b-57135e0d8540-node-bootstrap-token\") pod \"machine-config-server-gw4gv\" (UID: \"1946a181-5c5e-49f4-980b-57135e0d8540\") " pod="openshift-machine-config-operator/machine-config-server-gw4gv" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.048901 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntd7h\" (UniqueName: \"kubernetes.io/projected/1aa03714-6c33-42a8-a427-04a7ce839f21-kube-api-access-ntd7h\") pod \"ingress-operator-5b745b69d9-5smq4\" (UID: \"1aa03714-6c33-42a8-a427-04a7ce839f21\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5smq4" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.052144 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6ea95a80-9834-4f76-a250-6c6394c0df87-cert\") pod \"ingress-canary-kvbqn\" (UID: \"6ea95a80-9834-4f76-a250-6c6394c0df87\") " pod="openshift-ingress-canary/ingress-canary-kvbqn" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.055556 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-z7mx2" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.071605 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzpr7\" (UniqueName: \"kubernetes.io/projected/ac316465-87df-4f89-99d3-7633309d3158-kube-api-access-lzpr7\") pod \"authentication-operator-69f744f599-dffhx\" (UID: \"ac316465-87df-4f89-99d3-7633309d3158\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dffhx" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.084301 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-6d57b"] Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.090813 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlvvv\" (UniqueName: \"kubernetes.io/projected/9023a93b-33ee-447e-8278-7b4d752d7a53-kube-api-access-jlvvv\") pod \"controller-manager-879f6c89f-vd7f7\" (UID: \"9023a93b-33ee-447e-8278-7b4d752d7a53\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vd7f7" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.094013 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-lsngn"] Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.107581 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lz2ml\" (UniqueName: \"kubernetes.io/projected/b5f35d95-f6b7-4068-8036-dffcbf955272-kube-api-access-lz2ml\") pod \"machine-api-operator-5694c8668f-zhtqs\" (UID: \"b5f35d95-f6b7-4068-8036-dffcbf955272\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zhtqs" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.125511 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kn29l\" (UniqueName: \"kubernetes.io/projected/9db33ada-017a-4281-85c1-0c1530bc3bbc-kube-api-access-kn29l\") pod \"oauth-openshift-558db77b4-vxdmn\" (UID: \"9db33ada-017a-4281-85c1-0c1530bc3bbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-vxdmn" Feb 18 09:06:03 crc kubenswrapper[4556]: E0218 09:06:03.129635 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 09:06:03.629614602 +0000 UTC m=+120.646575583 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.130214 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.130278 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-8gpk5"] Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.130843 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wz622\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:06:03 crc kubenswrapper[4556]: E0218 09:06:03.131138 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 09:06:03.631115923 +0000 UTC m=+120.648076903 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wz622" (UID: "709f253c-93c3-4aaf-b3db-630775991aeb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.153562 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vrqj\" (UniqueName: \"kubernetes.io/projected/709f253c-93c3-4aaf-b3db-630775991aeb-kube-api-access-2vrqj\") pod \"image-registry-697d97f7c8-wz622\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.171807 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1aa03714-6c33-42a8-a427-04a7ce839f21-bound-sa-token\") pod \"ingress-operator-5b745b69d9-5smq4\" (UID: \"1aa03714-6c33-42a8-a427-04a7ce839f21\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5smq4" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.190398 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jtpw9\" (UniqueName: \"kubernetes.io/projected/db10e059-4e0c-417d-b457-31fcd8734cc9-kube-api-access-jtpw9\") pod \"dns-default-zlp6j\" (UID: \"db10e059-4e0c-417d-b457-31fcd8734cc9\") " pod="openshift-dns/dns-default-zlp6j" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.206318 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-2sdct"] Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.211913 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jr5d8\" (UniqueName: \"kubernetes.io/projected/288bd6a4-25b6-483f-a1e7-7b07da5d63b6-kube-api-access-jr5d8\") pod \"machine-approver-56656f9798-dncx2\" (UID: \"288bd6a4-25b6-483f-a1e7-7b07da5d63b6\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dncx2" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.228756 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-vd7f7" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.231585 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 09:06:03 crc kubenswrapper[4556]: E0218 09:06:03.231939 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 09:06:03.731926209 +0000 UTC m=+120.748887189 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.240997 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbxbm\" (UniqueName: \"kubernetes.io/projected/c567dda7-a353-4ff5-81fa-487517a51971-kube-api-access-rbxbm\") pod \"service-ca-operator-777779d784-pzbhd\" (UID: \"c567dda7-a353-4ff5-81fa-487517a51971\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-pzbhd" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.241640 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-j9ccg" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.241680 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-lv2k7"] Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.242597 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-zhtqs" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.245514 4556 patch_prober.go:28] interesting pod/router-default-5444994796-j9ccg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 18 09:06:03 crc kubenswrapper[4556]: [-]has-synced failed: reason withheld Feb 18 09:06:03 crc kubenswrapper[4556]: [+]process-running ok Feb 18 09:06:03 crc kubenswrapper[4556]: healthz check failed Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.245546 4556 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-j9ccg" podUID="9cf0e7d1-1645-47aa-add2-3a76bfc624a7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.250010 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-fmk6k"] Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.333218 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wz622\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:06:03 crc kubenswrapper[4556]: E0218 09:06:03.333748 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 09:06:03.833735682 +0000 UTC m=+120.850696662 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wz622" (UID: "709f253c-93c3-4aaf-b3db-630775991aeb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.335866 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-hhxz6"] Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.344225 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n9zlc"] Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.350200 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j5fdh"] Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.367629 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g58cn\" (UniqueName: \"kubernetes.io/projected/87338513-3a60-432f-8edf-02695687e1e2-kube-api-access-g58cn\") pod \"downloads-7954f5f757-mvvrj\" (UID: \"87338513-3a60-432f-8edf-02695687e1e2\") " pod="openshift-console/downloads-7954f5f757-mvvrj" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.394802 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-4vkc6"] Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.400578 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29523420-qb78x"] Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.421433 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-854xf\" (UniqueName: \"kubernetes.io/projected/1946a181-5c5e-49f4-980b-57135e0d8540-kube-api-access-854xf\") pod \"machine-config-server-gw4gv\" (UID: \"1946a181-5c5e-49f4-980b-57135e0d8540\") " pod="openshift-machine-config-operator/machine-config-server-gw4gv" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.426195 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.449965 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vc7vq"] Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.453261 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dncx2" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.457641 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 09:06:03 crc kubenswrapper[4556]: E0218 09:06:03.458200 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 09:06:03.958182655 +0000 UTC m=+120.975143634 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.459455 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.459620 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.465296 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-dffhx" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.475653 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.490799 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8d17a601-499d-4d9b-945a-557378ea7400-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-6ss98\" (UID: \"8d17a601-499d-4d9b-945a-557378ea7400\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6ss98" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.494479 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.499447 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-vxdmn" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.514414 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.520038 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-pzbhd" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.533484 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.541363 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-z7mx2"] Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.556608 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.559031 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wz622\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:06:03 crc kubenswrapper[4556]: E0218 09:06:03.559331 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 09:06:04.059319785 +0000 UTC m=+121.076280765 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wz622" (UID: "709f253c-93c3-4aaf-b3db-630775991aeb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.565995 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5smq4" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.572987 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Feb 18 09:06:03 crc kubenswrapper[4556]: W0218 09:06:03.589418 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb3ca6b25_cb73_47a9_867c_8d1f6f628077.slice/crio-9448f267c2b7cacaa0c71e231aea7f1df24cb04e98f1861e6f2a507c73e083f2 WatchSource:0}: Error finding container 9448f267c2b7cacaa0c71e231aea7f1df24cb04e98f1861e6f2a507c73e083f2: Status 404 returned error can't find the container with id 9448f267c2b7cacaa0c71e231aea7f1df24cb04e98f1861e6f2a507c73e083f2 Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.592491 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.612251 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.614582 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-zlp6j" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.652671 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.659351 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-677n8\" (UniqueName: \"kubernetes.io/projected/ff9016b7-a27a-4774-a7a2-498eee516d28-kube-api-access-677n8\") pod \"openshift-controller-manager-operator-756b6f6bc6-lkxrl\" (UID: \"ff9016b7-a27a-4774-a7a2-498eee516d28\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lkxrl" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.660448 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 09:06:03 crc kubenswrapper[4556]: E0218 09:06:03.660912 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 09:06:04.160893145 +0000 UTC m=+121.177854125 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.672350 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.693442 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.705187 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rlw5\" (UniqueName: \"kubernetes.io/projected/ccb1af48-7979-4f75-983e-7455ac471966-kube-api-access-6rlw5\") pod \"csi-hostpathplugin-qw4hv\" (UID: \"ccb1af48-7979-4f75-983e-7455ac471966\") " pod="hostpath-provisioner/csi-hostpathplugin-qw4hv" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.712610 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctf6x\" (UniqueName: \"kubernetes.io/projected/fd45fd0a-9812-4ff9-8bf7-91fda0bb0df3-kube-api-access-ctf6x\") pod \"migrator-59844c95c7-sxb49\" (UID: \"fd45fd0a-9812-4ff9-8bf7-91fda0bb0df3\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sxb49" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.713806 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.715027 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6ss98" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.743491 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.752944 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.764614 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-zhtqs"] Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.765191 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wz622\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.765428 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-mvvrj" Feb 18 09:06:03 crc kubenswrapper[4556]: E0218 09:06:03.765494 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 09:06:04.265482892 +0000 UTC m=+121.282443873 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wz622" (UID: "709f253c-93c3-4aaf-b3db-630775991aeb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.792861 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7fld\" (UniqueName: \"kubernetes.io/projected/6ea95a80-9834-4f76-a250-6c6394c0df87-kube-api-access-r7fld\") pod \"ingress-canary-kvbqn\" (UID: \"6ea95a80-9834-4f76-a250-6c6394c0df87\") " pod="openshift-ingress-canary/ingress-canary-kvbqn" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.796752 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.798277 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-gw4gv" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.841812 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.842909 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-vd7f7"] Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.844806 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sxb49" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.864063 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-z7mx2" event={"ID":"a737bcdf-9a16-4220-842e-75a130efb35d","Type":"ContainerStarted","Data":"9eed0626ac92748a854f7b9e9a15ac67ad713f7c6a12e95c30f5c644921eafec"} Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.865993 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 09:06:03 crc kubenswrapper[4556]: E0218 09:06:03.866272 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 09:06:04.366206105 +0000 UTC m=+121.383167085 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.866466 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wz622\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:06:03 crc kubenswrapper[4556]: E0218 09:06:03.866778 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 09:06:04.366766719 +0000 UTC m=+121.383727698 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wz622" (UID: "709f253c-93c3-4aaf-b3db-630775991aeb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.877177 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.886465 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lkxrl" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.887894 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-8gpk5" event={"ID":"f34c2f1a-5d5b-4719-82e3-d7534e16f7e6","Type":"ContainerStarted","Data":"7601dbdf99c7d9a7ad22314d5e2ee6774b6e2db5f7605c3cd3e6915f61d264dd"} Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.891427 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-dffhx"] Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.895731 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mgwg5" event={"ID":"d3cd4918-7d0a-409d-9fe3-e168ce2e80fa","Type":"ContainerStarted","Data":"78b849b684ce4468213e550e9d5dbad44568c4ab358b14b8e9db1ffe676c8dfd"} Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.896440 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mgwg5" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.909596 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-md5fm" event={"ID":"6bfd85fe-faa2-48f8-9375-3a5a15214496","Type":"ContainerStarted","Data":"8c237a79e79ddae15b97da100b158bb60285301462c54304eb9f3c51d9e55194"} Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.909689 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-md5fm" event={"ID":"6bfd85fe-faa2-48f8-9375-3a5a15214496","Type":"ContainerStarted","Data":"bfd5a09ce30c58d83f34fddd113655b92cf63ab7924d0f43da6f92ec543da6f2"} Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.909703 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-md5fm" event={"ID":"6bfd85fe-faa2-48f8-9375-3a5a15214496","Type":"ContainerStarted","Data":"4865b9488f6ed173bb9f122a06e5258e9d42252701e6aee5d4f07d4102ef69b8"} Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.915146 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mgwg5" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.921051 4556 generic.go:334] "Generic (PLEG): container finished" podID="8129ffba-fc55-45de-945f-66071d25c4eb" containerID="a1bb07bd9cd2dc66ffa42f3bf339b77e73500835df089762430b71eb78e833b6" exitCode=0 Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.921090 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxkx5" event={"ID":"8129ffba-fc55-45de-945f-66071d25c4eb","Type":"ContainerDied","Data":"a1bb07bd9cd2dc66ffa42f3bf339b77e73500835df089762430b71eb78e833b6"} Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.926572 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-fmk6k" event={"ID":"e8926df7-dcbc-480d-bded-d4941bcc0d91","Type":"ContainerStarted","Data":"f1c06ec7962dcc1e6d533a16001cb9859e0046eb604950adb5a604f6f13e7764"} Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.932391 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-vc7vq" event={"ID":"7fb132fe-af75-4b1a-828f-5b1dcd074fd6","Type":"ContainerStarted","Data":"8b09bc452f5187c3acd8beed22526cbb2d46907fea2cb19de628c721e12705ed"} Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.936714 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dncx2" event={"ID":"288bd6a4-25b6-483f-a1e7-7b07da5d63b6","Type":"ContainerStarted","Data":"19ed1c7e2c6daa157d34d88ff70bbe71c19d8e0862f4ae5139e7bab893131aec"} Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.940420 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2sdct" event={"ID":"6f9fe7cf-4384-4a95-abbb-20b487e7bc6c","Type":"ContainerStarted","Data":"92dfcf5c158636ba7d85bb08f3c643754a755c541be4e5c735fbeb96e693a266"} Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.940859 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2sdct" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.944369 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-hhxz6" event={"ID":"a303aa13-ef7e-4a65-a5d8-9763d2f0b4ea","Type":"ContainerStarted","Data":"7d4be5809bd98bdb9171f8062842621a504db9c9d81a67e4f848d3c1fc5742ec"} Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.944960 4556 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-2sdct container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.41:8443/healthz\": dial tcp 10.217.0.41:8443: connect: connection refused" start-of-body= Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.944994 4556 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2sdct" podUID="6f9fe7cf-4384-4a95-abbb-20b487e7bc6c" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.41:8443/healthz\": dial tcp 10.217.0.41:8443: connect: connection refused" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.947947 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-55hcm" event={"ID":"0c42eda9-04e2-4035-afb6-3bde005af88a","Type":"ContainerStarted","Data":"9fad56a47948bb4dbafeb1436e594797c724c3a9fe158a88f09f36d7d647bcd6"} Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.948812 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-55hcm" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.953697 4556 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-55hcm container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.10:5443/healthz\": dial tcp 10.217.0.10:5443: connect: connection refused" start-of-body= Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.953741 4556 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-55hcm" podUID="0c42eda9-04e2-4035-afb6-3bde005af88a" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.10:5443/healthz\": dial tcp 10.217.0.10:5443: connect: connection refused" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.960399 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lgh4n" event={"ID":"d1d69c14-a464-4f22-88ca-4e0cfe97111a","Type":"ContainerStarted","Data":"bece98172bdfa4b409f8c4f3063dd1b989541983c98f759120f0ff67b18d3cb5"} Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.960425 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lgh4n" event={"ID":"d1d69c14-a464-4f22-88ca-4e0cfe97111a","Type":"ContainerStarted","Data":"c1edf6c8188dd906a018428a4441af8fecf3ee2ef20274897d7c23ffeeced458"} Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.967584 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 09:06:03 crc kubenswrapper[4556]: E0218 09:06:03.967691 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 09:06:04.467670861 +0000 UTC m=+121.484631841 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.968479 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wz622\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:06:03 crc kubenswrapper[4556]: E0218 09:06:03.971272 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 09:06:04.471256549 +0000 UTC m=+121.488217528 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wz622" (UID: "709f253c-93c3-4aaf-b3db-630775991aeb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.977704 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j5fdh" event={"ID":"b3ca6b25-cb73-47a9-867c-8d1f6f628077","Type":"ContainerStarted","Data":"9448f267c2b7cacaa0c71e231aea7f1df24cb04e98f1861e6f2a507c73e083f2"} Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.981689 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r44z2" event={"ID":"c44b5f9b-46a0-4ca6-b6ac-456bcae851fa","Type":"ContainerStarted","Data":"34718e36f2711fbe768f4060b3db057de45f917eb0282fc193e7843ac4e0778c"} Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.981895 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r44z2" Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.985014 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4dfqf" event={"ID":"716bc347-cbf5-452c-a7c8-3eeff4e70870","Type":"ContainerStarted","Data":"0aa24c98018c1c99a89fdedbe196fd04eed418b88e4bacfa980d7c2360fcb35f"} Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.985047 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4dfqf" event={"ID":"716bc347-cbf5-452c-a7c8-3eeff4e70870","Type":"ContainerStarted","Data":"cd9fd4e13cd8b77c7a7a090164dd3900da674a250e28dda04081415ff6c3f250"} Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.990231 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-lsngn" event={"ID":"b1111d02-c210-4787-8dda-b118339cfc79","Type":"ContainerStarted","Data":"d020a202f26871eaf94c4dfb179397d8474445cffc77c27bbb651ac6e85ac873"} Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.993511 4556 generic.go:334] "Generic (PLEG): container finished" podID="9a47b8c4-1288-4b4b-9d82-4c6c22cc65b3" containerID="46f75bcc2e1421cebe8f3b3405a738e821699b16e5d6013bb41fb097b1924c89" exitCode=0 Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.993563 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-lzjw5" event={"ID":"9a47b8c4-1288-4b4b-9d82-4c6c22cc65b3","Type":"ContainerDied","Data":"46f75bcc2e1421cebe8f3b3405a738e821699b16e5d6013bb41fb097b1924c89"} Feb 18 09:06:03 crc kubenswrapper[4556]: I0218 09:06:03.996384 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-lv2k7" event={"ID":"75ba12a5-6494-4d1d-a2d8-b2674a55b3ca","Type":"ContainerStarted","Data":"bfa798243b0305435c714d8795a01f8e68e42a2f057018faae070d41bfdb28c8"} Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.012468 4556 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.016241 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-5wlpw" event={"ID":"944ea0f1-a1d8-4912-b117-86e73f15b0fe","Type":"ContainerStarted","Data":"c7188cdb9b1bfde950e11d61955f2c923e7e0556d9e961f9d02bdc319b09744c"} Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.022736 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-qw4hv" Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.054990 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.063169 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-kvbqn" Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.065805 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n9zlc" event={"ID":"585ea2bd-474c-41e6-aa73-e7a62d02eb1c","Type":"ContainerStarted","Data":"e4d5f1766d75038a24769fe8868c22d3cd63cd55c779a7a45a0ff67f45ea2603"} Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.069564 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 09:06:04 crc kubenswrapper[4556]: E0218 09:06:04.072300 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 09:06:04.572285946 +0000 UTC m=+121.589246926 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.073001 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n2cfq" event={"ID":"679a536c-c470-4231-b3aa-bf7004cbcb4f","Type":"ContainerStarted","Data":"98375b74461594c60736a1314d3038efc1026eaf4900d481fdf2458551c04fac"} Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.082030 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29523420-qb78x" event={"ID":"0aa2e566-fd95-498b-8df5-b63afb290572","Type":"ContainerStarted","Data":"f20e52a9326969859ebb286bd8c1bdaaf60692a63dec1c8d4018363660421633"} Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.092487 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-6d57b" event={"ID":"a7a96d74-e090-45ac-a088-46cda659a8e7","Type":"ContainerStarted","Data":"c8a748710564e2e441090789ec7ab62773aa380fc80336ec950848c6f2569088"} Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.097731 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-4vkc6" event={"ID":"a457911e-1ca9-4f57-8286-b13f51549de7","Type":"ContainerStarted","Data":"884bd7cb978691e010cef044b005284e23c07b05866c8ff2fac4877a3830cc5b"} Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.099305 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5chhg" event={"ID":"5bc2daed-4513-47bb-ad46-52cc8c14f02b","Type":"ContainerStarted","Data":"dd29f8a290c543d93dda9cb2127556fd2debec94059d6eec21c9f10f2f0ff402"} Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.106979 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-26rzb" Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.171238 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wz622\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:06:04 crc kubenswrapper[4556]: E0218 09:06:04.173167 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 09:06:04.673139143 +0000 UTC m=+121.690100123 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wz622" (UID: "709f253c-93c3-4aaf-b3db-630775991aeb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.248656 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-pzbhd"] Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.274147 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 09:06:04 crc kubenswrapper[4556]: E0218 09:06:04.274600 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 09:06:04.774586386 +0000 UTC m=+121.791547365 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.366057 4556 patch_prober.go:28] interesting pod/router-default-5444994796-j9ccg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 18 09:06:04 crc kubenswrapper[4556]: [-]has-synced failed: reason withheld Feb 18 09:06:04 crc kubenswrapper[4556]: [+]process-running ok Feb 18 09:06:04 crc kubenswrapper[4556]: healthz check failed Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.366268 4556 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-j9ccg" podUID="9cf0e7d1-1645-47aa-add2-3a76bfc624a7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.366509 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-brf28"] Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.371337 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-brf28" Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.380387 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-brf28"] Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.381512 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-vxdmn"] Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.388601 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wz622\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:06:04 crc kubenswrapper[4556]: E0218 09:06:04.389501 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 09:06:04.889488672 +0000 UTC m=+121.906449653 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wz622" (UID: "709f253c-93c3-4aaf-b3db-630775991aeb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.396244 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-5smq4"] Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.397928 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 18 09:06:04 crc kubenswrapper[4556]: W0218 09:06:04.449166 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9db33ada_017a_4281_85c1_0c1530bc3bbc.slice/crio-8aa857eb2ce6837468a5a1ffec5a01a532f05112cc366ffa17fda96412a25bd1 WatchSource:0}: Error finding container 8aa857eb2ce6837468a5a1ffec5a01a532f05112cc366ffa17fda96412a25bd1: Status 404 returned error can't find the container with id 8aa857eb2ce6837468a5a1ffec5a01a532f05112cc366ffa17fda96412a25bd1 Feb 18 09:06:04 crc kubenswrapper[4556]: W0218 09:06:04.473241 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1aa03714_6c33_42a8_a427_04a7ce839f21.slice/crio-d74bfa83adc732435f68733647a8f1b717041892c298b5415a7e102ca9e1ded3 WatchSource:0}: Error finding container d74bfa83adc732435f68733647a8f1b717041892c298b5415a7e102ca9e1ded3: Status 404 returned error can't find the container with id d74bfa83adc732435f68733647a8f1b717041892c298b5415a7e102ca9e1ded3 Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.489517 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.489751 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eac84811-d54a-47c2-ae5e-77f3a531fc4f-catalog-content\") pod \"certified-operators-brf28\" (UID: \"eac84811-d54a-47c2-ae5e-77f3a531fc4f\") " pod="openshift-marketplace/certified-operators-brf28" Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.489779 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eac84811-d54a-47c2-ae5e-77f3a531fc4f-utilities\") pod \"certified-operators-brf28\" (UID: \"eac84811-d54a-47c2-ae5e-77f3a531fc4f\") " pod="openshift-marketplace/certified-operators-brf28" Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.489839 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmc75\" (UniqueName: \"kubernetes.io/projected/eac84811-d54a-47c2-ae5e-77f3a531fc4f-kube-api-access-jmc75\") pod \"certified-operators-brf28\" (UID: \"eac84811-d54a-47c2-ae5e-77f3a531fc4f\") " pod="openshift-marketplace/certified-operators-brf28" Feb 18 09:06:04 crc kubenswrapper[4556]: E0218 09:06:04.490237 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 09:06:04.990215952 +0000 UTC m=+122.007176932 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.524349 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-mvvrj"] Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.560064 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-sd4dx"] Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.560918 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sd4dx" Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.570391 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-sd4dx"] Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.590617 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wz622\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:06:04 crc kubenswrapper[4556]: E0218 09:06:04.590870 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 09:06:05.090857231 +0000 UTC m=+122.107818211 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wz622" (UID: "709f253c-93c3-4aaf-b3db-630775991aeb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.590981 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eac84811-d54a-47c2-ae5e-77f3a531fc4f-catalog-content\") pod \"certified-operators-brf28\" (UID: \"eac84811-d54a-47c2-ae5e-77f3a531fc4f\") " pod="openshift-marketplace/certified-operators-brf28" Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.591005 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eac84811-d54a-47c2-ae5e-77f3a531fc4f-utilities\") pod \"certified-operators-brf28\" (UID: \"eac84811-d54a-47c2-ae5e-77f3a531fc4f\") " pod="openshift-marketplace/certified-operators-brf28" Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.591062 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmc75\" (UniqueName: \"kubernetes.io/projected/eac84811-d54a-47c2-ae5e-77f3a531fc4f-kube-api-access-jmc75\") pod \"certified-operators-brf28\" (UID: \"eac84811-d54a-47c2-ae5e-77f3a531fc4f\") " pod="openshift-marketplace/certified-operators-brf28" Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.591520 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eac84811-d54a-47c2-ae5e-77f3a531fc4f-catalog-content\") pod \"certified-operators-brf28\" (UID: \"eac84811-d54a-47c2-ae5e-77f3a531fc4f\") " pod="openshift-marketplace/certified-operators-brf28" Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.591630 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eac84811-d54a-47c2-ae5e-77f3a531fc4f-utilities\") pod \"certified-operators-brf28\" (UID: \"eac84811-d54a-47c2-ae5e-77f3a531fc4f\") " pod="openshift-marketplace/certified-operators-brf28" Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.592765 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.668790 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmc75\" (UniqueName: \"kubernetes.io/projected/eac84811-d54a-47c2-ae5e-77f3a531fc4f-kube-api-access-jmc75\") pod \"certified-operators-brf28\" (UID: \"eac84811-d54a-47c2-ae5e-77f3a531fc4f\") " pod="openshift-marketplace/certified-operators-brf28" Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.692503 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 09:06:04 crc kubenswrapper[4556]: E0218 09:06:04.692675 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 09:06:05.192657116 +0000 UTC m=+122.209618095 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.692840 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wz622\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.692934 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rcdq4\" (UniqueName: \"kubernetes.io/projected/90477a11-cc7c-48cf-a210-9d1d802360ce-kube-api-access-rcdq4\") pod \"community-operators-sd4dx\" (UID: \"90477a11-cc7c-48cf-a210-9d1d802360ce\") " pod="openshift-marketplace/community-operators-sd4dx" Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.692955 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90477a11-cc7c-48cf-a210-9d1d802360ce-catalog-content\") pod \"community-operators-sd4dx\" (UID: \"90477a11-cc7c-48cf-a210-9d1d802360ce\") " pod="openshift-marketplace/community-operators-sd4dx" Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.693017 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90477a11-cc7c-48cf-a210-9d1d802360ce-utilities\") pod \"community-operators-sd4dx\" (UID: \"90477a11-cc7c-48cf-a210-9d1d802360ce\") " pod="openshift-marketplace/community-operators-sd4dx" Feb 18 09:06:04 crc kubenswrapper[4556]: E0218 09:06:04.693333 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 09:06:05.193321874 +0000 UTC m=+122.210282854 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wz622" (UID: "709f253c-93c3-4aaf-b3db-630775991aeb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.726666 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-sxb49"] Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.728505 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6ss98"] Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.737436 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-zlp6j"] Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.757510 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-ss8sl"] Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.758348 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ss8sl" Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.771891 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ss8sl"] Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.799756 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.800103 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90477a11-cc7c-48cf-a210-9d1d802360ce-catalog-content\") pod \"community-operators-sd4dx\" (UID: \"90477a11-cc7c-48cf-a210-9d1d802360ce\") " pod="openshift-marketplace/community-operators-sd4dx" Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.800209 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7875d007-bdbc-428a-b291-283e8a1cdad0-catalog-content\") pod \"certified-operators-ss8sl\" (UID: \"7875d007-bdbc-428a-b291-283e8a1cdad0\") " pod="openshift-marketplace/certified-operators-ss8sl" Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.800262 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90477a11-cc7c-48cf-a210-9d1d802360ce-utilities\") pod \"community-operators-sd4dx\" (UID: \"90477a11-cc7c-48cf-a210-9d1d802360ce\") " pod="openshift-marketplace/community-operators-sd4dx" Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.800314 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktcxc\" (UniqueName: \"kubernetes.io/projected/7875d007-bdbc-428a-b291-283e8a1cdad0-kube-api-access-ktcxc\") pod \"certified-operators-ss8sl\" (UID: \"7875d007-bdbc-428a-b291-283e8a1cdad0\") " pod="openshift-marketplace/certified-operators-ss8sl" Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.800374 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7875d007-bdbc-428a-b291-283e8a1cdad0-utilities\") pod \"certified-operators-ss8sl\" (UID: \"7875d007-bdbc-428a-b291-283e8a1cdad0\") " pod="openshift-marketplace/certified-operators-ss8sl" Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.800418 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rcdq4\" (UniqueName: \"kubernetes.io/projected/90477a11-cc7c-48cf-a210-9d1d802360ce-kube-api-access-rcdq4\") pod \"community-operators-sd4dx\" (UID: \"90477a11-cc7c-48cf-a210-9d1d802360ce\") " pod="openshift-marketplace/community-operators-sd4dx" Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.801568 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90477a11-cc7c-48cf-a210-9d1d802360ce-catalog-content\") pod \"community-operators-sd4dx\" (UID: \"90477a11-cc7c-48cf-a210-9d1d802360ce\") " pod="openshift-marketplace/community-operators-sd4dx" Feb 18 09:06:04 crc kubenswrapper[4556]: E0218 09:06:04.801676 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 09:06:05.301653976 +0000 UTC m=+122.318614955 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.801804 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90477a11-cc7c-48cf-a210-9d1d802360ce-utilities\") pod \"community-operators-sd4dx\" (UID: \"90477a11-cc7c-48cf-a210-9d1d802360ce\") " pod="openshift-marketplace/community-operators-sd4dx" Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.810768 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-qw4hv"] Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.827018 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lkxrl"] Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.831672 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rcdq4\" (UniqueName: \"kubernetes.io/projected/90477a11-cc7c-48cf-a210-9d1d802360ce-kube-api-access-rcdq4\") pod \"community-operators-sd4dx\" (UID: \"90477a11-cc7c-48cf-a210-9d1d802360ce\") " pod="openshift-marketplace/community-operators-sd4dx" Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.838337 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-brf28" Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.854656 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-kvbqn"] Feb 18 09:06:04 crc kubenswrapper[4556]: W0218 09:06:04.885617 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podccb1af48_7979_4f75_983e_7455ac471966.slice/crio-49f092baecefcf56a2b0698cd897ffbbc51061189d46b128c5be9b46ec028bf1 WatchSource:0}: Error finding container 49f092baecefcf56a2b0698cd897ffbbc51061189d46b128c5be9b46ec028bf1: Status 404 returned error can't find the container with id 49f092baecefcf56a2b0698cd897ffbbc51061189d46b128c5be9b46ec028bf1 Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.912889 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7875d007-bdbc-428a-b291-283e8a1cdad0-utilities\") pod \"certified-operators-ss8sl\" (UID: \"7875d007-bdbc-428a-b291-283e8a1cdad0\") " pod="openshift-marketplace/certified-operators-ss8sl" Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.913909 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7875d007-bdbc-428a-b291-283e8a1cdad0-catalog-content\") pod \"certified-operators-ss8sl\" (UID: \"7875d007-bdbc-428a-b291-283e8a1cdad0\") " pod="openshift-marketplace/certified-operators-ss8sl" Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.914010 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wz622\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.914034 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktcxc\" (UniqueName: \"kubernetes.io/projected/7875d007-bdbc-428a-b291-283e8a1cdad0-kube-api-access-ktcxc\") pod \"certified-operators-ss8sl\" (UID: \"7875d007-bdbc-428a-b291-283e8a1cdad0\") " pod="openshift-marketplace/certified-operators-ss8sl" Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.914413 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7875d007-bdbc-428a-b291-283e8a1cdad0-catalog-content\") pod \"certified-operators-ss8sl\" (UID: \"7875d007-bdbc-428a-b291-283e8a1cdad0\") " pod="openshift-marketplace/certified-operators-ss8sl" Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.927441 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7875d007-bdbc-428a-b291-283e8a1cdad0-utilities\") pod \"certified-operators-ss8sl\" (UID: \"7875d007-bdbc-428a-b291-283e8a1cdad0\") " pod="openshift-marketplace/certified-operators-ss8sl" Feb 18 09:06:04 crc kubenswrapper[4556]: E0218 09:06:04.927673 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 09:06:05.427662452 +0000 UTC m=+122.444623432 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wz622" (UID: "709f253c-93c3-4aaf-b3db-630775991aeb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.928002 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sd4dx" Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.965693 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktcxc\" (UniqueName: \"kubernetes.io/projected/7875d007-bdbc-428a-b291-283e8a1cdad0-kube-api-access-ktcxc\") pod \"certified-operators-ss8sl\" (UID: \"7875d007-bdbc-428a-b291-283e8a1cdad0\") " pod="openshift-marketplace/certified-operators-ss8sl" Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.968651 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-wqmhk"] Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.979068 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wqmhk" Feb 18 09:06:04 crc kubenswrapper[4556]: I0218 09:06:04.997459 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wqmhk"] Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.014509 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.014722 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25cebcb5-2e64-4390-9d61-f06607eee280-catalog-content\") pod \"community-operators-wqmhk\" (UID: \"25cebcb5-2e64-4390-9d61-f06607eee280\") " pod="openshift-marketplace/community-operators-wqmhk" Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.014834 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25cebcb5-2e64-4390-9d61-f06607eee280-utilities\") pod \"community-operators-wqmhk\" (UID: \"25cebcb5-2e64-4390-9d61-f06607eee280\") " pod="openshift-marketplace/community-operators-wqmhk" Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.014883 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwpbb\" (UniqueName: \"kubernetes.io/projected/25cebcb5-2e64-4390-9d61-f06607eee280-kube-api-access-kwpbb\") pod \"community-operators-wqmhk\" (UID: \"25cebcb5-2e64-4390-9d61-f06607eee280\") " pod="openshift-marketplace/community-operators-wqmhk" Feb 18 09:06:05 crc kubenswrapper[4556]: E0218 09:06:05.016577 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 09:06:05.516559531 +0000 UTC m=+122.533520511 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.114164 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ss8sl" Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.115718 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25cebcb5-2e64-4390-9d61-f06607eee280-utilities\") pod \"community-operators-wqmhk\" (UID: \"25cebcb5-2e64-4390-9d61-f06607eee280\") " pod="openshift-marketplace/community-operators-wqmhk" Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.115772 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwpbb\" (UniqueName: \"kubernetes.io/projected/25cebcb5-2e64-4390-9d61-f06607eee280-kube-api-access-kwpbb\") pod \"community-operators-wqmhk\" (UID: \"25cebcb5-2e64-4390-9d61-f06607eee280\") " pod="openshift-marketplace/community-operators-wqmhk" Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.115797 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25cebcb5-2e64-4390-9d61-f06607eee280-catalog-content\") pod \"community-operators-wqmhk\" (UID: \"25cebcb5-2e64-4390-9d61-f06607eee280\") " pod="openshift-marketplace/community-operators-wqmhk" Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.115817 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wz622\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:06:05 crc kubenswrapper[4556]: E0218 09:06:05.116170 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 09:06:05.61614213 +0000 UTC m=+122.633103110 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wz622" (UID: "709f253c-93c3-4aaf-b3db-630775991aeb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.116507 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25cebcb5-2e64-4390-9d61-f06607eee280-utilities\") pod \"community-operators-wqmhk\" (UID: \"25cebcb5-2e64-4390-9d61-f06607eee280\") " pod="openshift-marketplace/community-operators-wqmhk" Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.116948 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25cebcb5-2e64-4390-9d61-f06607eee280-catalog-content\") pod \"community-operators-wqmhk\" (UID: \"25cebcb5-2e64-4390-9d61-f06607eee280\") " pod="openshift-marketplace/community-operators-wqmhk" Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.177024 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwpbb\" (UniqueName: \"kubernetes.io/projected/25cebcb5-2e64-4390-9d61-f06607eee280-kube-api-access-kwpbb\") pod \"community-operators-wqmhk\" (UID: \"25cebcb5-2e64-4390-9d61-f06607eee280\") " pod="openshift-marketplace/community-operators-wqmhk" Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.216713 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 09:06:05 crc kubenswrapper[4556]: E0218 09:06:05.217245 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 09:06:05.717227223 +0000 UTC m=+122.734188203 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.246072 4556 patch_prober.go:28] interesting pod/router-default-5444994796-j9ccg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 18 09:06:05 crc kubenswrapper[4556]: [-]has-synced failed: reason withheld Feb 18 09:06:05 crc kubenswrapper[4556]: [+]process-running ok Feb 18 09:06:05 crc kubenswrapper[4556]: healthz check failed Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.246299 4556 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-j9ccg" podUID="9cf0e7d1-1645-47aa-add2-3a76bfc624a7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.270557 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-zhtqs" event={"ID":"b5f35d95-f6b7-4068-8036-dffcbf955272","Type":"ContainerStarted","Data":"a12dc0721a22274ed2a665149e5adc431232ae7cf5cc7e368c9ace66b5c11ff1"} Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.270586 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-zhtqs" event={"ID":"b5f35d95-f6b7-4068-8036-dffcbf955272","Type":"ContainerStarted","Data":"1db18931cd63ee1db4666f3bafa16fbfb941d7fc5458cebc9ef32ce9772c09ba"} Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.270596 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-zhtqs" event={"ID":"b5f35d95-f6b7-4068-8036-dffcbf955272","Type":"ContainerStarted","Data":"4df8834e2d4e932aadd8012be384893070175101c469e41a02375327f826f3a5"} Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.317779 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wz622\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:06:05 crc kubenswrapper[4556]: E0218 09:06:05.318506 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 09:06:05.818493775 +0000 UTC m=+122.835454755 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wz622" (UID: "709f253c-93c3-4aaf-b3db-630775991aeb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.357998 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wqmhk" Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.413139 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r44z2" podStartSLOduration=99.413119429 podStartE2EDuration="1m39.413119429s" podCreationTimestamp="2026-02-18 09:04:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:06:05.409029224 +0000 UTC m=+122.425990204" watchObservedRunningTime="2026-02-18 09:06:05.413119429 +0000 UTC m=+122.430080409" Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.419273 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 09:06:05 crc kubenswrapper[4556]: E0218 09:06:05.419721 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 09:06:05.919703492 +0000 UTC m=+122.936664471 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.427421 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-lv2k7" event={"ID":"75ba12a5-6494-4d1d-a2d8-b2674a55b3ca","Type":"ContainerStarted","Data":"0980cbebcf26097e292835a8538fb4362c019768011bfc9a9cdf0cd364a42561"} Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.428252 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-lv2k7" Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.434821 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n9zlc" event={"ID":"585ea2bd-474c-41e6-aa73-e7a62d02eb1c","Type":"ContainerStarted","Data":"e2dbf9ec89fd990d1fc70612d5d419f97b1c2ef6ef3a19439369077e6b718494"} Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.435926 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-gw4gv" event={"ID":"1946a181-5c5e-49f4-980b-57135e0d8540","Type":"ContainerStarted","Data":"2f5171df2a3a744732336cd7c556d59285a1af855407954bb4c78c146e490107"} Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.435959 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-gw4gv" event={"ID":"1946a181-5c5e-49f4-980b-57135e0d8540","Type":"ContainerStarted","Data":"517765f51b770146484298dd96623faa6754f77d32fbd0140ee1f051abe92b16"} Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.446438 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5chhg" podStartSLOduration=99.44642773 podStartE2EDuration="1m39.44642773s" podCreationTimestamp="2026-02-18 09:04:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:06:05.442748947 +0000 UTC m=+122.459709927" watchObservedRunningTime="2026-02-18 09:06:05.44642773 +0000 UTC m=+122.463388710" Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.471804 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-brf28"] Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.481896 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n9zlc" podStartSLOduration=99.481878397 podStartE2EDuration="1m39.481878397s" podCreationTimestamp="2026-02-18 09:04:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:06:05.480420357 +0000 UTC m=+122.497381338" watchObservedRunningTime="2026-02-18 09:06:05.481878397 +0000 UTC m=+122.498839367" Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.487915 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-z7mx2" event={"ID":"a737bcdf-9a16-4220-842e-75a130efb35d","Type":"ContainerStarted","Data":"4f70366fe4db8c8eb43d94bb8a72fffee6678fc66081902b3bd70c53f250dd7e"} Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.492080 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sxb49" event={"ID":"fd45fd0a-9812-4ff9-8bf7-91fda0bb0df3","Type":"ContainerStarted","Data":"270209e34ec2e1fce39f8ca6570cabde6f84cf9f7b347de3207c2569d6cde418"} Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.512863 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dncx2" event={"ID":"288bd6a4-25b6-483f-a1e7-7b07da5d63b6","Type":"ContainerStarted","Data":"c2555d8fd172c8918db2fbd8804b54ff37a11cb467b9fc5aae30deb50e5ec876"} Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.512913 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dncx2" event={"ID":"288bd6a4-25b6-483f-a1e7-7b07da5d63b6","Type":"ContainerStarted","Data":"8ebfa676a764ca2c53b68e1d7b6054c0454d18aed5a08660d9eb1c7abbed7506"} Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.520484 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wz622\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:06:05 crc kubenswrapper[4556]: E0218 09:06:05.522169 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 09:06:06.022135746 +0000 UTC m=+123.039096726 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wz622" (UID: "709f253c-93c3-4aaf-b3db-630775991aeb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.529239 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-j9ccg" podStartSLOduration=99.529227843 podStartE2EDuration="1m39.529227843s" podCreationTimestamp="2026-02-18 09:04:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:06:05.527396351 +0000 UTC m=+122.544357331" watchObservedRunningTime="2026-02-18 09:06:05.529227843 +0000 UTC m=+122.546188823" Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.530360 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-mvvrj" event={"ID":"87338513-3a60-432f-8edf-02695687e1e2","Type":"ContainerStarted","Data":"b2dfbc7757f255d8f0ecd4f0dce1f100980f3188ffc840be64cff2669258a00d"} Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.531032 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-mvvrj" Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.549574 4556 patch_prober.go:28] interesting pod/downloads-7954f5f757-mvvrj container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body= Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.549612 4556 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-mvvrj" podUID="87338513-3a60-432f-8edf-02695687e1e2" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.557563 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5smq4" event={"ID":"1aa03714-6c33-42a8-a427-04a7ce839f21","Type":"ContainerStarted","Data":"763dbf641d38415984cccf51f33b11a5cca686a5a37aceb0135b099bc364e26e"} Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.557612 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5smq4" event={"ID":"1aa03714-6c33-42a8-a427-04a7ce839f21","Type":"ContainerStarted","Data":"d74bfa83adc732435f68733647a8f1b717041892c298b5415a7e102ca9e1ded3"} Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.566764 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-5wlpw" podStartSLOduration=99.566752617 podStartE2EDuration="1m39.566752617s" podCreationTimestamp="2026-02-18 09:04:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:06:05.566470356 +0000 UTC m=+122.583431336" watchObservedRunningTime="2026-02-18 09:06:05.566752617 +0000 UTC m=+122.583713596" Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.614305 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-zhtqs" podStartSLOduration=99.61428705 podStartE2EDuration="1m39.61428705s" podCreationTimestamp="2026-02-18 09:04:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:06:05.606012371 +0000 UTC m=+122.622973352" watchObservedRunningTime="2026-02-18 09:06:05.61428705 +0000 UTC m=+122.631248031" Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.616955 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-6d57b" event={"ID":"a7a96d74-e090-45ac-a088-46cda659a8e7","Type":"ContainerStarted","Data":"828956ecf5c1c54afce9526378b27531037e28131fe879c9c16b61277969c316"} Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.617001 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-6d57b" event={"ID":"a7a96d74-e090-45ac-a088-46cda659a8e7","Type":"ContainerStarted","Data":"662c6b35275ec4c91ec058acf5829f7eac3bdbdf3c5c5a93d7df1e5eb69e7b2d"} Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.632508 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 09:06:05 crc kubenswrapper[4556]: E0218 09:06:05.638510 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 09:06:06.138489139 +0000 UTC m=+123.155450119 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.656780 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-vc7vq" event={"ID":"7fb132fe-af75-4b1a-828f-5b1dcd074fd6","Type":"ContainerStarted","Data":"e01ec1bf5ac6396edac9a10179c99bb15ddd5777cc18550b56abeffcd1e6e39d"} Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.658724 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-vc7vq" Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.667511 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2sdct" podStartSLOduration=99.66748143 podStartE2EDuration="1m39.66748143s" podCreationTimestamp="2026-02-18 09:04:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:06:05.65769015 +0000 UTC m=+122.674651131" watchObservedRunningTime="2026-02-18 09:06:05.66748143 +0000 UTC m=+122.684442410" Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.686784 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29523420-qb78x" event={"ID":"0aa2e566-fd95-498b-8df5-b63afb290572","Type":"ContainerStarted","Data":"c2201de00ba0f9c778a8b2bbece184c31365c38bd6019f97cef435b6315fc2ab"} Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.692116 4556 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-vc7vq container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.32:8080/healthz\": dial tcp 10.217.0.32:8080: connect: connection refused" start-of-body= Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.692300 4556 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-vc7vq" podUID="7fb132fe-af75-4b1a-828f-5b1dcd074fd6" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.32:8080/healthz\": dial tcp 10.217.0.32:8080: connect: connection refused" Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.741510 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wz622\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:06:05 crc kubenswrapper[4556]: E0218 09:06:05.742513 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 09:06:06.242498156 +0000 UTC m=+123.259459136 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wz622" (UID: "709f253c-93c3-4aaf-b3db-630775991aeb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.742833 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-qw4hv" event={"ID":"ccb1af48-7979-4f75-983e-7455ac471966","Type":"ContainerStarted","Data":"49f092baecefcf56a2b0698cd897ffbbc51061189d46b128c5be9b46ec028bf1"} Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.775601 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-8gpk5" event={"ID":"f34c2f1a-5d5b-4719-82e3-d7534e16f7e6","Type":"ContainerStarted","Data":"b101476ee7791820a00adaa4bd676ed0412c9e2d157dee254b414bd6c62e163c"} Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.786624 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-lv2k7" Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.820786 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lgh4n" podStartSLOduration=99.820769528 podStartE2EDuration="1m39.820769528s" podCreationTimestamp="2026-02-18 09:04:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:06:05.820475837 +0000 UTC m=+122.837436806" watchObservedRunningTime="2026-02-18 09:06:05.820769528 +0000 UTC m=+122.837730508" Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.823627 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-55hcm" event={"ID":"0c42eda9-04e2-4035-afb6-3bde005af88a","Type":"ContainerStarted","Data":"ed66935f507932341c2d91cdd03abc55b09adaf255bcf8020be7651279754ffe"} Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.842575 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mgwg5" podStartSLOduration=99.842559604 podStartE2EDuration="1m39.842559604s" podCreationTimestamp="2026-02-18 09:04:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:06:05.777545675 +0000 UTC m=+122.794506655" watchObservedRunningTime="2026-02-18 09:06:05.842559604 +0000 UTC m=+122.859520585" Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.850199 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 09:06:05 crc kubenswrapper[4556]: E0218 09:06:05.850575 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 09:06:06.350554929 +0000 UTC m=+123.367515909 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.851674 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wz622\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:06:05 crc kubenswrapper[4556]: E0218 09:06:05.855583 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 09:06:06.355554303 +0000 UTC m=+123.372515283 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wz622" (UID: "709f253c-93c3-4aaf-b3db-630775991aeb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.865046 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lkxrl" event={"ID":"ff9016b7-a27a-4774-a7a2-498eee516d28","Type":"ContainerStarted","Data":"0a560a0e87f05a0d1c8a3df58bd3827a3c4d34ad6d551b0e23e70ad3b2fab986"} Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.866890 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-pzbhd" event={"ID":"c567dda7-a353-4ff5-81fa-487517a51971","Type":"ContainerStarted","Data":"03b5d7dca8ddbc75b9b9ff54cb5f2cb110e9e490c1017fa0a32ab4b68c27d96a"} Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.866913 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-pzbhd" event={"ID":"c567dda7-a353-4ff5-81fa-487517a51971","Type":"ContainerStarted","Data":"b579f345620e52d6559cf03ed4a684d5bcd4278671aa956d4f007dae9fcac26a"} Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.868548 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-vd7f7" event={"ID":"9023a93b-33ee-447e-8278-7b4d752d7a53","Type":"ContainerStarted","Data":"786f093e004a61c1e23c5430334210f564207c4356aeb3a685a705740c5d746e"} Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.868570 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-vd7f7" event={"ID":"9023a93b-33ee-447e-8278-7b4d752d7a53","Type":"ContainerStarted","Data":"271e02cd08d44d1fd7f6f684bfc79ff7aea467fdded5ca54f6586a450d7f6e0b"} Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.869190 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-vd7f7" Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.872679 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-vxdmn" event={"ID":"9db33ada-017a-4281-85c1-0c1530bc3bbc","Type":"ContainerStarted","Data":"8aa857eb2ce6837468a5a1ffec5a01a532f05112cc366ffa17fda96412a25bd1"} Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.875474 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-vxdmn" Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.887431 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-sd4dx"] Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.893772 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6ss98" event={"ID":"8d17a601-499d-4d9b-945a-557378ea7400","Type":"ContainerStarted","Data":"78b4427793da3b27d72c895b10aa110420e247a12520b0b3717be2a5a3ac6159"} Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.894665 4556 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-vxdmn container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.29:6443/healthz\": dial tcp 10.217.0.29:6443: connect: connection refused" start-of-body= Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.894704 4556 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-vxdmn" podUID="9db33ada-017a-4281-85c1-0c1530bc3bbc" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.29:6443/healthz\": dial tcp 10.217.0.29:6443: connect: connection refused" Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.906218 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ss8sl"] Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.906984 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-md5fm" podStartSLOduration=99.906971553 podStartE2EDuration="1m39.906971553s" podCreationTimestamp="2026-02-18 09:04:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:06:05.887873215 +0000 UTC m=+122.904834195" watchObservedRunningTime="2026-02-18 09:06:05.906971553 +0000 UTC m=+122.923932534" Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.909114 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-vd7f7" Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.921369 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-fmk6k" event={"ID":"e8926df7-dcbc-480d-bded-d4941bcc0d91","Type":"ContainerStarted","Data":"43b71859767725d32b8dd96bf8b1bf83108dd785136a25a6da5a3f444f74a85e"} Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.952948 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 09:06:05 crc kubenswrapper[4556]: E0218 09:06:05.954984 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 09:06:06.454956784 +0000 UTC m=+123.471917764 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:05 crc kubenswrapper[4556]: W0218 09:06:05.957854 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7875d007_bdbc_428a_b291_283e8a1cdad0.slice/crio-2fb3b6373c0c3b961335f85fd9735a8bb14ff7f1fe1eff214be4098697d47785 WatchSource:0}: Error finding container 2fb3b6373c0c3b961335f85fd9735a8bb14ff7f1fe1eff214be4098697d47785: Status 404 returned error can't find the container with id 2fb3b6373c0c3b961335f85fd9735a8bb14ff7f1fe1eff214be4098697d47785 Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.967990 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-4vkc6" event={"ID":"a457911e-1ca9-4f57-8286-b13f51549de7","Type":"ContainerStarted","Data":"5d4b705a0fae987c752c7e14ec1a443b04fda99987b469e49d8b43fe69748743"} Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.968027 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-4vkc6" event={"ID":"a457911e-1ca9-4f57-8286-b13f51549de7","Type":"ContainerStarted","Data":"8f718267bdd8e0dd579297782213009bedc4953c9bef6251ffaa6bfc84f78e38"} Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.976231 4556 generic.go:334] "Generic (PLEG): container finished" podID="b1111d02-c210-4787-8dda-b118339cfc79" containerID="220e38e245a9354172b99e9dda0165d2cddd33700b92b7001f3ebff54cf686fb" exitCode=0 Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.976287 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-lsngn" event={"ID":"b1111d02-c210-4787-8dda-b118339cfc79","Type":"ContainerStarted","Data":"07174c7a7f8058c35b93f16c1614ff5500e983eb22003f585dfb5e05990c3e43"} Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.976306 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-lsngn" event={"ID":"b1111d02-c210-4787-8dda-b118339cfc79","Type":"ContainerDied","Data":"220e38e245a9354172b99e9dda0165d2cddd33700b92b7001f3ebff54cf686fb"} Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.980984 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n2cfq" podStartSLOduration=99.980964235 podStartE2EDuration="1m39.980964235s" podCreationTimestamp="2026-02-18 09:04:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:06:05.978804527 +0000 UTC m=+122.995765507" watchObservedRunningTime="2026-02-18 09:06:05.980964235 +0000 UTC m=+122.997925215" Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.992453 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-dffhx" event={"ID":"ac316465-87df-4f89-99d3-7633309d3158","Type":"ContainerStarted","Data":"c0e3d3109958ea2d5b42d4f13726b6b831dc747ef7915bdd9a05dde73cdc3034"} Feb 18 09:06:05 crc kubenswrapper[4556]: I0218 09:06:05.992562 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-dffhx" event={"ID":"ac316465-87df-4f89-99d3-7633309d3158","Type":"ContainerStarted","Data":"81dd5c3e19bd46579329e0ce3339dc029fea8000b611d9caf8958c54c5b86917"} Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.021234 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-kvbqn" event={"ID":"6ea95a80-9834-4f76-a250-6c6394c0df87","Type":"ContainerStarted","Data":"203ce5c25d17bb39d76319df74a85bd0fc4b19f82657b5ddf4a4b02d41dfa1a4"} Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.054484 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-55hcm" Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.054944 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wz622\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:06:06 crc kubenswrapper[4556]: E0218 09:06:06.057046 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 09:06:06.557025454 +0000 UTC m=+123.573986434 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wz622" (UID: "709f253c-93c3-4aaf-b3db-630775991aeb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.057782 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4dfqf" podStartSLOduration=100.05711876 podStartE2EDuration="1m40.05711876s" podCreationTimestamp="2026-02-18 09:04:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:06:06.050520501 +0000 UTC m=+123.067481481" watchObservedRunningTime="2026-02-18 09:06:06.05711876 +0000 UTC m=+123.074079740" Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.076434 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-zlp6j" event={"ID":"db10e059-4e0c-417d-b457-31fcd8734cc9","Type":"ContainerStarted","Data":"e28573514af553511ae6ac9bc693c8e33b6cede556a90010227e41d01e15b264"} Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.111285 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxkx5" event={"ID":"8129ffba-fc55-45de-945f-66071d25c4eb","Type":"ContainerStarted","Data":"a8c0c89ea191b8c7adb148897e92b56ca24e68dcd7e8df47ace238201075c940"} Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.127967 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2sdct" event={"ID":"6f9fe7cf-4384-4a95-abbb-20b487e7bc6c","Type":"ContainerStarted","Data":"8373698cd99541c1c4ee31d8fe5af9ed9c9634f75e7ed7da6b5a85dd9a493f96"} Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.130865 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j5fdh" event={"ID":"b3ca6b25-cb73-47a9-867c-8d1f6f628077","Type":"ContainerStarted","Data":"f073ca8fe2e4433878dd5bd0375b1ab0962e16486c4dfc844ab6c3e3146d09ae"} Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.139888 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-lzjw5" event={"ID":"9a47b8c4-1288-4b4b-9d82-4c6c22cc65b3","Type":"ContainerStarted","Data":"8dc9a8de60d74936d16a8c5d81201636abcc79d285451fb63259a4b84abae683"} Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.140289 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-lzjw5" Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.146466 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2sdct" Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.159662 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 09:06:06 crc kubenswrapper[4556]: E0218 09:06:06.160772 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 09:06:06.660753313 +0000 UTC m=+123.677714293 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.167991 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-26rzb" podStartSLOduration=100.167967279 podStartE2EDuration="1m40.167967279s" podCreationTimestamp="2026-02-18 09:04:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:06:06.098212701 +0000 UTC m=+123.115173681" watchObservedRunningTime="2026-02-18 09:06:06.167967279 +0000 UTC m=+123.184928259" Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.169174 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-hhxz6" event={"ID":"a303aa13-ef7e-4a65-a5d8-9763d2f0b4ea","Type":"ContainerStarted","Data":"171117f57cb690d35bd30669a4076276562b2fe48ef1496459d2872ad92cdcb2"} Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.169224 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-hhxz6" event={"ID":"a303aa13-ef7e-4a65-a5d8-9763d2f0b4ea","Type":"ContainerStarted","Data":"b0eff79a2fe4895fdb8c0067f8e32fe87a659b93b5d95051c855b28ba881280b"} Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.261699 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wqmhk"] Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.270287 4556 patch_prober.go:28] interesting pod/router-default-5444994796-j9ccg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 18 09:06:06 crc kubenswrapper[4556]: [-]has-synced failed: reason withheld Feb 18 09:06:06 crc kubenswrapper[4556]: [+]process-running ok Feb 18 09:06:06 crc kubenswrapper[4556]: healthz check failed Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.270375 4556 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-j9ccg" podUID="9cf0e7d1-1645-47aa-add2-3a76bfc624a7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.272458 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wz622\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:06:06 crc kubenswrapper[4556]: E0218 09:06:06.275334 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 09:06:06.775308838 +0000 UTC m=+123.792269818 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wz622" (UID: "709f253c-93c3-4aaf-b3db-630775991aeb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.278236 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-55hcm" podStartSLOduration=100.278218545 podStartE2EDuration="1m40.278218545s" podCreationTimestamp="2026-02-18 09:04:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:06:06.27794393 +0000 UTC m=+123.294904910" watchObservedRunningTime="2026-02-18 09:06:06.278218545 +0000 UTC m=+123.295179525" Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.339017 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-fmk6k" podStartSLOduration=100.338966368 podStartE2EDuration="1m40.338966368s" podCreationTimestamp="2026-02-18 09:04:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:06:06.331453932 +0000 UTC m=+123.348414912" watchObservedRunningTime="2026-02-18 09:06:06.338966368 +0000 UTC m=+123.355927348" Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.368188 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-ndnpr"] Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.369747 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ndnpr" Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.375822 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.376587 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 09:06:06 crc kubenswrapper[4556]: E0218 09:06:06.376841 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 09:06:06.876820522 +0000 UTC m=+123.893781502 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.377180 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wz622\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:06:06 crc kubenswrapper[4556]: E0218 09:06:06.379190 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 09:06:06.879176929 +0000 UTC m=+123.896137909 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wz622" (UID: "709f253c-93c3-4aaf-b3db-630775991aeb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.392465 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxkx5" podStartSLOduration=100.392444402 podStartE2EDuration="1m40.392444402s" podCreationTimestamp="2026-02-18 09:04:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:06:06.375591432 +0000 UTC m=+123.392552412" watchObservedRunningTime="2026-02-18 09:06:06.392444402 +0000 UTC m=+123.409405381" Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.394371 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ndnpr"] Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.441475 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lkxrl" podStartSLOduration=100.44146121 podStartE2EDuration="1m40.44146121s" podCreationTimestamp="2026-02-18 09:04:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:06:06.419607824 +0000 UTC m=+123.436568805" watchObservedRunningTime="2026-02-18 09:06:06.44146121 +0000 UTC m=+123.458422190" Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.469920 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-4vkc6" podStartSLOduration=100.469902766 podStartE2EDuration="1m40.469902766s" podCreationTimestamp="2026-02-18 09:04:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:06:06.468571955 +0000 UTC m=+123.485532935" watchObservedRunningTime="2026-02-18 09:06:06.469902766 +0000 UTC m=+123.486863746" Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.485567 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.485801 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngl5n\" (UniqueName: \"kubernetes.io/projected/e23bd1bf-79ec-4cc0-95c4-16959eab3aa7-kube-api-access-ngl5n\") pod \"redhat-marketplace-ndnpr\" (UID: \"e23bd1bf-79ec-4cc0-95c4-16959eab3aa7\") " pod="openshift-marketplace/redhat-marketplace-ndnpr" Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.485858 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e23bd1bf-79ec-4cc0-95c4-16959eab3aa7-catalog-content\") pod \"redhat-marketplace-ndnpr\" (UID: \"e23bd1bf-79ec-4cc0-95c4-16959eab3aa7\") " pod="openshift-marketplace/redhat-marketplace-ndnpr" Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.485894 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e23bd1bf-79ec-4cc0-95c4-16959eab3aa7-utilities\") pod \"redhat-marketplace-ndnpr\" (UID: \"e23bd1bf-79ec-4cc0-95c4-16959eab3aa7\") " pod="openshift-marketplace/redhat-marketplace-ndnpr" Feb 18 09:06:06 crc kubenswrapper[4556]: E0218 09:06:06.486102 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 09:06:06.986073352 +0000 UTC m=+124.003034332 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.556434 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-hhxz6" podStartSLOduration=100.556412279 podStartE2EDuration="1m40.556412279s" podCreationTimestamp="2026-02-18 09:04:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:06:06.555339383 +0000 UTC m=+123.572300362" watchObservedRunningTime="2026-02-18 09:06:06.556412279 +0000 UTC m=+123.573373258" Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.590073 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e23bd1bf-79ec-4cc0-95c4-16959eab3aa7-utilities\") pod \"redhat-marketplace-ndnpr\" (UID: \"e23bd1bf-79ec-4cc0-95c4-16959eab3aa7\") " pod="openshift-marketplace/redhat-marketplace-ndnpr" Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.590251 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngl5n\" (UniqueName: \"kubernetes.io/projected/e23bd1bf-79ec-4cc0-95c4-16959eab3aa7-kube-api-access-ngl5n\") pod \"redhat-marketplace-ndnpr\" (UID: \"e23bd1bf-79ec-4cc0-95c4-16959eab3aa7\") " pod="openshift-marketplace/redhat-marketplace-ndnpr" Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.590325 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e23bd1bf-79ec-4cc0-95c4-16959eab3aa7-catalog-content\") pod \"redhat-marketplace-ndnpr\" (UID: \"e23bd1bf-79ec-4cc0-95c4-16959eab3aa7\") " pod="openshift-marketplace/redhat-marketplace-ndnpr" Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.590372 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wz622\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.590411 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-z7mx2" podStartSLOduration=100.590388785 podStartE2EDuration="1m40.590388785s" podCreationTimestamp="2026-02-18 09:04:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:06:06.589435944 +0000 UTC m=+123.606396924" watchObservedRunningTime="2026-02-18 09:06:06.590388785 +0000 UTC m=+123.607349755" Feb 18 09:06:06 crc kubenswrapper[4556]: E0218 09:06:06.590628 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 09:06:07.090617264 +0000 UTC m=+124.107578244 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wz622" (UID: "709f253c-93c3-4aaf-b3db-630775991aeb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.590953 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e23bd1bf-79ec-4cc0-95c4-16959eab3aa7-utilities\") pod \"redhat-marketplace-ndnpr\" (UID: \"e23bd1bf-79ec-4cc0-95c4-16959eab3aa7\") " pod="openshift-marketplace/redhat-marketplace-ndnpr" Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.591398 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e23bd1bf-79ec-4cc0-95c4-16959eab3aa7-catalog-content\") pod \"redhat-marketplace-ndnpr\" (UID: \"e23bd1bf-79ec-4cc0-95c4-16959eab3aa7\") " pod="openshift-marketplace/redhat-marketplace-ndnpr" Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.616608 4556 csr.go:261] certificate signing request csr-zlpwz is approved, waiting to be issued Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.621960 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-6d57b" podStartSLOduration=100.621946145 podStartE2EDuration="1m40.621946145s" podCreationTimestamp="2026-02-18 09:04:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:06:06.61945287 +0000 UTC m=+123.636413850" watchObservedRunningTime="2026-02-18 09:06:06.621946145 +0000 UTC m=+123.638907124" Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.622980 4556 csr.go:257] certificate signing request csr-zlpwz is issued Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.648682 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-vd7f7" podStartSLOduration=100.648660795 podStartE2EDuration="1m40.648660795s" podCreationTimestamp="2026-02-18 09:04:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:06:06.647489104 +0000 UTC m=+123.664450083" watchObservedRunningTime="2026-02-18 09:06:06.648660795 +0000 UTC m=+123.665621776" Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.648703 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngl5n\" (UniqueName: \"kubernetes.io/projected/e23bd1bf-79ec-4cc0-95c4-16959eab3aa7-kube-api-access-ngl5n\") pod \"redhat-marketplace-ndnpr\" (UID: \"e23bd1bf-79ec-4cc0-95c4-16959eab3aa7\") " pod="openshift-marketplace/redhat-marketplace-ndnpr" Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.685759 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-pzbhd" podStartSLOduration=100.685739231 podStartE2EDuration="1m40.685739231s" podCreationTimestamp="2026-02-18 09:04:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:06:06.684426154 +0000 UTC m=+123.701387133" watchObservedRunningTime="2026-02-18 09:06:06.685739231 +0000 UTC m=+123.702700211" Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.690738 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 09:06:06 crc kubenswrapper[4556]: E0218 09:06:06.691046 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 09:06:07.191031055 +0000 UTC m=+124.207992035 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.736744 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29523420-qb78x" podStartSLOduration=100.736728527 podStartE2EDuration="1m40.736728527s" podCreationTimestamp="2026-02-18 09:04:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:06:06.735466614 +0000 UTC m=+123.752427595" watchObservedRunningTime="2026-02-18 09:06:06.736728527 +0000 UTC m=+123.753689507" Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.769817 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-gw4gv" podStartSLOduration=6.76979917 podStartE2EDuration="6.76979917s" podCreationTimestamp="2026-02-18 09:06:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:06:06.767512744 +0000 UTC m=+123.784473725" watchObservedRunningTime="2026-02-18 09:06:06.76979917 +0000 UTC m=+123.786760151" Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.772804 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-ntvpp"] Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.773824 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ntvpp" Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.777783 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ntvpp"] Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.791755 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wz622\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.791839 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc2d3f5c-f685-43ea-bf13-4bec129bf889-catalog-content\") pod \"redhat-marketplace-ntvpp\" (UID: \"bc2d3f5c-f685-43ea-bf13-4bec129bf889\") " pod="openshift-marketplace/redhat-marketplace-ntvpp" Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.791878 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc2d3f5c-f685-43ea-bf13-4bec129bf889-utilities\") pod \"redhat-marketplace-ntvpp\" (UID: \"bc2d3f5c-f685-43ea-bf13-4bec129bf889\") " pod="openshift-marketplace/redhat-marketplace-ntvpp" Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.791909 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbfxx\" (UniqueName: \"kubernetes.io/projected/bc2d3f5c-f685-43ea-bf13-4bec129bf889-kube-api-access-nbfxx\") pod \"redhat-marketplace-ntvpp\" (UID: \"bc2d3f5c-f685-43ea-bf13-4bec129bf889\") " pod="openshift-marketplace/redhat-marketplace-ntvpp" Feb 18 09:06:06 crc kubenswrapper[4556]: E0218 09:06:06.792258 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 09:06:07.292244598 +0000 UTC m=+124.309205578 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wz622" (UID: "709f253c-93c3-4aaf-b3db-630775991aeb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.798254 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-kvbqn" podStartSLOduration=6.798243912 podStartE2EDuration="6.798243912s" podCreationTimestamp="2026-02-18 09:06:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:06:06.797084162 +0000 UTC m=+123.814045143" watchObservedRunningTime="2026-02-18 09:06:06.798243912 +0000 UTC m=+123.815204892" Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.805769 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ndnpr" Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.842546 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dncx2" podStartSLOduration=100.842533749 podStartE2EDuration="1m40.842533749s" podCreationTimestamp="2026-02-18 09:04:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:06:06.840685044 +0000 UTC m=+123.857646024" watchObservedRunningTime="2026-02-18 09:06:06.842533749 +0000 UTC m=+123.859494729" Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.875369 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6ss98" podStartSLOduration=100.875346679 podStartE2EDuration="1m40.875346679s" podCreationTimestamp="2026-02-18 09:04:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:06:06.86803602 +0000 UTC m=+123.884997000" watchObservedRunningTime="2026-02-18 09:06:06.875346679 +0000 UTC m=+123.892307658" Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.893578 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.893939 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc2d3f5c-f685-43ea-bf13-4bec129bf889-catalog-content\") pod \"redhat-marketplace-ntvpp\" (UID: \"bc2d3f5c-f685-43ea-bf13-4bec129bf889\") " pod="openshift-marketplace/redhat-marketplace-ntvpp" Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.893982 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc2d3f5c-f685-43ea-bf13-4bec129bf889-utilities\") pod \"redhat-marketplace-ntvpp\" (UID: \"bc2d3f5c-f685-43ea-bf13-4bec129bf889\") " pod="openshift-marketplace/redhat-marketplace-ntvpp" Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.894019 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbfxx\" (UniqueName: \"kubernetes.io/projected/bc2d3f5c-f685-43ea-bf13-4bec129bf889-kube-api-access-nbfxx\") pod \"redhat-marketplace-ntvpp\" (UID: \"bc2d3f5c-f685-43ea-bf13-4bec129bf889\") " pod="openshift-marketplace/redhat-marketplace-ntvpp" Feb 18 09:06:06 crc kubenswrapper[4556]: E0218 09:06:06.894277 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 09:06:07.394255871 +0000 UTC m=+124.411216852 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.894826 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc2d3f5c-f685-43ea-bf13-4bec129bf889-catalog-content\") pod \"redhat-marketplace-ntvpp\" (UID: \"bc2d3f5c-f685-43ea-bf13-4bec129bf889\") " pod="openshift-marketplace/redhat-marketplace-ntvpp" Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.895049 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc2d3f5c-f685-43ea-bf13-4bec129bf889-utilities\") pod \"redhat-marketplace-ntvpp\" (UID: \"bc2d3f5c-f685-43ea-bf13-4bec129bf889\") " pod="openshift-marketplace/redhat-marketplace-ntvpp" Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.950105 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbfxx\" (UniqueName: \"kubernetes.io/projected/bc2d3f5c-f685-43ea-bf13-4bec129bf889-kube-api-access-nbfxx\") pod \"redhat-marketplace-ntvpp\" (UID: \"bc2d3f5c-f685-43ea-bf13-4bec129bf889\") " pod="openshift-marketplace/redhat-marketplace-ntvpp" Feb 18 09:06:06 crc kubenswrapper[4556]: I0218 09:06:06.997772 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wz622\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:06:06 crc kubenswrapper[4556]: E0218 09:06:06.998123 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 09:06:07.498109435 +0000 UTC m=+124.515070415 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wz622" (UID: "709f253c-93c3-4aaf-b3db-630775991aeb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.046020 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j5fdh" podStartSLOduration=101.046000429 podStartE2EDuration="1m41.046000429s" podCreationTimestamp="2026-02-18 09:04:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:06:07.044076073 +0000 UTC m=+124.061037053" watchObservedRunningTime="2026-02-18 09:06:07.046000429 +0000 UTC m=+124.062961409" Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.094959 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ntvpp" Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.101308 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 09:06:07 crc kubenswrapper[4556]: E0218 09:06:07.101617 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 09:06:07.601597564 +0000 UTC m=+124.618558544 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.174431 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-lzjw5" podStartSLOduration=101.174413424 podStartE2EDuration="1m41.174413424s" podCreationTimestamp="2026-02-18 09:04:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:06:07.141349633 +0000 UTC m=+124.158310613" watchObservedRunningTime="2026-02-18 09:06:07.174413424 +0000 UTC m=+124.191374404" Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.175732 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-dffhx" podStartSLOduration=101.175724378 podStartE2EDuration="1m41.175724378s" podCreationTimestamp="2026-02-18 09:04:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:06:07.172984559 +0000 UTC m=+124.189945539" watchObservedRunningTime="2026-02-18 09:06:07.175724378 +0000 UTC m=+124.192685358" Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.202541 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-mvvrj" event={"ID":"87338513-3a60-432f-8edf-02695687e1e2","Type":"ContainerStarted","Data":"8b5de01a5872e7b62e4cf541a025cdb6afaeb1c78fa1158ef7a70011922a7f8f"} Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.203571 4556 patch_prober.go:28] interesting pod/downloads-7954f5f757-mvvrj container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body= Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.203602 4556 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-mvvrj" podUID="87338513-3a60-432f-8edf-02695687e1e2" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.205089 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wz622\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:06:07 crc kubenswrapper[4556]: E0218 09:06:07.206141 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 09:06:07.706130364 +0000 UTC m=+124.723091344 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wz622" (UID: "709f253c-93c3-4aaf-b3db-630775991aeb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.207619 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxkx5" Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.209659 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxkx5" Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.213347 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-vc7vq" podStartSLOduration=101.213335835 podStartE2EDuration="1m41.213335835s" podCreationTimestamp="2026-02-18 09:04:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:06:07.211971682 +0000 UTC m=+124.228932682" watchObservedRunningTime="2026-02-18 09:06:07.213335835 +0000 UTC m=+124.230296815" Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.214275 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-lsngn" event={"ID":"b1111d02-c210-4787-8dda-b118339cfc79","Type":"ContainerStarted","Data":"78818a1fbcfc84c6c0a1d0472aaaef1f452e0795367ee5c9e937d3c24b2c70db"} Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.217127 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxkx5" Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.228476 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5smq4" event={"ID":"1aa03714-6c33-42a8-a427-04a7ce839f21","Type":"ContainerStarted","Data":"227273d3983227b02323eb4a6605c0b482f7fa25eb5179e1e8140ec366652eff"} Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.230801 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-zlp6j" event={"ID":"db10e059-4e0c-417d-b457-31fcd8734cc9","Type":"ContainerStarted","Data":"2026ae0c57e357908c9fa8b966dbd07a5f7be4e26102d20398c85d3efe046c6e"} Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.230851 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-zlp6j" event={"ID":"db10e059-4e0c-417d-b457-31fcd8734cc9","Type":"ContainerStarted","Data":"3b3288e407ccc4e1a2b9b85e05e245caf92bcac8ee2e4585f6dd3e1a12f2258a"} Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.233975 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-zlp6j" Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.250085 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-mvvrj" podStartSLOduration=101.250064052 podStartE2EDuration="1m41.250064052s" podCreationTimestamp="2026-02-18 09:04:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:06:07.248385518 +0000 UTC m=+124.265346498" watchObservedRunningTime="2026-02-18 09:06:07.250064052 +0000 UTC m=+124.267025032" Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.251660 4556 generic.go:334] "Generic (PLEG): container finished" podID="25cebcb5-2e64-4390-9d61-f06607eee280" containerID="21164cb24c7ae3c208c4b8151a1fce3c76cc94a1e3dfa324207b42c2c85785f0" exitCode=0 Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.251731 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wqmhk" event={"ID":"25cebcb5-2e64-4390-9d61-f06607eee280","Type":"ContainerDied","Data":"21164cb24c7ae3c208c4b8151a1fce3c76cc94a1e3dfa324207b42c2c85785f0"} Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.251756 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wqmhk" event={"ID":"25cebcb5-2e64-4390-9d61-f06607eee280","Type":"ContainerStarted","Data":"5c554bf507ddc09a88bd7cbc6738f8f0a6481718839f00552ffbfa2bb81ad115"} Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.253405 4556 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.258058 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-kvbqn" event={"ID":"6ea95a80-9834-4f76-a250-6c6394c0df87","Type":"ContainerStarted","Data":"368558ca6c8abdf73a16ed837463547fd8d0cb53a4d5fb4d0e1475bc08ef5077"} Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.259629 4556 generic.go:334] "Generic (PLEG): container finished" podID="90477a11-cc7c-48cf-a210-9d1d802360ce" containerID="c623b3371772b683d554cf6cc7a299bd1973b82f4a7504608f10972d19a633e9" exitCode=0 Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.259683 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sd4dx" event={"ID":"90477a11-cc7c-48cf-a210-9d1d802360ce","Type":"ContainerDied","Data":"c623b3371772b683d554cf6cc7a299bd1973b82f4a7504608f10972d19a633e9"} Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.259720 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sd4dx" event={"ID":"90477a11-cc7c-48cf-a210-9d1d802360ce","Type":"ContainerStarted","Data":"1606cfc7ff8627098dd6180d64cf31ed84f3331507f250eaa8b250f3995e57a1"} Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.275229 4556 patch_prober.go:28] interesting pod/router-default-5444994796-j9ccg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 18 09:06:07 crc kubenswrapper[4556]: [-]has-synced failed: reason withheld Feb 18 09:06:07 crc kubenswrapper[4556]: [+]process-running ok Feb 18 09:06:07 crc kubenswrapper[4556]: healthz check failed Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.275274 4556 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-j9ccg" podUID="9cf0e7d1-1645-47aa-add2-3a76bfc624a7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.275241 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6ss98" event={"ID":"8d17a601-499d-4d9b-945a-557378ea7400","Type":"ContainerStarted","Data":"ac4cad44a333a39e7d95f6d8f54bf8a858d3a3f6f981c0422d117361fc87f3b1"} Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.295911 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-vxdmn" podStartSLOduration=101.29589819 podStartE2EDuration="1m41.29589819s" podCreationTimestamp="2026-02-18 09:04:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:06:07.292443309 +0000 UTC m=+124.309404289" watchObservedRunningTime="2026-02-18 09:06:07.29589819 +0000 UTC m=+124.312859170" Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.306727 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-qw4hv" event={"ID":"ccb1af48-7979-4f75-983e-7455ac471966","Type":"ContainerStarted","Data":"e76504b3aec47cc0475d610dddaea8b3e916668a974ca88bd2858f1576060d70"} Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.307711 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 09:06:07 crc kubenswrapper[4556]: E0218 09:06:07.308870 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 09:06:07.808857322 +0000 UTC m=+124.825818302 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.315548 4556 generic.go:334] "Generic (PLEG): container finished" podID="7875d007-bdbc-428a-b291-283e8a1cdad0" containerID="449d08a8d66546a788eda025d228560e03d01cbafbe835839b2982d1aa177301" exitCode=0 Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.316257 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ss8sl" event={"ID":"7875d007-bdbc-428a-b291-283e8a1cdad0","Type":"ContainerDied","Data":"449d08a8d66546a788eda025d228560e03d01cbafbe835839b2982d1aa177301"} Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.316308 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ss8sl" event={"ID":"7875d007-bdbc-428a-b291-283e8a1cdad0","Type":"ContainerStarted","Data":"2fb3b6373c0c3b961335f85fd9735a8bb14ff7f1fe1eff214be4098697d47785"} Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.336419 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sxb49" event={"ID":"fd45fd0a-9812-4ff9-8bf7-91fda0bb0df3","Type":"ContainerStarted","Data":"77c466d433607a7474986a26b868a2cc752810aac5d675b36e8bbcf09a24657a"} Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.336697 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sxb49" event={"ID":"fd45fd0a-9812-4ff9-8bf7-91fda0bb0df3","Type":"ContainerStarted","Data":"b6376e387906feab0368d01467c3824809beae2709a6234a0bb004bd63c2bf4c"} Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.346178 4556 generic.go:334] "Generic (PLEG): container finished" podID="eac84811-d54a-47c2-ae5e-77f3a531fc4f" containerID="6f3e016b9541f7fb845611ae98130a4b33004793082ce622b1c15b7c7e02297d" exitCode=0 Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.346234 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-brf28" event={"ID":"eac84811-d54a-47c2-ae5e-77f3a531fc4f","Type":"ContainerDied","Data":"6f3e016b9541f7fb845611ae98130a4b33004793082ce622b1c15b7c7e02297d"} Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.346253 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-brf28" event={"ID":"eac84811-d54a-47c2-ae5e-77f3a531fc4f","Type":"ContainerStarted","Data":"8630aa61bf54c6289842987772ae6787bdc62980758815c13ed806b02bc25265"} Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.366507 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lkxrl" event={"ID":"ff9016b7-a27a-4774-a7a2-498eee516d28","Type":"ContainerStarted","Data":"c9c97333c818c3beef327e211fb28b393c4c7310c1c3eeddb0efc4afd0b77724"} Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.367782 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-8gpk5" podStartSLOduration=101.367768344 podStartE2EDuration="1m41.367768344s" podCreationTimestamp="2026-02-18 09:04:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:06:07.366517763 +0000 UTC m=+124.383478743" watchObservedRunningTime="2026-02-18 09:06:07.367768344 +0000 UTC m=+124.384729324" Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.372254 4556 generic.go:334] "Generic (PLEG): container finished" podID="0aa2e566-fd95-498b-8df5-b63afb290572" containerID="c2201de00ba0f9c778a8b2bbece184c31365c38bd6019f97cef435b6315fc2ab" exitCode=0 Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.372375 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29523420-qb78x" event={"ID":"0aa2e566-fd95-498b-8df5-b63afb290572","Type":"ContainerDied","Data":"c2201de00ba0f9c778a8b2bbece184c31365c38bd6019f97cef435b6315fc2ab"} Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.379398 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ndnpr"] Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.411383 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wz622\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:06:07 crc kubenswrapper[4556]: E0218 09:06:07.411745 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 09:06:07.911731467 +0000 UTC m=+124.928692447 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wz622" (UID: "709f253c-93c3-4aaf-b3db-630775991aeb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:07 crc kubenswrapper[4556]: W0218 09:06:07.412496 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode23bd1bf_79ec_4cc0_95c4_16959eab3aa7.slice/crio-4804b76870038a935a002ae8e9b4d53a56df4542126f2424cd995637a12631b8 WatchSource:0}: Error finding container 4804b76870038a935a002ae8e9b4d53a56df4542126f2424cd995637a12631b8: Status 404 returned error can't find the container with id 4804b76870038a935a002ae8e9b4d53a56df4542126f2424cd995637a12631b8 Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.449461 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-vxdmn" event={"ID":"9db33ada-017a-4281-85c1-0c1530bc3bbc","Type":"ContainerStarted","Data":"f95e8072a9929cfc1b3f58f632a3d8a2bdbaf2dafc7a7f7a6cb653d94674a089"} Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.462331 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxkx5" Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.463665 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-vc7vq" Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.465468 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-lv2k7" podStartSLOduration=101.465308685 podStartE2EDuration="1m41.465308685s" podCreationTimestamp="2026-02-18 09:04:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:06:07.45039477 +0000 UTC m=+124.467355750" watchObservedRunningTime="2026-02-18 09:06:07.465308685 +0000 UTC m=+124.482269665" Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.488208 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-lzjw5" Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.490978 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-vxdmn" Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.491539 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-lsngn" podStartSLOduration=101.491526011 podStartE2EDuration="1m41.491526011s" podCreationTimestamp="2026-02-18 09:04:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:06:07.488755755 +0000 UTC m=+124.505716735" watchObservedRunningTime="2026-02-18 09:06:07.491526011 +0000 UTC m=+124.508486991" Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.526814 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 09:06:07 crc kubenswrapper[4556]: E0218 09:06:07.528986 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 09:06:08.028966777 +0000 UTC m=+125.045927757 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.608677 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-lsngn" Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.609165 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-lsngn" Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.632295 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wz622\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:06:07 crc kubenswrapper[4556]: E0218 09:06:07.632907 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 09:06:08.132891314 +0000 UTC m=+125.149852285 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wz622" (UID: "709f253c-93c3-4aaf-b3db-630775991aeb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.633205 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-18 09:01:06 +0000 UTC, rotation deadline is 2027-01-02 21:02:53.649808917 +0000 UTC Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.633306 4556 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 7643h56m46.016505708s for next certificate rotation Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.645496 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bwwg2"] Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.646690 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bwwg2" Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.648083 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.652654 4556 patch_prober.go:28] interesting pod/apiserver-76f77b778f-lsngn container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Feb 18 09:06:07 crc kubenswrapper[4556]: [+]log ok Feb 18 09:06:07 crc kubenswrapper[4556]: [+]etcd ok Feb 18 09:06:07 crc kubenswrapper[4556]: [+]poststarthook/start-apiserver-admission-initializer ok Feb 18 09:06:07 crc kubenswrapper[4556]: [+]poststarthook/generic-apiserver-start-informers ok Feb 18 09:06:07 crc kubenswrapper[4556]: [+]poststarthook/max-in-flight-filter ok Feb 18 09:06:07 crc kubenswrapper[4556]: [+]poststarthook/storage-object-count-tracker-hook ok Feb 18 09:06:07 crc kubenswrapper[4556]: [+]poststarthook/image.openshift.io-apiserver-caches ok Feb 18 09:06:07 crc kubenswrapper[4556]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Feb 18 09:06:07 crc kubenswrapper[4556]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Feb 18 09:06:07 crc kubenswrapper[4556]: [+]poststarthook/project.openshift.io-projectcache ok Feb 18 09:06:07 crc kubenswrapper[4556]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Feb 18 09:06:07 crc kubenswrapper[4556]: [+]poststarthook/openshift.io-startinformers ok Feb 18 09:06:07 crc kubenswrapper[4556]: [+]poststarthook/openshift.io-restmapperupdater ok Feb 18 09:06:07 crc kubenswrapper[4556]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Feb 18 09:06:07 crc kubenswrapper[4556]: livez check failed Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.652703 4556 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-lsngn" podUID="b1111d02-c210-4787-8dda-b118339cfc79" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.656750 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bwwg2"] Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.676281 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ntvpp"] Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.676828 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-zlp6j" podStartSLOduration=7.676808592 podStartE2EDuration="7.676808592s" podCreationTimestamp="2026-02-18 09:06:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:06:07.646988555 +0000 UTC m=+124.663949535" watchObservedRunningTime="2026-02-18 09:06:07.676808592 +0000 UTC m=+124.693769572" Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.739657 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.739917 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fdd9dfd4-872b-42fc-960f-5d4c51e6e66e-catalog-content\") pod \"redhat-operators-bwwg2\" (UID: \"fdd9dfd4-872b-42fc-960f-5d4c51e6e66e\") " pod="openshift-marketplace/redhat-operators-bwwg2" Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.740012 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xgm2\" (UniqueName: \"kubernetes.io/projected/fdd9dfd4-872b-42fc-960f-5d4c51e6e66e-kube-api-access-2xgm2\") pod \"redhat-operators-bwwg2\" (UID: \"fdd9dfd4-872b-42fc-960f-5d4c51e6e66e\") " pod="openshift-marketplace/redhat-operators-bwwg2" Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.740108 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fdd9dfd4-872b-42fc-960f-5d4c51e6e66e-utilities\") pod \"redhat-operators-bwwg2\" (UID: \"fdd9dfd4-872b-42fc-960f-5d4c51e6e66e\") " pod="openshift-marketplace/redhat-operators-bwwg2" Feb 18 09:06:07 crc kubenswrapper[4556]: E0218 09:06:07.740276 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 09:06:08.240259103 +0000 UTC m=+125.257220084 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.773788 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.774648 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.783530 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.783727 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.784022 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.844005 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7cccf6d1-154d-456c-96f5-9af880a6a87b-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"7cccf6d1-154d-456c-96f5-9af880a6a87b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.844260 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wz622\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.844356 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fdd9dfd4-872b-42fc-960f-5d4c51e6e66e-utilities\") pod \"redhat-operators-bwwg2\" (UID: \"fdd9dfd4-872b-42fc-960f-5d4c51e6e66e\") " pod="openshift-marketplace/redhat-operators-bwwg2" Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.844565 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7cccf6d1-154d-456c-96f5-9af880a6a87b-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"7cccf6d1-154d-456c-96f5-9af880a6a87b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.844645 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fdd9dfd4-872b-42fc-960f-5d4c51e6e66e-catalog-content\") pod \"redhat-operators-bwwg2\" (UID: \"fdd9dfd4-872b-42fc-960f-5d4c51e6e66e\") " pod="openshift-marketplace/redhat-operators-bwwg2" Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.844832 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xgm2\" (UniqueName: \"kubernetes.io/projected/fdd9dfd4-872b-42fc-960f-5d4c51e6e66e-kube-api-access-2xgm2\") pod \"redhat-operators-bwwg2\" (UID: \"fdd9dfd4-872b-42fc-960f-5d4c51e6e66e\") " pod="openshift-marketplace/redhat-operators-bwwg2" Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.844967 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fdd9dfd4-872b-42fc-960f-5d4c51e6e66e-utilities\") pod \"redhat-operators-bwwg2\" (UID: \"fdd9dfd4-872b-42fc-960f-5d4c51e6e66e\") " pod="openshift-marketplace/redhat-operators-bwwg2" Feb 18 09:06:07 crc kubenswrapper[4556]: E0218 09:06:07.845342 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 09:06:08.345308304 +0000 UTC m=+125.362269285 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wz622" (UID: "709f253c-93c3-4aaf-b3db-630775991aeb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.845488 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fdd9dfd4-872b-42fc-960f-5d4c51e6e66e-catalog-content\") pod \"redhat-operators-bwwg2\" (UID: \"fdd9dfd4-872b-42fc-960f-5d4c51e6e66e\") " pod="openshift-marketplace/redhat-operators-bwwg2" Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.880315 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xgm2\" (UniqueName: \"kubernetes.io/projected/fdd9dfd4-872b-42fc-960f-5d4c51e6e66e-kube-api-access-2xgm2\") pod \"redhat-operators-bwwg2\" (UID: \"fdd9dfd4-872b-42fc-960f-5d4c51e6e66e\") " pod="openshift-marketplace/redhat-operators-bwwg2" Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.912329 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sxb49" podStartSLOduration=101.91231022 podStartE2EDuration="1m41.91231022s" podCreationTimestamp="2026-02-18 09:04:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:06:07.911686548 +0000 UTC m=+124.928647529" watchObservedRunningTime="2026-02-18 09:06:07.91231022 +0000 UTC m=+124.929271200" Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.947778 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.947992 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7cccf6d1-154d-456c-96f5-9af880a6a87b-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"7cccf6d1-154d-456c-96f5-9af880a6a87b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.948195 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7cccf6d1-154d-456c-96f5-9af880a6a87b-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"7cccf6d1-154d-456c-96f5-9af880a6a87b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 18 09:06:07 crc kubenswrapper[4556]: E0218 09:06:07.948518 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 09:06:08.44849665 +0000 UTC m=+125.465457630 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.948554 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7cccf6d1-154d-456c-96f5-9af880a6a87b-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"7cccf6d1-154d-456c-96f5-9af880a6a87b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.970521 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5smq4" podStartSLOduration=101.970505266 podStartE2EDuration="1m41.970505266s" podCreationTimestamp="2026-02-18 09:04:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:06:07.95450528 +0000 UTC m=+124.971466260" watchObservedRunningTime="2026-02-18 09:06:07.970505266 +0000 UTC m=+124.987466246" Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.971650 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-cctvp"] Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.972547 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cctvp" Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.984766 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cctvp"] Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.987266 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7cccf6d1-154d-456c-96f5-9af880a6a87b-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"7cccf6d1-154d-456c-96f5-9af880a6a87b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 18 09:06:07 crc kubenswrapper[4556]: I0218 09:06:07.999728 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bwwg2" Feb 18 09:06:08 crc kubenswrapper[4556]: I0218 09:06:08.051777 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/accd4caa-1c67-4430-aa3a-db398194a103-utilities\") pod \"redhat-operators-cctvp\" (UID: \"accd4caa-1c67-4430-aa3a-db398194a103\") " pod="openshift-marketplace/redhat-operators-cctvp" Feb 18 09:06:08 crc kubenswrapper[4556]: I0218 09:06:08.051900 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjssn\" (UniqueName: \"kubernetes.io/projected/accd4caa-1c67-4430-aa3a-db398194a103-kube-api-access-tjssn\") pod \"redhat-operators-cctvp\" (UID: \"accd4caa-1c67-4430-aa3a-db398194a103\") " pod="openshift-marketplace/redhat-operators-cctvp" Feb 18 09:06:08 crc kubenswrapper[4556]: I0218 09:06:08.051967 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wz622\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:06:08 crc kubenswrapper[4556]: I0218 09:06:08.051992 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/accd4caa-1c67-4430-aa3a-db398194a103-catalog-content\") pod \"redhat-operators-cctvp\" (UID: \"accd4caa-1c67-4430-aa3a-db398194a103\") " pod="openshift-marketplace/redhat-operators-cctvp" Feb 18 09:06:08 crc kubenswrapper[4556]: E0218 09:06:08.052333 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 09:06:08.55232186 +0000 UTC m=+125.569282841 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wz622" (UID: "709f253c-93c3-4aaf-b3db-630775991aeb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:08 crc kubenswrapper[4556]: I0218 09:06:08.103120 4556 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Feb 18 09:06:08 crc kubenswrapper[4556]: I0218 09:06:08.128951 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 18 09:06:08 crc kubenswrapper[4556]: I0218 09:06:08.152575 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 09:06:08 crc kubenswrapper[4556]: I0218 09:06:08.152833 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/accd4caa-1c67-4430-aa3a-db398194a103-catalog-content\") pod \"redhat-operators-cctvp\" (UID: \"accd4caa-1c67-4430-aa3a-db398194a103\") " pod="openshift-marketplace/redhat-operators-cctvp" Feb 18 09:06:08 crc kubenswrapper[4556]: I0218 09:06:08.152878 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/accd4caa-1c67-4430-aa3a-db398194a103-utilities\") pod \"redhat-operators-cctvp\" (UID: \"accd4caa-1c67-4430-aa3a-db398194a103\") " pod="openshift-marketplace/redhat-operators-cctvp" Feb 18 09:06:08 crc kubenswrapper[4556]: I0218 09:06:08.152944 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjssn\" (UniqueName: \"kubernetes.io/projected/accd4caa-1c67-4430-aa3a-db398194a103-kube-api-access-tjssn\") pod \"redhat-operators-cctvp\" (UID: \"accd4caa-1c67-4430-aa3a-db398194a103\") " pod="openshift-marketplace/redhat-operators-cctvp" Feb 18 09:06:08 crc kubenswrapper[4556]: E0218 09:06:08.153247 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 09:06:08.653234959 +0000 UTC m=+125.670195939 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:08 crc kubenswrapper[4556]: I0218 09:06:08.153555 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/accd4caa-1c67-4430-aa3a-db398194a103-catalog-content\") pod \"redhat-operators-cctvp\" (UID: \"accd4caa-1c67-4430-aa3a-db398194a103\") " pod="openshift-marketplace/redhat-operators-cctvp" Feb 18 09:06:08 crc kubenswrapper[4556]: I0218 09:06:08.153760 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/accd4caa-1c67-4430-aa3a-db398194a103-utilities\") pod \"redhat-operators-cctvp\" (UID: \"accd4caa-1c67-4430-aa3a-db398194a103\") " pod="openshift-marketplace/redhat-operators-cctvp" Feb 18 09:06:08 crc kubenswrapper[4556]: I0218 09:06:08.173579 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjssn\" (UniqueName: \"kubernetes.io/projected/accd4caa-1c67-4430-aa3a-db398194a103-kube-api-access-tjssn\") pod \"redhat-operators-cctvp\" (UID: \"accd4caa-1c67-4430-aa3a-db398194a103\") " pod="openshift-marketplace/redhat-operators-cctvp" Feb 18 09:06:08 crc kubenswrapper[4556]: I0218 09:06:08.243582 4556 patch_prober.go:28] interesting pod/router-default-5444994796-j9ccg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 18 09:06:08 crc kubenswrapper[4556]: [-]has-synced failed: reason withheld Feb 18 09:06:08 crc kubenswrapper[4556]: [+]process-running ok Feb 18 09:06:08 crc kubenswrapper[4556]: healthz check failed Feb 18 09:06:08 crc kubenswrapper[4556]: I0218 09:06:08.243640 4556 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-j9ccg" podUID="9cf0e7d1-1645-47aa-add2-3a76bfc624a7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 18 09:06:08 crc kubenswrapper[4556]: I0218 09:06:08.264388 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wz622\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:06:08 crc kubenswrapper[4556]: E0218 09:06:08.264941 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 09:06:08.76491969 +0000 UTC m=+125.781880670 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wz622" (UID: "709f253c-93c3-4aaf-b3db-630775991aeb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:08 crc kubenswrapper[4556]: I0218 09:06:08.288480 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bwwg2"] Feb 18 09:06:08 crc kubenswrapper[4556]: I0218 09:06:08.290028 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cctvp" Feb 18 09:06:08 crc kubenswrapper[4556]: W0218 09:06:08.358817 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfdd9dfd4_872b_42fc_960f_5d4c51e6e66e.slice/crio-887663eb39f115ab07d951379bc7310153d46c4bf8928ed2dc23a27df0654b1e WatchSource:0}: Error finding container 887663eb39f115ab07d951379bc7310153d46c4bf8928ed2dc23a27df0654b1e: Status 404 returned error can't find the container with id 887663eb39f115ab07d951379bc7310153d46c4bf8928ed2dc23a27df0654b1e Feb 18 09:06:08 crc kubenswrapper[4556]: I0218 09:06:08.366846 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 09:06:08 crc kubenswrapper[4556]: E0218 09:06:08.371100 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 09:06:08.871067215 +0000 UTC m=+125.888028195 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:08 crc kubenswrapper[4556]: I0218 09:06:08.371538 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wz622\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:06:08 crc kubenswrapper[4556]: E0218 09:06:08.371862 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 09:06:08.871855116 +0000 UTC m=+125.888816096 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wz622" (UID: "709f253c-93c3-4aaf-b3db-630775991aeb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 09:06:08 crc kubenswrapper[4556]: I0218 09:06:08.399906 4556 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-02-18T09:06:08.10313772Z","Handler":null,"Name":""} Feb 18 09:06:08 crc kubenswrapper[4556]: I0218 09:06:08.425588 4556 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Feb 18 09:06:08 crc kubenswrapper[4556]: I0218 09:06:08.425626 4556 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Feb 18 09:06:08 crc kubenswrapper[4556]: I0218 09:06:08.442448 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bwwg2" event={"ID":"fdd9dfd4-872b-42fc-960f-5d4c51e6e66e","Type":"ContainerStarted","Data":"887663eb39f115ab07d951379bc7310153d46c4bf8928ed2dc23a27df0654b1e"} Feb 18 09:06:08 crc kubenswrapper[4556]: I0218 09:06:08.444023 4556 generic.go:334] "Generic (PLEG): container finished" podID="e23bd1bf-79ec-4cc0-95c4-16959eab3aa7" containerID="fb51382c701dac9eff637f394b33ff49ba492534b3ec98a703fe8770b728aab1" exitCode=0 Feb 18 09:06:08 crc kubenswrapper[4556]: I0218 09:06:08.444075 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ndnpr" event={"ID":"e23bd1bf-79ec-4cc0-95c4-16959eab3aa7","Type":"ContainerDied","Data":"fb51382c701dac9eff637f394b33ff49ba492534b3ec98a703fe8770b728aab1"} Feb 18 09:06:08 crc kubenswrapper[4556]: I0218 09:06:08.444126 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ndnpr" event={"ID":"e23bd1bf-79ec-4cc0-95c4-16959eab3aa7","Type":"ContainerStarted","Data":"4804b76870038a935a002ae8e9b4d53a56df4542126f2424cd995637a12631b8"} Feb 18 09:06:08 crc kubenswrapper[4556]: I0218 09:06:08.460536 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-qw4hv" event={"ID":"ccb1af48-7979-4f75-983e-7455ac471966","Type":"ContainerStarted","Data":"2a4abc8d8a5aa56bdbb5f0fac7874688d0f16a392f78f5e16c07b23fbbcd732b"} Feb 18 09:06:08 crc kubenswrapper[4556]: I0218 09:06:08.460623 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-qw4hv" event={"ID":"ccb1af48-7979-4f75-983e-7455ac471966","Type":"ContainerStarted","Data":"98f975e07a76d8104d496b0aa69242075314c4edba603f2339c8fd16fc49f51d"} Feb 18 09:06:08 crc kubenswrapper[4556]: I0218 09:06:08.469830 4556 generic.go:334] "Generic (PLEG): container finished" podID="bc2d3f5c-f685-43ea-bf13-4bec129bf889" containerID="27e59ca8d3945d13facd4466d769159dd0dd7ece817888020cefdb60ad599b24" exitCode=0 Feb 18 09:06:08 crc kubenswrapper[4556]: I0218 09:06:08.471018 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ntvpp" event={"ID":"bc2d3f5c-f685-43ea-bf13-4bec129bf889","Type":"ContainerDied","Data":"27e59ca8d3945d13facd4466d769159dd0dd7ece817888020cefdb60ad599b24"} Feb 18 09:06:08 crc kubenswrapper[4556]: I0218 09:06:08.471111 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ntvpp" event={"ID":"bc2d3f5c-f685-43ea-bf13-4bec129bf889","Type":"ContainerStarted","Data":"3adc525ce621d004dcc2338e52a240b04c5fc126a1af2ebd25f5eeae1d3421dd"} Feb 18 09:06:08 crc kubenswrapper[4556]: I0218 09:06:08.472807 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 09:06:08 crc kubenswrapper[4556]: I0218 09:06:08.473296 4556 patch_prober.go:28] interesting pod/downloads-7954f5f757-mvvrj container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body= Feb 18 09:06:08 crc kubenswrapper[4556]: I0218 09:06:08.473331 4556 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-mvvrj" podUID="87338513-3a60-432f-8edf-02695687e1e2" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" Feb 18 09:06:08 crc kubenswrapper[4556]: I0218 09:06:08.484674 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 18 09:06:08 crc kubenswrapper[4556]: I0218 09:06:08.493145 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 18 09:06:08 crc kubenswrapper[4556]: I0218 09:06:08.575934 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wz622\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:06:08 crc kubenswrapper[4556]: I0218 09:06:08.583444 4556 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 18 09:06:08 crc kubenswrapper[4556]: I0218 09:06:08.583477 4556 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wz622\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:06:08 crc kubenswrapper[4556]: I0218 09:06:08.662260 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wz622\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:06:08 crc kubenswrapper[4556]: I0218 09:06:08.793609 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29523420-qb78x" Feb 18 09:06:08 crc kubenswrapper[4556]: I0218 09:06:08.849288 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cctvp"] Feb 18 09:06:08 crc kubenswrapper[4556]: I0218 09:06:08.881905 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2gtkk\" (UniqueName: \"kubernetes.io/projected/0aa2e566-fd95-498b-8df5-b63afb290572-kube-api-access-2gtkk\") pod \"0aa2e566-fd95-498b-8df5-b63afb290572\" (UID: \"0aa2e566-fd95-498b-8df5-b63afb290572\") " Feb 18 09:06:08 crc kubenswrapper[4556]: I0218 09:06:08.883187 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0aa2e566-fd95-498b-8df5-b63afb290572-config-volume\") pod \"0aa2e566-fd95-498b-8df5-b63afb290572\" (UID: \"0aa2e566-fd95-498b-8df5-b63afb290572\") " Feb 18 09:06:08 crc kubenswrapper[4556]: I0218 09:06:08.883230 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0aa2e566-fd95-498b-8df5-b63afb290572-secret-volume\") pod \"0aa2e566-fd95-498b-8df5-b63afb290572\" (UID: \"0aa2e566-fd95-498b-8df5-b63afb290572\") " Feb 18 09:06:08 crc kubenswrapper[4556]: I0218 09:06:08.883896 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0aa2e566-fd95-498b-8df5-b63afb290572-config-volume" (OuterVolumeSpecName: "config-volume") pod "0aa2e566-fd95-498b-8df5-b63afb290572" (UID: "0aa2e566-fd95-498b-8df5-b63afb290572"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:06:08 crc kubenswrapper[4556]: I0218 09:06:08.888601 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0aa2e566-fd95-498b-8df5-b63afb290572-kube-api-access-2gtkk" (OuterVolumeSpecName: "kube-api-access-2gtkk") pod "0aa2e566-fd95-498b-8df5-b63afb290572" (UID: "0aa2e566-fd95-498b-8df5-b63afb290572"). InnerVolumeSpecName "kube-api-access-2gtkk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:06:08 crc kubenswrapper[4556]: I0218 09:06:08.892623 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0aa2e566-fd95-498b-8df5-b63afb290572-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "0aa2e566-fd95-498b-8df5-b63afb290572" (UID: "0aa2e566-fd95-498b-8df5-b63afb290572"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:06:08 crc kubenswrapper[4556]: W0218 09:06:08.905247 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaccd4caa_1c67_4430_aa3a_db398194a103.slice/crio-9052dc47a6ad03afebbea1f5a4cf2a9869e0f1e1c0025ef9a9eebddb95f420b0 WatchSource:0}: Error finding container 9052dc47a6ad03afebbea1f5a4cf2a9869e0f1e1c0025ef9a9eebddb95f420b0: Status 404 returned error can't find the container with id 9052dc47a6ad03afebbea1f5a4cf2a9869e0f1e1c0025ef9a9eebddb95f420b0 Feb 18 09:06:08 crc kubenswrapper[4556]: I0218 09:06:08.926871 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Feb 18 09:06:08 crc kubenswrapper[4556]: I0218 09:06:08.936238 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:06:08 crc kubenswrapper[4556]: I0218 09:06:08.985424 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2gtkk\" (UniqueName: \"kubernetes.io/projected/0aa2e566-fd95-498b-8df5-b63afb290572-kube-api-access-2gtkk\") on node \"crc\" DevicePath \"\"" Feb 18 09:06:08 crc kubenswrapper[4556]: I0218 09:06:08.985455 4556 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0aa2e566-fd95-498b-8df5-b63afb290572-config-volume\") on node \"crc\" DevicePath \"\"" Feb 18 09:06:08 crc kubenswrapper[4556]: I0218 09:06:08.985464 4556 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0aa2e566-fd95-498b-8df5-b63afb290572-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 18 09:06:09 crc kubenswrapper[4556]: I0218 09:06:09.242420 4556 patch_prober.go:28] interesting pod/router-default-5444994796-j9ccg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 18 09:06:09 crc kubenswrapper[4556]: [-]has-synced failed: reason withheld Feb 18 09:06:09 crc kubenswrapper[4556]: [+]process-running ok Feb 18 09:06:09 crc kubenswrapper[4556]: healthz check failed Feb 18 09:06:09 crc kubenswrapper[4556]: I0218 09:06:09.242729 4556 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-j9ccg" podUID="9cf0e7d1-1645-47aa-add2-3a76bfc624a7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 18 09:06:09 crc kubenswrapper[4556]: I0218 09:06:09.295957 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Feb 18 09:06:09 crc kubenswrapper[4556]: I0218 09:06:09.518258 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"7cccf6d1-154d-456c-96f5-9af880a6a87b","Type":"ContainerStarted","Data":"743642d9ce9a0dd4257cf1b2241886e562c638cdc0a2e04356ca1e437d1ab01b"} Feb 18 09:06:09 crc kubenswrapper[4556]: I0218 09:06:09.518543 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"7cccf6d1-154d-456c-96f5-9af880a6a87b","Type":"ContainerStarted","Data":"b3dc267e123f31d0fef28a6d788abe3851f058d8ddf1cd899b823a18a4b7b4dc"} Feb 18 09:06:09 crc kubenswrapper[4556]: I0218 09:06:09.548664 4556 generic.go:334] "Generic (PLEG): container finished" podID="fdd9dfd4-872b-42fc-960f-5d4c51e6e66e" containerID="071fb0569134209452182b6f81b2cb0454f0201b5b5056bb3fdb2e29f34c768d" exitCode=0 Feb 18 09:06:09 crc kubenswrapper[4556]: I0218 09:06:09.548999 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bwwg2" event={"ID":"fdd9dfd4-872b-42fc-960f-5d4c51e6e66e","Type":"ContainerDied","Data":"071fb0569134209452182b6f81b2cb0454f0201b5b5056bb3fdb2e29f34c768d"} Feb 18 09:06:09 crc kubenswrapper[4556]: I0218 09:06:09.552301 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.55228492 podStartE2EDuration="2.55228492s" podCreationTimestamp="2026-02-18 09:06:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:06:09.551969196 +0000 UTC m=+126.568930177" watchObservedRunningTime="2026-02-18 09:06:09.55228492 +0000 UTC m=+126.569245900" Feb 18 09:06:09 crc kubenswrapper[4556]: I0218 09:06:09.564058 4556 generic.go:334] "Generic (PLEG): container finished" podID="accd4caa-1c67-4430-aa3a-db398194a103" containerID="24d53d9bb4d59e5757a7974f266afa373b85706f9438d5ad3988fd380d8c0312" exitCode=0 Feb 18 09:06:09 crc kubenswrapper[4556]: I0218 09:06:09.564147 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cctvp" event={"ID":"accd4caa-1c67-4430-aa3a-db398194a103","Type":"ContainerDied","Data":"24d53d9bb4d59e5757a7974f266afa373b85706f9438d5ad3988fd380d8c0312"} Feb 18 09:06:09 crc kubenswrapper[4556]: I0218 09:06:09.564194 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cctvp" event={"ID":"accd4caa-1c67-4430-aa3a-db398194a103","Type":"ContainerStarted","Data":"9052dc47a6ad03afebbea1f5a4cf2a9869e0f1e1c0025ef9a9eebddb95f420b0"} Feb 18 09:06:09 crc kubenswrapper[4556]: I0218 09:06:09.572893 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29523420-qb78x" event={"ID":"0aa2e566-fd95-498b-8df5-b63afb290572","Type":"ContainerDied","Data":"f20e52a9326969859ebb286bd8c1bdaaf60692a63dec1c8d4018363660421633"} Feb 18 09:06:09 crc kubenswrapper[4556]: I0218 09:06:09.572927 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f20e52a9326969859ebb286bd8c1bdaaf60692a63dec1c8d4018363660421633" Feb 18 09:06:09 crc kubenswrapper[4556]: I0218 09:06:09.572948 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29523420-qb78x" Feb 18 09:06:09 crc kubenswrapper[4556]: I0218 09:06:09.577453 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-qw4hv" event={"ID":"ccb1af48-7979-4f75-983e-7455ac471966","Type":"ContainerStarted","Data":"a62cbe9996ef1f1f14bb936a636fe69405a3e5f9ccd3e7fa7ecb9faeee9fcf68"} Feb 18 09:06:09 crc kubenswrapper[4556]: I0218 09:06:09.598415 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-wz622"] Feb 18 09:06:09 crc kubenswrapper[4556]: I0218 09:06:09.619465 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-qw4hv" podStartSLOduration=9.619444561 podStartE2EDuration="9.619444561s" podCreationTimestamp="2026-02-18 09:06:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:06:09.61662876 +0000 UTC m=+126.633589740" watchObservedRunningTime="2026-02-18 09:06:09.619444561 +0000 UTC m=+126.636405541" Feb 18 09:06:10 crc kubenswrapper[4556]: I0218 09:06:10.246088 4556 patch_prober.go:28] interesting pod/router-default-5444994796-j9ccg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 18 09:06:10 crc kubenswrapper[4556]: [-]has-synced failed: reason withheld Feb 18 09:06:10 crc kubenswrapper[4556]: [+]process-running ok Feb 18 09:06:10 crc kubenswrapper[4556]: healthz check failed Feb 18 09:06:10 crc kubenswrapper[4556]: I0218 09:06:10.246193 4556 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-j9ccg" podUID="9cf0e7d1-1645-47aa-add2-3a76bfc624a7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 18 09:06:10 crc kubenswrapper[4556]: I0218 09:06:10.596934 4556 generic.go:334] "Generic (PLEG): container finished" podID="7cccf6d1-154d-456c-96f5-9af880a6a87b" containerID="743642d9ce9a0dd4257cf1b2241886e562c638cdc0a2e04356ca1e437d1ab01b" exitCode=0 Feb 18 09:06:10 crc kubenswrapper[4556]: I0218 09:06:10.597646 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"7cccf6d1-154d-456c-96f5-9af880a6a87b","Type":"ContainerDied","Data":"743642d9ce9a0dd4257cf1b2241886e562c638cdc0a2e04356ca1e437d1ab01b"} Feb 18 09:06:10 crc kubenswrapper[4556]: I0218 09:06:10.602042 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-wz622" event={"ID":"709f253c-93c3-4aaf-b3db-630775991aeb","Type":"ContainerStarted","Data":"910086a9ffd35fe9f573814c9d702a1ea8340e0bc824e9adcdc009f52d38ffda"} Feb 18 09:06:10 crc kubenswrapper[4556]: I0218 09:06:10.602080 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-wz622" event={"ID":"709f253c-93c3-4aaf-b3db-630775991aeb","Type":"ContainerStarted","Data":"a9de573ef0d283c2f4720857e5f6a03636743796f09c3e066c47154a783e9a04"} Feb 18 09:06:10 crc kubenswrapper[4556]: I0218 09:06:10.602093 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:06:10 crc kubenswrapper[4556]: I0218 09:06:10.625428 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-wz622" podStartSLOduration=104.625411757 podStartE2EDuration="1m44.625411757s" podCreationTimestamp="2026-02-18 09:04:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:06:10.624334302 +0000 UTC m=+127.641295282" watchObservedRunningTime="2026-02-18 09:06:10.625411757 +0000 UTC m=+127.642372737" Feb 18 09:06:11 crc kubenswrapper[4556]: I0218 09:06:11.240062 4556 patch_prober.go:28] interesting pod/router-default-5444994796-j9ccg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 18 09:06:11 crc kubenswrapper[4556]: [-]has-synced failed: reason withheld Feb 18 09:06:11 crc kubenswrapper[4556]: [+]process-running ok Feb 18 09:06:11 crc kubenswrapper[4556]: healthz check failed Feb 18 09:06:11 crc kubenswrapper[4556]: I0218 09:06:11.240131 4556 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-j9ccg" podUID="9cf0e7d1-1645-47aa-add2-3a76bfc624a7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 18 09:06:11 crc kubenswrapper[4556]: I0218 09:06:11.878738 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 18 09:06:12 crc kubenswrapper[4556]: I0218 09:06:12.044658 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7cccf6d1-154d-456c-96f5-9af880a6a87b-kubelet-dir\") pod \"7cccf6d1-154d-456c-96f5-9af880a6a87b\" (UID: \"7cccf6d1-154d-456c-96f5-9af880a6a87b\") " Feb 18 09:06:12 crc kubenswrapper[4556]: I0218 09:06:12.044722 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7cccf6d1-154d-456c-96f5-9af880a6a87b-kube-api-access\") pod \"7cccf6d1-154d-456c-96f5-9af880a6a87b\" (UID: \"7cccf6d1-154d-456c-96f5-9af880a6a87b\") " Feb 18 09:06:12 crc kubenswrapper[4556]: I0218 09:06:12.044829 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7cccf6d1-154d-456c-96f5-9af880a6a87b-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "7cccf6d1-154d-456c-96f5-9af880a6a87b" (UID: "7cccf6d1-154d-456c-96f5-9af880a6a87b"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 09:06:12 crc kubenswrapper[4556]: I0218 09:06:12.045104 4556 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7cccf6d1-154d-456c-96f5-9af880a6a87b-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 18 09:06:12 crc kubenswrapper[4556]: I0218 09:06:12.051346 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7cccf6d1-154d-456c-96f5-9af880a6a87b-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "7cccf6d1-154d-456c-96f5-9af880a6a87b" (UID: "7cccf6d1-154d-456c-96f5-9af880a6a87b"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:06:12 crc kubenswrapper[4556]: I0218 09:06:12.145988 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7cccf6d1-154d-456c-96f5-9af880a6a87b-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 18 09:06:12 crc kubenswrapper[4556]: I0218 09:06:12.237778 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-j9ccg" Feb 18 09:06:12 crc kubenswrapper[4556]: I0218 09:06:12.242289 4556 patch_prober.go:28] interesting pod/router-default-5444994796-j9ccg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 18 09:06:12 crc kubenswrapper[4556]: [-]has-synced failed: reason withheld Feb 18 09:06:12 crc kubenswrapper[4556]: [+]process-running ok Feb 18 09:06:12 crc kubenswrapper[4556]: healthz check failed Feb 18 09:06:12 crc kubenswrapper[4556]: I0218 09:06:12.242388 4556 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-j9ccg" podUID="9cf0e7d1-1645-47aa-add2-3a76bfc624a7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 18 09:06:12 crc kubenswrapper[4556]: I0218 09:06:12.601015 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-lsngn" Feb 18 09:06:12 crc kubenswrapper[4556]: I0218 09:06:12.605709 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-8gpk5" Feb 18 09:06:12 crc kubenswrapper[4556]: I0218 09:06:12.605753 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-8gpk5" Feb 18 09:06:12 crc kubenswrapper[4556]: I0218 09:06:12.606206 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-lsngn" Feb 18 09:06:12 crc kubenswrapper[4556]: I0218 09:06:12.607667 4556 patch_prober.go:28] interesting pod/console-f9d7485db-8gpk5 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.27:8443/health\": dial tcp 10.217.0.27:8443: connect: connection refused" start-of-body= Feb 18 09:06:12 crc kubenswrapper[4556]: I0218 09:06:12.607720 4556 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-8gpk5" podUID="f34c2f1a-5d5b-4719-82e3-d7534e16f7e6" containerName="console" probeResult="failure" output="Get \"https://10.217.0.27:8443/health\": dial tcp 10.217.0.27:8443: connect: connection refused" Feb 18 09:06:12 crc kubenswrapper[4556]: I0218 09:06:12.681308 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 18 09:06:12 crc kubenswrapper[4556]: I0218 09:06:12.683462 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"7cccf6d1-154d-456c-96f5-9af880a6a87b","Type":"ContainerDied","Data":"b3dc267e123f31d0fef28a6d788abe3851f058d8ddf1cd899b823a18a4b7b4dc"} Feb 18 09:06:12 crc kubenswrapper[4556]: I0218 09:06:12.683525 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b3dc267e123f31d0fef28a6d788abe3851f058d8ddf1cd899b823a18a4b7b4dc" Feb 18 09:06:13 crc kubenswrapper[4556]: I0218 09:06:13.240212 4556 patch_prober.go:28] interesting pod/router-default-5444994796-j9ccg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 18 09:06:13 crc kubenswrapper[4556]: [-]has-synced failed: reason withheld Feb 18 09:06:13 crc kubenswrapper[4556]: [+]process-running ok Feb 18 09:06:13 crc kubenswrapper[4556]: healthz check failed Feb 18 09:06:13 crc kubenswrapper[4556]: I0218 09:06:13.240300 4556 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-j9ccg" podUID="9cf0e7d1-1645-47aa-add2-3a76bfc624a7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 18 09:06:13 crc kubenswrapper[4556]: I0218 09:06:13.781674 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-mvvrj" Feb 18 09:06:14 crc kubenswrapper[4556]: I0218 09:06:14.240972 4556 patch_prober.go:28] interesting pod/router-default-5444994796-j9ccg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 18 09:06:14 crc kubenswrapper[4556]: [-]has-synced failed: reason withheld Feb 18 09:06:14 crc kubenswrapper[4556]: [+]process-running ok Feb 18 09:06:14 crc kubenswrapper[4556]: healthz check failed Feb 18 09:06:14 crc kubenswrapper[4556]: I0218 09:06:14.241497 4556 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-j9ccg" podUID="9cf0e7d1-1645-47aa-add2-3a76bfc624a7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 18 09:06:15 crc kubenswrapper[4556]: I0218 09:06:15.243670 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 18 09:06:15 crc kubenswrapper[4556]: E0218 09:06:15.243857 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cccf6d1-154d-456c-96f5-9af880a6a87b" containerName="pruner" Feb 18 09:06:15 crc kubenswrapper[4556]: I0218 09:06:15.243869 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cccf6d1-154d-456c-96f5-9af880a6a87b" containerName="pruner" Feb 18 09:06:15 crc kubenswrapper[4556]: E0218 09:06:15.243877 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0aa2e566-fd95-498b-8df5-b63afb290572" containerName="collect-profiles" Feb 18 09:06:15 crc kubenswrapper[4556]: I0218 09:06:15.243883 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="0aa2e566-fd95-498b-8df5-b63afb290572" containerName="collect-profiles" Feb 18 09:06:15 crc kubenswrapper[4556]: I0218 09:06:15.243967 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="0aa2e566-fd95-498b-8df5-b63afb290572" containerName="collect-profiles" Feb 18 09:06:15 crc kubenswrapper[4556]: I0218 09:06:15.243978 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="7cccf6d1-154d-456c-96f5-9af880a6a87b" containerName="pruner" Feb 18 09:06:15 crc kubenswrapper[4556]: I0218 09:06:15.248404 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 18 09:06:15 crc kubenswrapper[4556]: I0218 09:06:15.251058 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 18 09:06:15 crc kubenswrapper[4556]: I0218 09:06:15.251376 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 18 09:06:15 crc kubenswrapper[4556]: I0218 09:06:15.251623 4556 patch_prober.go:28] interesting pod/router-default-5444994796-j9ccg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 18 09:06:15 crc kubenswrapper[4556]: [-]has-synced failed: reason withheld Feb 18 09:06:15 crc kubenswrapper[4556]: [+]process-running ok Feb 18 09:06:15 crc kubenswrapper[4556]: healthz check failed Feb 18 09:06:15 crc kubenswrapper[4556]: I0218 09:06:15.251645 4556 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-j9ccg" podUID="9cf0e7d1-1645-47aa-add2-3a76bfc624a7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 18 09:06:15 crc kubenswrapper[4556]: I0218 09:06:15.259178 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 18 09:06:15 crc kubenswrapper[4556]: I0218 09:06:15.339780 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/50482765-d4b6-4bbf-9fd7-6d077e861c98-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"50482765-d4b6-4bbf-9fd7-6d077e861c98\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 18 09:06:15 crc kubenswrapper[4556]: I0218 09:06:15.339840 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/50482765-d4b6-4bbf-9fd7-6d077e861c98-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"50482765-d4b6-4bbf-9fd7-6d077e861c98\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 18 09:06:15 crc kubenswrapper[4556]: I0218 09:06:15.440828 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/50482765-d4b6-4bbf-9fd7-6d077e861c98-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"50482765-d4b6-4bbf-9fd7-6d077e861c98\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 18 09:06:15 crc kubenswrapper[4556]: I0218 09:06:15.440911 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/50482765-d4b6-4bbf-9fd7-6d077e861c98-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"50482765-d4b6-4bbf-9fd7-6d077e861c98\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 18 09:06:15 crc kubenswrapper[4556]: I0218 09:06:15.441011 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/50482765-d4b6-4bbf-9fd7-6d077e861c98-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"50482765-d4b6-4bbf-9fd7-6d077e861c98\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 18 09:06:15 crc kubenswrapper[4556]: I0218 09:06:15.457678 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/50482765-d4b6-4bbf-9fd7-6d077e861c98-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"50482765-d4b6-4bbf-9fd7-6d077e861c98\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 18 09:06:15 crc kubenswrapper[4556]: I0218 09:06:15.575174 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 18 09:06:16 crc kubenswrapper[4556]: I0218 09:06:16.242169 4556 patch_prober.go:28] interesting pod/router-default-5444994796-j9ccg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 18 09:06:16 crc kubenswrapper[4556]: [-]has-synced failed: reason withheld Feb 18 09:06:16 crc kubenswrapper[4556]: [+]process-running ok Feb 18 09:06:16 crc kubenswrapper[4556]: healthz check failed Feb 18 09:06:16 crc kubenswrapper[4556]: I0218 09:06:16.242232 4556 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-j9ccg" podUID="9cf0e7d1-1645-47aa-add2-3a76bfc624a7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 18 09:06:17 crc kubenswrapper[4556]: I0218 09:06:17.240098 4556 patch_prober.go:28] interesting pod/router-default-5444994796-j9ccg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 18 09:06:17 crc kubenswrapper[4556]: [-]has-synced failed: reason withheld Feb 18 09:06:17 crc kubenswrapper[4556]: [+]process-running ok Feb 18 09:06:17 crc kubenswrapper[4556]: healthz check failed Feb 18 09:06:17 crc kubenswrapper[4556]: I0218 09:06:17.240163 4556 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-j9ccg" podUID="9cf0e7d1-1645-47aa-add2-3a76bfc624a7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 18 09:06:18 crc kubenswrapper[4556]: I0218 09:06:18.245218 4556 patch_prober.go:28] interesting pod/router-default-5444994796-j9ccg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 18 09:06:18 crc kubenswrapper[4556]: [-]has-synced failed: reason withheld Feb 18 09:06:18 crc kubenswrapper[4556]: [+]process-running ok Feb 18 09:06:18 crc kubenswrapper[4556]: healthz check failed Feb 18 09:06:18 crc kubenswrapper[4556]: I0218 09:06:18.253385 4556 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-j9ccg" podUID="9cf0e7d1-1645-47aa-add2-3a76bfc624a7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 18 09:06:18 crc kubenswrapper[4556]: I0218 09:06:18.618400 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-zlp6j" Feb 18 09:06:19 crc kubenswrapper[4556]: I0218 09:06:19.241018 4556 patch_prober.go:28] interesting pod/router-default-5444994796-j9ccg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 18 09:06:19 crc kubenswrapper[4556]: [-]has-synced failed: reason withheld Feb 18 09:06:19 crc kubenswrapper[4556]: [+]process-running ok Feb 18 09:06:19 crc kubenswrapper[4556]: healthz check failed Feb 18 09:06:19 crc kubenswrapper[4556]: I0218 09:06:19.241321 4556 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-j9ccg" podUID="9cf0e7d1-1645-47aa-add2-3a76bfc624a7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 18 09:06:19 crc kubenswrapper[4556]: I0218 09:06:19.408868 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 18 09:06:19 crc kubenswrapper[4556]: W0218 09:06:19.432497 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod50482765_d4b6_4bbf_9fd7_6d077e861c98.slice/crio-284041010bd1426706243f9b445525f723a14846c07adcb3e7067a1a9abe2739 WatchSource:0}: Error finding container 284041010bd1426706243f9b445525f723a14846c07adcb3e7067a1a9abe2739: Status 404 returned error can't find the container with id 284041010bd1426706243f9b445525f723a14846c07adcb3e7067a1a9abe2739 Feb 18 09:06:19 crc kubenswrapper[4556]: I0218 09:06:19.740692 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"50482765-d4b6-4bbf-9fd7-6d077e861c98","Type":"ContainerStarted","Data":"284041010bd1426706243f9b445525f723a14846c07adcb3e7067a1a9abe2739"} Feb 18 09:06:20 crc kubenswrapper[4556]: I0218 09:06:20.240727 4556 patch_prober.go:28] interesting pod/router-default-5444994796-j9ccg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 18 09:06:20 crc kubenswrapper[4556]: [-]has-synced failed: reason withheld Feb 18 09:06:20 crc kubenswrapper[4556]: [+]process-running ok Feb 18 09:06:20 crc kubenswrapper[4556]: healthz check failed Feb 18 09:06:20 crc kubenswrapper[4556]: I0218 09:06:20.240775 4556 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-j9ccg" podUID="9cf0e7d1-1645-47aa-add2-3a76bfc624a7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 18 09:06:20 crc kubenswrapper[4556]: I0218 09:06:20.746182 4556 generic.go:334] "Generic (PLEG): container finished" podID="50482765-d4b6-4bbf-9fd7-6d077e861c98" containerID="bf76fa61685f6c0f0a7f35ea542ef8bccb3001b41138b3730e6886f9fb8439d6" exitCode=0 Feb 18 09:06:20 crc kubenswrapper[4556]: I0218 09:06:20.746219 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"50482765-d4b6-4bbf-9fd7-6d077e861c98","Type":"ContainerDied","Data":"bf76fa61685f6c0f0a7f35ea542ef8bccb3001b41138b3730e6886f9fb8439d6"} Feb 18 09:06:21 crc kubenswrapper[4556]: I0218 09:06:21.240008 4556 patch_prober.go:28] interesting pod/router-default-5444994796-j9ccg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 18 09:06:21 crc kubenswrapper[4556]: [-]has-synced failed: reason withheld Feb 18 09:06:21 crc kubenswrapper[4556]: [+]process-running ok Feb 18 09:06:21 crc kubenswrapper[4556]: healthz check failed Feb 18 09:06:21 crc kubenswrapper[4556]: I0218 09:06:21.240074 4556 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-j9ccg" podUID="9cf0e7d1-1645-47aa-add2-3a76bfc624a7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 18 09:06:22 crc kubenswrapper[4556]: I0218 09:06:22.240872 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-j9ccg" Feb 18 09:06:22 crc kubenswrapper[4556]: I0218 09:06:22.243885 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-j9ccg" Feb 18 09:06:22 crc kubenswrapper[4556]: I0218 09:06:22.606831 4556 patch_prober.go:28] interesting pod/console-f9d7485db-8gpk5 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.27:8443/health\": dial tcp 10.217.0.27:8443: connect: connection refused" start-of-body= Feb 18 09:06:22 crc kubenswrapper[4556]: I0218 09:06:22.606878 4556 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-8gpk5" podUID="f34c2f1a-5d5b-4719-82e3-d7534e16f7e6" containerName="console" probeResult="failure" output="Get \"https://10.217.0.27:8443/health\": dial tcp 10.217.0.27:8443: connect: connection refused" Feb 18 09:06:26 crc kubenswrapper[4556]: I0218 09:06:26.934934 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 18 09:06:27 crc kubenswrapper[4556]: I0218 09:06:27.024190 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/50482765-d4b6-4bbf-9fd7-6d077e861c98-kube-api-access\") pod \"50482765-d4b6-4bbf-9fd7-6d077e861c98\" (UID: \"50482765-d4b6-4bbf-9fd7-6d077e861c98\") " Feb 18 09:06:27 crc kubenswrapper[4556]: I0218 09:06:27.024653 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/50482765-d4b6-4bbf-9fd7-6d077e861c98-kubelet-dir\") pod \"50482765-d4b6-4bbf-9fd7-6d077e861c98\" (UID: \"50482765-d4b6-4bbf-9fd7-6d077e861c98\") " Feb 18 09:06:27 crc kubenswrapper[4556]: I0218 09:06:27.025142 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/50482765-d4b6-4bbf-9fd7-6d077e861c98-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "50482765-d4b6-4bbf-9fd7-6d077e861c98" (UID: "50482765-d4b6-4bbf-9fd7-6d077e861c98"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 09:06:27 crc kubenswrapper[4556]: I0218 09:06:27.031057 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50482765-d4b6-4bbf-9fd7-6d077e861c98-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "50482765-d4b6-4bbf-9fd7-6d077e861c98" (UID: "50482765-d4b6-4bbf-9fd7-6d077e861c98"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:06:27 crc kubenswrapper[4556]: I0218 09:06:27.126071 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/50482765-d4b6-4bbf-9fd7-6d077e861c98-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 18 09:06:27 crc kubenswrapper[4556]: I0218 09:06:27.126099 4556 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/50482765-d4b6-4bbf-9fd7-6d077e861c98-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 18 09:06:27 crc kubenswrapper[4556]: I0218 09:06:27.781615 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"50482765-d4b6-4bbf-9fd7-6d077e861c98","Type":"ContainerDied","Data":"284041010bd1426706243f9b445525f723a14846c07adcb3e7067a1a9abe2739"} Feb 18 09:06:27 crc kubenswrapper[4556]: I0218 09:06:27.781660 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="284041010bd1426706243f9b445525f723a14846c07adcb3e7067a1a9abe2739" Feb 18 09:06:27 crc kubenswrapper[4556]: I0218 09:06:27.781679 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 18 09:06:28 crc kubenswrapper[4556]: I0218 09:06:28.946620 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:06:29 crc kubenswrapper[4556]: I0218 09:06:29.261392 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:06:29 crc kubenswrapper[4556]: I0218 09:06:29.261482 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:06:29 crc kubenswrapper[4556]: I0218 09:06:29.263420 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Feb 18 09:06:29 crc kubenswrapper[4556]: I0218 09:06:29.263729 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Feb 18 09:06:29 crc kubenswrapper[4556]: I0218 09:06:29.273639 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:06:29 crc kubenswrapper[4556]: I0218 09:06:29.278273 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:06:29 crc kubenswrapper[4556]: I0218 09:06:29.362265 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:06:29 crc kubenswrapper[4556]: I0218 09:06:29.362359 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:06:29 crc kubenswrapper[4556]: I0218 09:06:29.363590 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Feb 18 09:06:29 crc kubenswrapper[4556]: I0218 09:06:29.375034 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Feb 18 09:06:29 crc kubenswrapper[4556]: I0218 09:06:29.387118 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:06:29 crc kubenswrapper[4556]: I0218 09:06:29.387140 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:06:29 crc kubenswrapper[4556]: I0218 09:06:29.491794 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 09:06:29 crc kubenswrapper[4556]: I0218 09:06:29.498255 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:06:29 crc kubenswrapper[4556]: I0218 09:06:29.503501 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 09:06:29 crc kubenswrapper[4556]: E0218 09:06:29.966677 4556 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Feb 18 09:06:29 crc kubenswrapper[4556]: E0218 09:06:29.966923 4556 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rcdq4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-sd4dx_openshift-marketplace(90477a11-cc7c-48cf-a210-9d1d802360ce): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 18 09:06:29 crc kubenswrapper[4556]: E0218 09:06:29.968030 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-sd4dx" podUID="90477a11-cc7c-48cf-a210-9d1d802360ce" Feb 18 09:06:30 crc kubenswrapper[4556]: W0218 09:06:30.328723 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-b9ceaa3c4374d29e2056aaa0d3c1ce624845f54dbc6a32b3335c5c702482b4ec WatchSource:0}: Error finding container b9ceaa3c4374d29e2056aaa0d3c1ce624845f54dbc6a32b3335c5c702482b4ec: Status 404 returned error can't find the container with id b9ceaa3c4374d29e2056aaa0d3c1ce624845f54dbc6a32b3335c5c702482b4ec Feb 18 09:06:30 crc kubenswrapper[4556]: W0218 09:06:30.329379 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-3838e90c763b525d2df6ea284d10332208e7fc5d36cc9e594bee000be0940c0c WatchSource:0}: Error finding container 3838e90c763b525d2df6ea284d10332208e7fc5d36cc9e594bee000be0940c0c: Status 404 returned error can't find the container with id 3838e90c763b525d2df6ea284d10332208e7fc5d36cc9e594bee000be0940c0c Feb 18 09:06:30 crc kubenswrapper[4556]: W0218 09:06:30.411413 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-ade8cecea7abd6ca5ae2b6af7424264b5353bf57fba55f30df4f1c21920277d3 WatchSource:0}: Error finding container ade8cecea7abd6ca5ae2b6af7424264b5353bf57fba55f30df4f1c21920277d3: Status 404 returned error can't find the container with id ade8cecea7abd6ca5ae2b6af7424264b5353bf57fba55f30df4f1c21920277d3 Feb 18 09:06:30 crc kubenswrapper[4556]: I0218 09:06:30.799401 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"1ad8ab6f8b79bd12805676af3fa80e4e1be11fbeaea7a20274db024017fca8bb"} Feb 18 09:06:30 crc kubenswrapper[4556]: I0218 09:06:30.799731 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"b9ceaa3c4374d29e2056aaa0d3c1ce624845f54dbc6a32b3335c5c702482b4ec"} Feb 18 09:06:30 crc kubenswrapper[4556]: I0218 09:06:30.799912 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:06:30 crc kubenswrapper[4556]: I0218 09:06:30.802848 4556 generic.go:334] "Generic (PLEG): container finished" podID="accd4caa-1c67-4430-aa3a-db398194a103" containerID="9a8ee12e70164c058fa80e3f90b3871596bdba73fc9301d8e6bd435015787c72" exitCode=0 Feb 18 09:06:30 crc kubenswrapper[4556]: I0218 09:06:30.802919 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cctvp" event={"ID":"accd4caa-1c67-4430-aa3a-db398194a103","Type":"ContainerDied","Data":"9a8ee12e70164c058fa80e3f90b3871596bdba73fc9301d8e6bd435015787c72"} Feb 18 09:06:30 crc kubenswrapper[4556]: I0218 09:06:30.804987 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"25321c3dd8df2b06abdfb27c361b2b659422da833aae9e7a33693760f1a935f0"} Feb 18 09:06:30 crc kubenswrapper[4556]: I0218 09:06:30.805044 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"3838e90c763b525d2df6ea284d10332208e7fc5d36cc9e594bee000be0940c0c"} Feb 18 09:06:30 crc kubenswrapper[4556]: I0218 09:06:30.809044 4556 generic.go:334] "Generic (PLEG): container finished" podID="e23bd1bf-79ec-4cc0-95c4-16959eab3aa7" containerID="737dda86681f77545e2f061e569a8489f2e9b178bdbffe9bc502e44f548eea16" exitCode=0 Feb 18 09:06:30 crc kubenswrapper[4556]: I0218 09:06:30.809129 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ndnpr" event={"ID":"e23bd1bf-79ec-4cc0-95c4-16959eab3aa7","Type":"ContainerDied","Data":"737dda86681f77545e2f061e569a8489f2e9b178bdbffe9bc502e44f548eea16"} Feb 18 09:06:30 crc kubenswrapper[4556]: I0218 09:06:30.814119 4556 generic.go:334] "Generic (PLEG): container finished" podID="7875d007-bdbc-428a-b291-283e8a1cdad0" containerID="95a69f4487fff8166687ccd402c34aaccdea04bb52085712cc69bc783ad2fbef" exitCode=0 Feb 18 09:06:30 crc kubenswrapper[4556]: I0218 09:06:30.814190 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ss8sl" event={"ID":"7875d007-bdbc-428a-b291-283e8a1cdad0","Type":"ContainerDied","Data":"95a69f4487fff8166687ccd402c34aaccdea04bb52085712cc69bc783ad2fbef"} Feb 18 09:06:30 crc kubenswrapper[4556]: I0218 09:06:30.818052 4556 generic.go:334] "Generic (PLEG): container finished" podID="bc2d3f5c-f685-43ea-bf13-4bec129bf889" containerID="9da730c32a7c60d54821313f40e255e0f0f65429a94f934648a343466930518c" exitCode=0 Feb 18 09:06:30 crc kubenswrapper[4556]: I0218 09:06:30.818114 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ntvpp" event={"ID":"bc2d3f5c-f685-43ea-bf13-4bec129bf889","Type":"ContainerDied","Data":"9da730c32a7c60d54821313f40e255e0f0f65429a94f934648a343466930518c"} Feb 18 09:06:30 crc kubenswrapper[4556]: I0218 09:06:30.820898 4556 generic.go:334] "Generic (PLEG): container finished" podID="eac84811-d54a-47c2-ae5e-77f3a531fc4f" containerID="0fce5558b046231804fca44b8ec9931210abf6bfc754bd3033f4e3a85819c0f1" exitCode=0 Feb 18 09:06:30 crc kubenswrapper[4556]: I0218 09:06:30.820943 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-brf28" event={"ID":"eac84811-d54a-47c2-ae5e-77f3a531fc4f","Type":"ContainerDied","Data":"0fce5558b046231804fca44b8ec9931210abf6bfc754bd3033f4e3a85819c0f1"} Feb 18 09:06:30 crc kubenswrapper[4556]: I0218 09:06:30.826597 4556 generic.go:334] "Generic (PLEG): container finished" podID="25cebcb5-2e64-4390-9d61-f06607eee280" containerID="64dca3deba8f1f18af92dc864b2f5fbb5b624d3afce3d18f9dab3d6f1e538b26" exitCode=0 Feb 18 09:06:30 crc kubenswrapper[4556]: I0218 09:06:30.826664 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wqmhk" event={"ID":"25cebcb5-2e64-4390-9d61-f06607eee280","Type":"ContainerDied","Data":"64dca3deba8f1f18af92dc864b2f5fbb5b624d3afce3d18f9dab3d6f1e538b26"} Feb 18 09:06:30 crc kubenswrapper[4556]: I0218 09:06:30.832669 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bwwg2" event={"ID":"fdd9dfd4-872b-42fc-960f-5d4c51e6e66e","Type":"ContainerDied","Data":"b1e538edcfe7d69af3ccabf373f71ff4f59d9eed02244cc39e055c7904706da5"} Feb 18 09:06:30 crc kubenswrapper[4556]: I0218 09:06:30.832808 4556 generic.go:334] "Generic (PLEG): container finished" podID="fdd9dfd4-872b-42fc-960f-5d4c51e6e66e" containerID="b1e538edcfe7d69af3ccabf373f71ff4f59d9eed02244cc39e055c7904706da5" exitCode=0 Feb 18 09:06:30 crc kubenswrapper[4556]: I0218 09:06:30.834733 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"85918edf2b23f949d40e103034cb00e755e8e9f0556b456a97f941354b2dfd8a"} Feb 18 09:06:30 crc kubenswrapper[4556]: I0218 09:06:30.834827 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"ade8cecea7abd6ca5ae2b6af7424264b5353bf57fba55f30df4f1c21920277d3"} Feb 18 09:06:30 crc kubenswrapper[4556]: E0218 09:06:30.838069 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-sd4dx" podUID="90477a11-cc7c-48cf-a210-9d1d802360ce" Feb 18 09:06:31 crc kubenswrapper[4556]: I0218 09:06:31.727974 4556 patch_prober.go:28] interesting pod/machine-config-daemon-f76hs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 09:06:31 crc kubenswrapper[4556]: I0218 09:06:31.728668 4556 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 09:06:31 crc kubenswrapper[4556]: I0218 09:06:31.841477 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wqmhk" event={"ID":"25cebcb5-2e64-4390-9d61-f06607eee280","Type":"ContainerStarted","Data":"bc03719a50264fe8b32082004059ef994cf5a497fc0943e571729c8c1de412d4"} Feb 18 09:06:31 crc kubenswrapper[4556]: I0218 09:06:31.842999 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bwwg2" event={"ID":"fdd9dfd4-872b-42fc-960f-5d4c51e6e66e","Type":"ContainerStarted","Data":"aeff40b439e6e8751290c8681fc2c889491f6e822c6c530d08c1d651305a72f6"} Feb 18 09:06:31 crc kubenswrapper[4556]: I0218 09:06:31.844577 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cctvp" event={"ID":"accd4caa-1c67-4430-aa3a-db398194a103","Type":"ContainerStarted","Data":"3e8a16f1d3647125ea2809825d88ccaa93b3dc1ae82e8fb0e11121cf3e6f5164"} Feb 18 09:06:31 crc kubenswrapper[4556]: I0218 09:06:31.845992 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ndnpr" event={"ID":"e23bd1bf-79ec-4cc0-95c4-16959eab3aa7","Type":"ContainerStarted","Data":"2b7ded25b1a59a7579dc6d99250e77bc4deae52ae4a074b1393c9ac3324ab332"} Feb 18 09:06:31 crc kubenswrapper[4556]: I0218 09:06:31.848143 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ntvpp" event={"ID":"bc2d3f5c-f685-43ea-bf13-4bec129bf889","Type":"ContainerStarted","Data":"a0ca9a87c9be3946c62e8774eec8e01501df09a41c668b77946cb9f9423257a6"} Feb 18 09:06:31 crc kubenswrapper[4556]: I0218 09:06:31.849891 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ss8sl" event={"ID":"7875d007-bdbc-428a-b291-283e8a1cdad0","Type":"ContainerStarted","Data":"43dcd4a4cacc561cfe972cbc840cff9d82992908726b740e6971a69a2dd195ba"} Feb 18 09:06:31 crc kubenswrapper[4556]: I0218 09:06:31.851647 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-brf28" event={"ID":"eac84811-d54a-47c2-ae5e-77f3a531fc4f","Type":"ContainerStarted","Data":"5bb7c88c3a0aaf429e291c99988ed6c53f910bd094b7e378bae706c0b21db144"} Feb 18 09:06:31 crc kubenswrapper[4556]: I0218 09:06:31.861271 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-wqmhk" podStartSLOduration=3.42539745 podStartE2EDuration="27.861257181s" podCreationTimestamp="2026-02-18 09:06:04 +0000 UTC" firstStartedPulling="2026-02-18 09:06:07.253134642 +0000 UTC m=+124.270095622" lastFinishedPulling="2026-02-18 09:06:31.688994374 +0000 UTC m=+148.705955353" observedRunningTime="2026-02-18 09:06:31.859868594 +0000 UTC m=+148.876829575" watchObservedRunningTime="2026-02-18 09:06:31.861257181 +0000 UTC m=+148.878218161" Feb 18 09:06:31 crc kubenswrapper[4556]: I0218 09:06:31.878940 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-ss8sl" podStartSLOduration=3.7556986390000002 podStartE2EDuration="27.878927307s" podCreationTimestamp="2026-02-18 09:06:04 +0000 UTC" firstStartedPulling="2026-02-18 09:06:07.317228353 +0000 UTC m=+124.334189333" lastFinishedPulling="2026-02-18 09:06:31.440457022 +0000 UTC m=+148.457418001" observedRunningTime="2026-02-18 09:06:31.878333819 +0000 UTC m=+148.895294799" watchObservedRunningTime="2026-02-18 09:06:31.878927307 +0000 UTC m=+148.895888287" Feb 18 09:06:31 crc kubenswrapper[4556]: I0218 09:06:31.890732 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-cctvp" podStartSLOduration=3.008565431 podStartE2EDuration="24.890718589s" podCreationTimestamp="2026-02-18 09:06:07 +0000 UTC" firstStartedPulling="2026-02-18 09:06:09.574967754 +0000 UTC m=+126.591928734" lastFinishedPulling="2026-02-18 09:06:31.457120912 +0000 UTC m=+148.474081892" observedRunningTime="2026-02-18 09:06:31.889637834 +0000 UTC m=+148.906598814" watchObservedRunningTime="2026-02-18 09:06:31.890718589 +0000 UTC m=+148.907679570" Feb 18 09:06:31 crc kubenswrapper[4556]: I0218 09:06:31.938795 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-ntvpp" podStartSLOduration=3.032744921 podStartE2EDuration="25.938782542s" podCreationTimestamp="2026-02-18 09:06:06 +0000 UTC" firstStartedPulling="2026-02-18 09:06:08.473389151 +0000 UTC m=+125.490350122" lastFinishedPulling="2026-02-18 09:06:31.379426763 +0000 UTC m=+148.396387743" observedRunningTime="2026-02-18 09:06:31.914375173 +0000 UTC m=+148.931336153" watchObservedRunningTime="2026-02-18 09:06:31.938782542 +0000 UTC m=+148.955743522" Feb 18 09:06:31 crc kubenswrapper[4556]: I0218 09:06:31.953123 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bwwg2" podStartSLOduration=3.208391056 podStartE2EDuration="24.953107949s" podCreationTimestamp="2026-02-18 09:06:07 +0000 UTC" firstStartedPulling="2026-02-18 09:06:09.556886065 +0000 UTC m=+126.573847046" lastFinishedPulling="2026-02-18 09:06:31.301602959 +0000 UTC m=+148.318563939" observedRunningTime="2026-02-18 09:06:31.939327829 +0000 UTC m=+148.956288809" watchObservedRunningTime="2026-02-18 09:06:31.953107949 +0000 UTC m=+148.970068928" Feb 18 09:06:31 crc kubenswrapper[4556]: I0218 09:06:31.953895 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-brf28" podStartSLOduration=3.990397434 podStartE2EDuration="27.953888859s" podCreationTimestamp="2026-02-18 09:06:04 +0000 UTC" firstStartedPulling="2026-02-18 09:06:07.348956965 +0000 UTC m=+124.365917945" lastFinishedPulling="2026-02-18 09:06:31.312448389 +0000 UTC m=+148.329409370" observedRunningTime="2026-02-18 09:06:31.950872597 +0000 UTC m=+148.967833577" watchObservedRunningTime="2026-02-18 09:06:31.953888859 +0000 UTC m=+148.970849839" Feb 18 09:06:31 crc kubenswrapper[4556]: I0218 09:06:31.965726 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-ndnpr" podStartSLOduration=3.138081827 podStartE2EDuration="25.965716079s" podCreationTimestamp="2026-02-18 09:06:06 +0000 UTC" firstStartedPulling="2026-02-18 09:06:08.46611891 +0000 UTC m=+125.483079890" lastFinishedPulling="2026-02-18 09:06:31.293753162 +0000 UTC m=+148.310714142" observedRunningTime="2026-02-18 09:06:31.962905314 +0000 UTC m=+148.979866294" watchObservedRunningTime="2026-02-18 09:06:31.965716079 +0000 UTC m=+148.982677059" Feb 18 09:06:32 crc kubenswrapper[4556]: I0218 09:06:32.609327 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-8gpk5" Feb 18 09:06:32 crc kubenswrapper[4556]: I0218 09:06:32.612208 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-8gpk5" Feb 18 09:06:34 crc kubenswrapper[4556]: I0218 09:06:34.839036 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-brf28" Feb 18 09:06:34 crc kubenswrapper[4556]: I0218 09:06:34.839404 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-brf28" Feb 18 09:06:34 crc kubenswrapper[4556]: I0218 09:06:34.921464 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-brf28" Feb 18 09:06:35 crc kubenswrapper[4556]: I0218 09:06:35.115043 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-ss8sl" Feb 18 09:06:35 crc kubenswrapper[4556]: I0218 09:06:35.115110 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-ss8sl" Feb 18 09:06:35 crc kubenswrapper[4556]: I0218 09:06:35.145325 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-ss8sl" Feb 18 09:06:35 crc kubenswrapper[4556]: I0218 09:06:35.359622 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-wqmhk" Feb 18 09:06:35 crc kubenswrapper[4556]: I0218 09:06:35.359660 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-wqmhk" Feb 18 09:06:35 crc kubenswrapper[4556]: I0218 09:06:35.387787 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-wqmhk" Feb 18 09:06:36 crc kubenswrapper[4556]: I0218 09:06:36.806656 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-ndnpr" Feb 18 09:06:36 crc kubenswrapper[4556]: I0218 09:06:36.807009 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-ndnpr" Feb 18 09:06:36 crc kubenswrapper[4556]: I0218 09:06:36.837466 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-ndnpr" Feb 18 09:06:36 crc kubenswrapper[4556]: I0218 09:06:36.901586 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-ndnpr" Feb 18 09:06:37 crc kubenswrapper[4556]: I0218 09:06:37.096203 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-ntvpp" Feb 18 09:06:37 crc kubenswrapper[4556]: I0218 09:06:37.096265 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-ntvpp" Feb 18 09:06:37 crc kubenswrapper[4556]: I0218 09:06:37.117939 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-vxdmn"] Feb 18 09:06:37 crc kubenswrapper[4556]: I0218 09:06:37.149898 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-ntvpp" Feb 18 09:06:37 crc kubenswrapper[4556]: I0218 09:06:37.910410 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-ntvpp" Feb 18 09:06:38 crc kubenswrapper[4556]: I0218 09:06:38.000624 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bwwg2" Feb 18 09:06:38 crc kubenswrapper[4556]: I0218 09:06:38.000869 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bwwg2" Feb 18 09:06:38 crc kubenswrapper[4556]: I0218 09:06:38.029985 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bwwg2" Feb 18 09:06:38 crc kubenswrapper[4556]: I0218 09:06:38.292292 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-cctvp" Feb 18 09:06:38 crc kubenswrapper[4556]: I0218 09:06:38.292397 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-cctvp" Feb 18 09:06:38 crc kubenswrapper[4556]: I0218 09:06:38.322924 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-cctvp" Feb 18 09:06:38 crc kubenswrapper[4556]: I0218 09:06:38.916358 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bwwg2" Feb 18 09:06:38 crc kubenswrapper[4556]: I0218 09:06:38.917760 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-cctvp" Feb 18 09:06:39 crc kubenswrapper[4556]: I0218 09:06:39.254060 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ntvpp"] Feb 18 09:06:39 crc kubenswrapper[4556]: I0218 09:06:39.892599 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-ntvpp" podUID="bc2d3f5c-f685-43ea-bf13-4bec129bf889" containerName="registry-server" containerID="cri-o://a0ca9a87c9be3946c62e8774eec8e01501df09a41c668b77946cb9f9423257a6" gracePeriod=2 Feb 18 09:06:40 crc kubenswrapper[4556]: I0218 09:06:40.192762 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ntvpp" Feb 18 09:06:40 crc kubenswrapper[4556]: I0218 09:06:40.331223 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc2d3f5c-f685-43ea-bf13-4bec129bf889-catalog-content\") pod \"bc2d3f5c-f685-43ea-bf13-4bec129bf889\" (UID: \"bc2d3f5c-f685-43ea-bf13-4bec129bf889\") " Feb 18 09:06:40 crc kubenswrapper[4556]: I0218 09:06:40.331403 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nbfxx\" (UniqueName: \"kubernetes.io/projected/bc2d3f5c-f685-43ea-bf13-4bec129bf889-kube-api-access-nbfxx\") pod \"bc2d3f5c-f685-43ea-bf13-4bec129bf889\" (UID: \"bc2d3f5c-f685-43ea-bf13-4bec129bf889\") " Feb 18 09:06:40 crc kubenswrapper[4556]: I0218 09:06:40.331444 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc2d3f5c-f685-43ea-bf13-4bec129bf889-utilities\") pod \"bc2d3f5c-f685-43ea-bf13-4bec129bf889\" (UID: \"bc2d3f5c-f685-43ea-bf13-4bec129bf889\") " Feb 18 09:06:40 crc kubenswrapper[4556]: I0218 09:06:40.332179 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc2d3f5c-f685-43ea-bf13-4bec129bf889-utilities" (OuterVolumeSpecName: "utilities") pod "bc2d3f5c-f685-43ea-bf13-4bec129bf889" (UID: "bc2d3f5c-f685-43ea-bf13-4bec129bf889"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:06:40 crc kubenswrapper[4556]: I0218 09:06:40.337003 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc2d3f5c-f685-43ea-bf13-4bec129bf889-kube-api-access-nbfxx" (OuterVolumeSpecName: "kube-api-access-nbfxx") pod "bc2d3f5c-f685-43ea-bf13-4bec129bf889" (UID: "bc2d3f5c-f685-43ea-bf13-4bec129bf889"). InnerVolumeSpecName "kube-api-access-nbfxx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:06:40 crc kubenswrapper[4556]: I0218 09:06:40.350785 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc2d3f5c-f685-43ea-bf13-4bec129bf889-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bc2d3f5c-f685-43ea-bf13-4bec129bf889" (UID: "bc2d3f5c-f685-43ea-bf13-4bec129bf889"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:06:40 crc kubenswrapper[4556]: I0218 09:06:40.433437 4556 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc2d3f5c-f685-43ea-bf13-4bec129bf889-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 09:06:40 crc kubenswrapper[4556]: I0218 09:06:40.433472 4556 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc2d3f5c-f685-43ea-bf13-4bec129bf889-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 09:06:40 crc kubenswrapper[4556]: I0218 09:06:40.433487 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nbfxx\" (UniqueName: \"kubernetes.io/projected/bc2d3f5c-f685-43ea-bf13-4bec129bf889-kube-api-access-nbfxx\") on node \"crc\" DevicePath \"\"" Feb 18 09:06:40 crc kubenswrapper[4556]: I0218 09:06:40.900082 4556 generic.go:334] "Generic (PLEG): container finished" podID="bc2d3f5c-f685-43ea-bf13-4bec129bf889" containerID="a0ca9a87c9be3946c62e8774eec8e01501df09a41c668b77946cb9f9423257a6" exitCode=0 Feb 18 09:06:40 crc kubenswrapper[4556]: I0218 09:06:40.900476 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ntvpp" event={"ID":"bc2d3f5c-f685-43ea-bf13-4bec129bf889","Type":"ContainerDied","Data":"a0ca9a87c9be3946c62e8774eec8e01501df09a41c668b77946cb9f9423257a6"} Feb 18 09:06:40 crc kubenswrapper[4556]: I0218 09:06:40.900548 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ntvpp" event={"ID":"bc2d3f5c-f685-43ea-bf13-4bec129bf889","Type":"ContainerDied","Data":"3adc525ce621d004dcc2338e52a240b04c5fc126a1af2ebd25f5eeae1d3421dd"} Feb 18 09:06:40 crc kubenswrapper[4556]: I0218 09:06:40.900567 4556 scope.go:117] "RemoveContainer" containerID="a0ca9a87c9be3946c62e8774eec8e01501df09a41c668b77946cb9f9423257a6" Feb 18 09:06:40 crc kubenswrapper[4556]: I0218 09:06:40.900802 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ntvpp" Feb 18 09:06:40 crc kubenswrapper[4556]: I0218 09:06:40.915071 4556 scope.go:117] "RemoveContainer" containerID="9da730c32a7c60d54821313f40e255e0f0f65429a94f934648a343466930518c" Feb 18 09:06:40 crc kubenswrapper[4556]: I0218 09:06:40.922273 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ntvpp"] Feb 18 09:06:40 crc kubenswrapper[4556]: I0218 09:06:40.925661 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-ntvpp"] Feb 18 09:06:40 crc kubenswrapper[4556]: I0218 09:06:40.942836 4556 scope.go:117] "RemoveContainer" containerID="27e59ca8d3945d13facd4466d769159dd0dd7ece817888020cefdb60ad599b24" Feb 18 09:06:40 crc kubenswrapper[4556]: I0218 09:06:40.956400 4556 scope.go:117] "RemoveContainer" containerID="a0ca9a87c9be3946c62e8774eec8e01501df09a41c668b77946cb9f9423257a6" Feb 18 09:06:40 crc kubenswrapper[4556]: E0218 09:06:40.956780 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0ca9a87c9be3946c62e8774eec8e01501df09a41c668b77946cb9f9423257a6\": container with ID starting with a0ca9a87c9be3946c62e8774eec8e01501df09a41c668b77946cb9f9423257a6 not found: ID does not exist" containerID="a0ca9a87c9be3946c62e8774eec8e01501df09a41c668b77946cb9f9423257a6" Feb 18 09:06:40 crc kubenswrapper[4556]: I0218 09:06:40.956814 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0ca9a87c9be3946c62e8774eec8e01501df09a41c668b77946cb9f9423257a6"} err="failed to get container status \"a0ca9a87c9be3946c62e8774eec8e01501df09a41c668b77946cb9f9423257a6\": rpc error: code = NotFound desc = could not find container \"a0ca9a87c9be3946c62e8774eec8e01501df09a41c668b77946cb9f9423257a6\": container with ID starting with a0ca9a87c9be3946c62e8774eec8e01501df09a41c668b77946cb9f9423257a6 not found: ID does not exist" Feb 18 09:06:40 crc kubenswrapper[4556]: I0218 09:06:40.956852 4556 scope.go:117] "RemoveContainer" containerID="9da730c32a7c60d54821313f40e255e0f0f65429a94f934648a343466930518c" Feb 18 09:06:40 crc kubenswrapper[4556]: E0218 09:06:40.957119 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9da730c32a7c60d54821313f40e255e0f0f65429a94f934648a343466930518c\": container with ID starting with 9da730c32a7c60d54821313f40e255e0f0f65429a94f934648a343466930518c not found: ID does not exist" containerID="9da730c32a7c60d54821313f40e255e0f0f65429a94f934648a343466930518c" Feb 18 09:06:40 crc kubenswrapper[4556]: I0218 09:06:40.957168 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9da730c32a7c60d54821313f40e255e0f0f65429a94f934648a343466930518c"} err="failed to get container status \"9da730c32a7c60d54821313f40e255e0f0f65429a94f934648a343466930518c\": rpc error: code = NotFound desc = could not find container \"9da730c32a7c60d54821313f40e255e0f0f65429a94f934648a343466930518c\": container with ID starting with 9da730c32a7c60d54821313f40e255e0f0f65429a94f934648a343466930518c not found: ID does not exist" Feb 18 09:06:40 crc kubenswrapper[4556]: I0218 09:06:40.957195 4556 scope.go:117] "RemoveContainer" containerID="27e59ca8d3945d13facd4466d769159dd0dd7ece817888020cefdb60ad599b24" Feb 18 09:06:40 crc kubenswrapper[4556]: E0218 09:06:40.957576 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27e59ca8d3945d13facd4466d769159dd0dd7ece817888020cefdb60ad599b24\": container with ID starting with 27e59ca8d3945d13facd4466d769159dd0dd7ece817888020cefdb60ad599b24 not found: ID does not exist" containerID="27e59ca8d3945d13facd4466d769159dd0dd7ece817888020cefdb60ad599b24" Feb 18 09:06:40 crc kubenswrapper[4556]: I0218 09:06:40.957619 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27e59ca8d3945d13facd4466d769159dd0dd7ece817888020cefdb60ad599b24"} err="failed to get container status \"27e59ca8d3945d13facd4466d769159dd0dd7ece817888020cefdb60ad599b24\": rpc error: code = NotFound desc = could not find container \"27e59ca8d3945d13facd4466d769159dd0dd7ece817888020cefdb60ad599b24\": container with ID starting with 27e59ca8d3945d13facd4466d769159dd0dd7ece817888020cefdb60ad599b24 not found: ID does not exist" Feb 18 09:06:41 crc kubenswrapper[4556]: I0218 09:06:41.288200 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc2d3f5c-f685-43ea-bf13-4bec129bf889" path="/var/lib/kubelet/pods/bc2d3f5c-f685-43ea-bf13-4bec129bf889/volumes" Feb 18 09:06:41 crc kubenswrapper[4556]: I0218 09:06:41.652313 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cctvp"] Feb 18 09:06:41 crc kubenswrapper[4556]: I0218 09:06:41.906393 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-cctvp" podUID="accd4caa-1c67-4430-aa3a-db398194a103" containerName="registry-server" containerID="cri-o://3e8a16f1d3647125ea2809825d88ccaa93b3dc1ae82e8fb0e11121cf3e6f5164" gracePeriod=2 Feb 18 09:06:42 crc kubenswrapper[4556]: I0218 09:06:42.187531 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cctvp" Feb 18 09:06:42 crc kubenswrapper[4556]: I0218 09:06:42.253645 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/accd4caa-1c67-4430-aa3a-db398194a103-catalog-content\") pod \"accd4caa-1c67-4430-aa3a-db398194a103\" (UID: \"accd4caa-1c67-4430-aa3a-db398194a103\") " Feb 18 09:06:42 crc kubenswrapper[4556]: I0218 09:06:42.253689 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/accd4caa-1c67-4430-aa3a-db398194a103-utilities\") pod \"accd4caa-1c67-4430-aa3a-db398194a103\" (UID: \"accd4caa-1c67-4430-aa3a-db398194a103\") " Feb 18 09:06:42 crc kubenswrapper[4556]: I0218 09:06:42.253712 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tjssn\" (UniqueName: \"kubernetes.io/projected/accd4caa-1c67-4430-aa3a-db398194a103-kube-api-access-tjssn\") pod \"accd4caa-1c67-4430-aa3a-db398194a103\" (UID: \"accd4caa-1c67-4430-aa3a-db398194a103\") " Feb 18 09:06:42 crc kubenswrapper[4556]: I0218 09:06:42.255013 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/accd4caa-1c67-4430-aa3a-db398194a103-utilities" (OuterVolumeSpecName: "utilities") pod "accd4caa-1c67-4430-aa3a-db398194a103" (UID: "accd4caa-1c67-4430-aa3a-db398194a103"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:06:42 crc kubenswrapper[4556]: I0218 09:06:42.257464 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/accd4caa-1c67-4430-aa3a-db398194a103-kube-api-access-tjssn" (OuterVolumeSpecName: "kube-api-access-tjssn") pod "accd4caa-1c67-4430-aa3a-db398194a103" (UID: "accd4caa-1c67-4430-aa3a-db398194a103"). InnerVolumeSpecName "kube-api-access-tjssn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:06:42 crc kubenswrapper[4556]: I0218 09:06:42.269907 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r44z2" Feb 18 09:06:42 crc kubenswrapper[4556]: I0218 09:06:42.344671 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/accd4caa-1c67-4430-aa3a-db398194a103-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "accd4caa-1c67-4430-aa3a-db398194a103" (UID: "accd4caa-1c67-4430-aa3a-db398194a103"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:06:42 crc kubenswrapper[4556]: I0218 09:06:42.356072 4556 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/accd4caa-1c67-4430-aa3a-db398194a103-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 09:06:42 crc kubenswrapper[4556]: I0218 09:06:42.356100 4556 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/accd4caa-1c67-4430-aa3a-db398194a103-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 09:06:42 crc kubenswrapper[4556]: I0218 09:06:42.356110 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tjssn\" (UniqueName: \"kubernetes.io/projected/accd4caa-1c67-4430-aa3a-db398194a103-kube-api-access-tjssn\") on node \"crc\" DevicePath \"\"" Feb 18 09:06:42 crc kubenswrapper[4556]: I0218 09:06:42.912515 4556 generic.go:334] "Generic (PLEG): container finished" podID="accd4caa-1c67-4430-aa3a-db398194a103" containerID="3e8a16f1d3647125ea2809825d88ccaa93b3dc1ae82e8fb0e11121cf3e6f5164" exitCode=0 Feb 18 09:06:42 crc kubenswrapper[4556]: I0218 09:06:42.912574 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cctvp" Feb 18 09:06:42 crc kubenswrapper[4556]: I0218 09:06:42.912784 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cctvp" event={"ID":"accd4caa-1c67-4430-aa3a-db398194a103","Type":"ContainerDied","Data":"3e8a16f1d3647125ea2809825d88ccaa93b3dc1ae82e8fb0e11121cf3e6f5164"} Feb 18 09:06:42 crc kubenswrapper[4556]: I0218 09:06:42.912876 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cctvp" event={"ID":"accd4caa-1c67-4430-aa3a-db398194a103","Type":"ContainerDied","Data":"9052dc47a6ad03afebbea1f5a4cf2a9869e0f1e1c0025ef9a9eebddb95f420b0"} Feb 18 09:06:42 crc kubenswrapper[4556]: I0218 09:06:42.912944 4556 scope.go:117] "RemoveContainer" containerID="3e8a16f1d3647125ea2809825d88ccaa93b3dc1ae82e8fb0e11121cf3e6f5164" Feb 18 09:06:42 crc kubenswrapper[4556]: I0218 09:06:42.929517 4556 scope.go:117] "RemoveContainer" containerID="9a8ee12e70164c058fa80e3f90b3871596bdba73fc9301d8e6bd435015787c72" Feb 18 09:06:42 crc kubenswrapper[4556]: I0218 09:06:42.938346 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cctvp"] Feb 18 09:06:42 crc kubenswrapper[4556]: I0218 09:06:42.940748 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-cctvp"] Feb 18 09:06:42 crc kubenswrapper[4556]: I0218 09:06:42.964051 4556 scope.go:117] "RemoveContainer" containerID="24d53d9bb4d59e5757a7974f266afa373b85706f9438d5ad3988fd380d8c0312" Feb 18 09:06:42 crc kubenswrapper[4556]: I0218 09:06:42.979001 4556 scope.go:117] "RemoveContainer" containerID="3e8a16f1d3647125ea2809825d88ccaa93b3dc1ae82e8fb0e11121cf3e6f5164" Feb 18 09:06:42 crc kubenswrapper[4556]: E0218 09:06:42.979336 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e8a16f1d3647125ea2809825d88ccaa93b3dc1ae82e8fb0e11121cf3e6f5164\": container with ID starting with 3e8a16f1d3647125ea2809825d88ccaa93b3dc1ae82e8fb0e11121cf3e6f5164 not found: ID does not exist" containerID="3e8a16f1d3647125ea2809825d88ccaa93b3dc1ae82e8fb0e11121cf3e6f5164" Feb 18 09:06:42 crc kubenswrapper[4556]: I0218 09:06:42.979376 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e8a16f1d3647125ea2809825d88ccaa93b3dc1ae82e8fb0e11121cf3e6f5164"} err="failed to get container status \"3e8a16f1d3647125ea2809825d88ccaa93b3dc1ae82e8fb0e11121cf3e6f5164\": rpc error: code = NotFound desc = could not find container \"3e8a16f1d3647125ea2809825d88ccaa93b3dc1ae82e8fb0e11121cf3e6f5164\": container with ID starting with 3e8a16f1d3647125ea2809825d88ccaa93b3dc1ae82e8fb0e11121cf3e6f5164 not found: ID does not exist" Feb 18 09:06:42 crc kubenswrapper[4556]: I0218 09:06:42.979401 4556 scope.go:117] "RemoveContainer" containerID="9a8ee12e70164c058fa80e3f90b3871596bdba73fc9301d8e6bd435015787c72" Feb 18 09:06:42 crc kubenswrapper[4556]: E0218 09:06:42.979707 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a8ee12e70164c058fa80e3f90b3871596bdba73fc9301d8e6bd435015787c72\": container with ID starting with 9a8ee12e70164c058fa80e3f90b3871596bdba73fc9301d8e6bd435015787c72 not found: ID does not exist" containerID="9a8ee12e70164c058fa80e3f90b3871596bdba73fc9301d8e6bd435015787c72" Feb 18 09:06:42 crc kubenswrapper[4556]: I0218 09:06:42.979750 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a8ee12e70164c058fa80e3f90b3871596bdba73fc9301d8e6bd435015787c72"} err="failed to get container status \"9a8ee12e70164c058fa80e3f90b3871596bdba73fc9301d8e6bd435015787c72\": rpc error: code = NotFound desc = could not find container \"9a8ee12e70164c058fa80e3f90b3871596bdba73fc9301d8e6bd435015787c72\": container with ID starting with 9a8ee12e70164c058fa80e3f90b3871596bdba73fc9301d8e6bd435015787c72 not found: ID does not exist" Feb 18 09:06:42 crc kubenswrapper[4556]: I0218 09:06:42.979779 4556 scope.go:117] "RemoveContainer" containerID="24d53d9bb4d59e5757a7974f266afa373b85706f9438d5ad3988fd380d8c0312" Feb 18 09:06:42 crc kubenswrapper[4556]: E0218 09:06:42.980056 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24d53d9bb4d59e5757a7974f266afa373b85706f9438d5ad3988fd380d8c0312\": container with ID starting with 24d53d9bb4d59e5757a7974f266afa373b85706f9438d5ad3988fd380d8c0312 not found: ID does not exist" containerID="24d53d9bb4d59e5757a7974f266afa373b85706f9438d5ad3988fd380d8c0312" Feb 18 09:06:42 crc kubenswrapper[4556]: I0218 09:06:42.980081 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24d53d9bb4d59e5757a7974f266afa373b85706f9438d5ad3988fd380d8c0312"} err="failed to get container status \"24d53d9bb4d59e5757a7974f266afa373b85706f9438d5ad3988fd380d8c0312\": rpc error: code = NotFound desc = could not find container \"24d53d9bb4d59e5757a7974f266afa373b85706f9438d5ad3988fd380d8c0312\": container with ID starting with 24d53d9bb4d59e5757a7974f266afa373b85706f9438d5ad3988fd380d8c0312 not found: ID does not exist" Feb 18 09:06:43 crc kubenswrapper[4556]: I0218 09:06:43.287582 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="accd4caa-1c67-4430-aa3a-db398194a103" path="/var/lib/kubelet/pods/accd4caa-1c67-4430-aa3a-db398194a103/volumes" Feb 18 09:06:44 crc kubenswrapper[4556]: I0218 09:06:44.869302 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-brf28" Feb 18 09:06:44 crc kubenswrapper[4556]: I0218 09:06:44.924663 4556 generic.go:334] "Generic (PLEG): container finished" podID="90477a11-cc7c-48cf-a210-9d1d802360ce" containerID="c0208355264ca4b2cb097ab6e495c05ecd84248a539dc45f09174336766c649a" exitCode=0 Feb 18 09:06:44 crc kubenswrapper[4556]: I0218 09:06:44.924707 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sd4dx" event={"ID":"90477a11-cc7c-48cf-a210-9d1d802360ce","Type":"ContainerDied","Data":"c0208355264ca4b2cb097ab6e495c05ecd84248a539dc45f09174336766c649a"} Feb 18 09:06:45 crc kubenswrapper[4556]: I0218 09:06:45.141021 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-ss8sl" Feb 18 09:06:45 crc kubenswrapper[4556]: I0218 09:06:45.384836 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-wqmhk" Feb 18 09:06:45 crc kubenswrapper[4556]: I0218 09:06:45.930997 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sd4dx" event={"ID":"90477a11-cc7c-48cf-a210-9d1d802360ce","Type":"ContainerStarted","Data":"832cdf04bbfa34d0e4091a6792478f2d3cbe62887d49c17406613a2b92772956"} Feb 18 09:06:45 crc kubenswrapper[4556]: I0218 09:06:45.949564 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-sd4dx" podStartSLOduration=3.776044218 podStartE2EDuration="41.94954626s" podCreationTimestamp="2026-02-18 09:06:04 +0000 UTC" firstStartedPulling="2026-02-18 09:06:07.267318114 +0000 UTC m=+124.284279094" lastFinishedPulling="2026-02-18 09:06:45.440820157 +0000 UTC m=+162.457781136" observedRunningTime="2026-02-18 09:06:45.947479036 +0000 UTC m=+162.964440015" watchObservedRunningTime="2026-02-18 09:06:45.94954626 +0000 UTC m=+162.966507240" Feb 18 09:06:47 crc kubenswrapper[4556]: I0218 09:06:47.051324 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ss8sl"] Feb 18 09:06:47 crc kubenswrapper[4556]: I0218 09:06:47.051523 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-ss8sl" podUID="7875d007-bdbc-428a-b291-283e8a1cdad0" containerName="registry-server" containerID="cri-o://43dcd4a4cacc561cfe972cbc840cff9d82992908726b740e6971a69a2dd195ba" gracePeriod=2 Feb 18 09:06:47 crc kubenswrapper[4556]: I0218 09:06:47.328115 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ss8sl" Feb 18 09:06:47 crc kubenswrapper[4556]: I0218 09:06:47.409889 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7875d007-bdbc-428a-b291-283e8a1cdad0-utilities\") pod \"7875d007-bdbc-428a-b291-283e8a1cdad0\" (UID: \"7875d007-bdbc-428a-b291-283e8a1cdad0\") " Feb 18 09:06:47 crc kubenswrapper[4556]: I0218 09:06:47.409970 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7875d007-bdbc-428a-b291-283e8a1cdad0-catalog-content\") pod \"7875d007-bdbc-428a-b291-283e8a1cdad0\" (UID: \"7875d007-bdbc-428a-b291-283e8a1cdad0\") " Feb 18 09:06:47 crc kubenswrapper[4556]: I0218 09:06:47.410038 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ktcxc\" (UniqueName: \"kubernetes.io/projected/7875d007-bdbc-428a-b291-283e8a1cdad0-kube-api-access-ktcxc\") pod \"7875d007-bdbc-428a-b291-283e8a1cdad0\" (UID: \"7875d007-bdbc-428a-b291-283e8a1cdad0\") " Feb 18 09:06:47 crc kubenswrapper[4556]: I0218 09:06:47.410567 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7875d007-bdbc-428a-b291-283e8a1cdad0-utilities" (OuterVolumeSpecName: "utilities") pod "7875d007-bdbc-428a-b291-283e8a1cdad0" (UID: "7875d007-bdbc-428a-b291-283e8a1cdad0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:06:47 crc kubenswrapper[4556]: I0218 09:06:47.410999 4556 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7875d007-bdbc-428a-b291-283e8a1cdad0-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 09:06:47 crc kubenswrapper[4556]: I0218 09:06:47.415090 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7875d007-bdbc-428a-b291-283e8a1cdad0-kube-api-access-ktcxc" (OuterVolumeSpecName: "kube-api-access-ktcxc") pod "7875d007-bdbc-428a-b291-283e8a1cdad0" (UID: "7875d007-bdbc-428a-b291-283e8a1cdad0"). InnerVolumeSpecName "kube-api-access-ktcxc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:06:47 crc kubenswrapper[4556]: I0218 09:06:47.449350 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7875d007-bdbc-428a-b291-283e8a1cdad0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7875d007-bdbc-428a-b291-283e8a1cdad0" (UID: "7875d007-bdbc-428a-b291-283e8a1cdad0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:06:47 crc kubenswrapper[4556]: I0218 09:06:47.512008 4556 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7875d007-bdbc-428a-b291-283e8a1cdad0-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 09:06:47 crc kubenswrapper[4556]: I0218 09:06:47.512039 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ktcxc\" (UniqueName: \"kubernetes.io/projected/7875d007-bdbc-428a-b291-283e8a1cdad0-kube-api-access-ktcxc\") on node \"crc\" DevicePath \"\"" Feb 18 09:06:47 crc kubenswrapper[4556]: I0218 09:06:47.942385 4556 generic.go:334] "Generic (PLEG): container finished" podID="7875d007-bdbc-428a-b291-283e8a1cdad0" containerID="43dcd4a4cacc561cfe972cbc840cff9d82992908726b740e6971a69a2dd195ba" exitCode=0 Feb 18 09:06:47 crc kubenswrapper[4556]: I0218 09:06:47.942428 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ss8sl" event={"ID":"7875d007-bdbc-428a-b291-283e8a1cdad0","Type":"ContainerDied","Data":"43dcd4a4cacc561cfe972cbc840cff9d82992908726b740e6971a69a2dd195ba"} Feb 18 09:06:47 crc kubenswrapper[4556]: I0218 09:06:47.942435 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ss8sl" Feb 18 09:06:47 crc kubenswrapper[4556]: I0218 09:06:47.942453 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ss8sl" event={"ID":"7875d007-bdbc-428a-b291-283e8a1cdad0","Type":"ContainerDied","Data":"2fb3b6373c0c3b961335f85fd9735a8bb14ff7f1fe1eff214be4098697d47785"} Feb 18 09:06:47 crc kubenswrapper[4556]: I0218 09:06:47.942470 4556 scope.go:117] "RemoveContainer" containerID="43dcd4a4cacc561cfe972cbc840cff9d82992908726b740e6971a69a2dd195ba" Feb 18 09:06:47 crc kubenswrapper[4556]: I0218 09:06:47.953702 4556 scope.go:117] "RemoveContainer" containerID="95a69f4487fff8166687ccd402c34aaccdea04bb52085712cc69bc783ad2fbef" Feb 18 09:06:47 crc kubenswrapper[4556]: I0218 09:06:47.973287 4556 scope.go:117] "RemoveContainer" containerID="449d08a8d66546a788eda025d228560e03d01cbafbe835839b2982d1aa177301" Feb 18 09:06:47 crc kubenswrapper[4556]: I0218 09:06:47.977713 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ss8sl"] Feb 18 09:06:47 crc kubenswrapper[4556]: I0218 09:06:47.983422 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-ss8sl"] Feb 18 09:06:47 crc kubenswrapper[4556]: I0218 09:06:47.997393 4556 scope.go:117] "RemoveContainer" containerID="43dcd4a4cacc561cfe972cbc840cff9d82992908726b740e6971a69a2dd195ba" Feb 18 09:06:47 crc kubenswrapper[4556]: E0218 09:06:47.997734 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43dcd4a4cacc561cfe972cbc840cff9d82992908726b740e6971a69a2dd195ba\": container with ID starting with 43dcd4a4cacc561cfe972cbc840cff9d82992908726b740e6971a69a2dd195ba not found: ID does not exist" containerID="43dcd4a4cacc561cfe972cbc840cff9d82992908726b740e6971a69a2dd195ba" Feb 18 09:06:47 crc kubenswrapper[4556]: I0218 09:06:47.997770 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43dcd4a4cacc561cfe972cbc840cff9d82992908726b740e6971a69a2dd195ba"} err="failed to get container status \"43dcd4a4cacc561cfe972cbc840cff9d82992908726b740e6971a69a2dd195ba\": rpc error: code = NotFound desc = could not find container \"43dcd4a4cacc561cfe972cbc840cff9d82992908726b740e6971a69a2dd195ba\": container with ID starting with 43dcd4a4cacc561cfe972cbc840cff9d82992908726b740e6971a69a2dd195ba not found: ID does not exist" Feb 18 09:06:47 crc kubenswrapper[4556]: I0218 09:06:47.997793 4556 scope.go:117] "RemoveContainer" containerID="95a69f4487fff8166687ccd402c34aaccdea04bb52085712cc69bc783ad2fbef" Feb 18 09:06:47 crc kubenswrapper[4556]: E0218 09:06:47.998044 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95a69f4487fff8166687ccd402c34aaccdea04bb52085712cc69bc783ad2fbef\": container with ID starting with 95a69f4487fff8166687ccd402c34aaccdea04bb52085712cc69bc783ad2fbef not found: ID does not exist" containerID="95a69f4487fff8166687ccd402c34aaccdea04bb52085712cc69bc783ad2fbef" Feb 18 09:06:47 crc kubenswrapper[4556]: I0218 09:06:47.998065 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95a69f4487fff8166687ccd402c34aaccdea04bb52085712cc69bc783ad2fbef"} err="failed to get container status \"95a69f4487fff8166687ccd402c34aaccdea04bb52085712cc69bc783ad2fbef\": rpc error: code = NotFound desc = could not find container \"95a69f4487fff8166687ccd402c34aaccdea04bb52085712cc69bc783ad2fbef\": container with ID starting with 95a69f4487fff8166687ccd402c34aaccdea04bb52085712cc69bc783ad2fbef not found: ID does not exist" Feb 18 09:06:47 crc kubenswrapper[4556]: I0218 09:06:47.998078 4556 scope.go:117] "RemoveContainer" containerID="449d08a8d66546a788eda025d228560e03d01cbafbe835839b2982d1aa177301" Feb 18 09:06:47 crc kubenswrapper[4556]: E0218 09:06:47.998427 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"449d08a8d66546a788eda025d228560e03d01cbafbe835839b2982d1aa177301\": container with ID starting with 449d08a8d66546a788eda025d228560e03d01cbafbe835839b2982d1aa177301 not found: ID does not exist" containerID="449d08a8d66546a788eda025d228560e03d01cbafbe835839b2982d1aa177301" Feb 18 09:06:47 crc kubenswrapper[4556]: I0218 09:06:47.998445 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"449d08a8d66546a788eda025d228560e03d01cbafbe835839b2982d1aa177301"} err="failed to get container status \"449d08a8d66546a788eda025d228560e03d01cbafbe835839b2982d1aa177301\": rpc error: code = NotFound desc = could not find container \"449d08a8d66546a788eda025d228560e03d01cbafbe835839b2982d1aa177301\": container with ID starting with 449d08a8d66546a788eda025d228560e03d01cbafbe835839b2982d1aa177301 not found: ID does not exist" Feb 18 09:06:48 crc kubenswrapper[4556]: I0218 09:06:48.322625 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d2f0e36c-cb29-4b06-bb22-6afd59466cab-metrics-certs\") pod \"network-metrics-daemon-cgwv9\" (UID: \"d2f0e36c-cb29-4b06-bb22-6afd59466cab\") " pod="openshift-multus/network-metrics-daemon-cgwv9" Feb 18 09:06:48 crc kubenswrapper[4556]: I0218 09:06:48.324102 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Feb 18 09:06:48 crc kubenswrapper[4556]: I0218 09:06:48.339475 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d2f0e36c-cb29-4b06-bb22-6afd59466cab-metrics-certs\") pod \"network-metrics-daemon-cgwv9\" (UID: \"d2f0e36c-cb29-4b06-bb22-6afd59466cab\") " pod="openshift-multus/network-metrics-daemon-cgwv9" Feb 18 09:06:48 crc kubenswrapper[4556]: I0218 09:06:48.409810 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Feb 18 09:06:48 crc kubenswrapper[4556]: I0218 09:06:48.419658 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cgwv9" Feb 18 09:06:48 crc kubenswrapper[4556]: I0218 09:06:48.746468 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-cgwv9"] Feb 18 09:06:48 crc kubenswrapper[4556]: W0218 09:06:48.750033 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd2f0e36c_cb29_4b06_bb22_6afd59466cab.slice/crio-570777f7d290ef3e4daac0c544cf9de11619aacc77a7747eb17506d5236675b8 WatchSource:0}: Error finding container 570777f7d290ef3e4daac0c544cf9de11619aacc77a7747eb17506d5236675b8: Status 404 returned error can't find the container with id 570777f7d290ef3e4daac0c544cf9de11619aacc77a7747eb17506d5236675b8 Feb 18 09:06:48 crc kubenswrapper[4556]: I0218 09:06:48.948665 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-cgwv9" event={"ID":"d2f0e36c-cb29-4b06-bb22-6afd59466cab","Type":"ContainerStarted","Data":"3f5d092ee57f7f9e85b036decde58ee68e94b6d3feccf6d782fcf64e4ba714a1"} Feb 18 09:06:48 crc kubenswrapper[4556]: I0218 09:06:48.948872 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-cgwv9" event={"ID":"d2f0e36c-cb29-4b06-bb22-6afd59466cab","Type":"ContainerStarted","Data":"570777f7d290ef3e4daac0c544cf9de11619aacc77a7747eb17506d5236675b8"} Feb 18 09:06:49 crc kubenswrapper[4556]: I0218 09:06:49.288573 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7875d007-bdbc-428a-b291-283e8a1cdad0" path="/var/lib/kubelet/pods/7875d007-bdbc-428a-b291-283e8a1cdad0/volumes" Feb 18 09:06:49 crc kubenswrapper[4556]: I0218 09:06:49.450453 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wqmhk"] Feb 18 09:06:49 crc kubenswrapper[4556]: I0218 09:06:49.450686 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-wqmhk" podUID="25cebcb5-2e64-4390-9d61-f06607eee280" containerName="registry-server" containerID="cri-o://bc03719a50264fe8b32082004059ef994cf5a497fc0943e571729c8c1de412d4" gracePeriod=2 Feb 18 09:06:49 crc kubenswrapper[4556]: I0218 09:06:49.722446 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wqmhk" Feb 18 09:06:49 crc kubenswrapper[4556]: I0218 09:06:49.838668 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kwpbb\" (UniqueName: \"kubernetes.io/projected/25cebcb5-2e64-4390-9d61-f06607eee280-kube-api-access-kwpbb\") pod \"25cebcb5-2e64-4390-9d61-f06607eee280\" (UID: \"25cebcb5-2e64-4390-9d61-f06607eee280\") " Feb 18 09:06:49 crc kubenswrapper[4556]: I0218 09:06:49.838704 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25cebcb5-2e64-4390-9d61-f06607eee280-utilities\") pod \"25cebcb5-2e64-4390-9d61-f06607eee280\" (UID: \"25cebcb5-2e64-4390-9d61-f06607eee280\") " Feb 18 09:06:49 crc kubenswrapper[4556]: I0218 09:06:49.838743 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25cebcb5-2e64-4390-9d61-f06607eee280-catalog-content\") pod \"25cebcb5-2e64-4390-9d61-f06607eee280\" (UID: \"25cebcb5-2e64-4390-9d61-f06607eee280\") " Feb 18 09:06:49 crc kubenswrapper[4556]: I0218 09:06:49.839535 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25cebcb5-2e64-4390-9d61-f06607eee280-utilities" (OuterVolumeSpecName: "utilities") pod "25cebcb5-2e64-4390-9d61-f06607eee280" (UID: "25cebcb5-2e64-4390-9d61-f06607eee280"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:06:49 crc kubenswrapper[4556]: I0218 09:06:49.842852 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25cebcb5-2e64-4390-9d61-f06607eee280-kube-api-access-kwpbb" (OuterVolumeSpecName: "kube-api-access-kwpbb") pod "25cebcb5-2e64-4390-9d61-f06607eee280" (UID: "25cebcb5-2e64-4390-9d61-f06607eee280"). InnerVolumeSpecName "kube-api-access-kwpbb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:06:49 crc kubenswrapper[4556]: I0218 09:06:49.875696 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25cebcb5-2e64-4390-9d61-f06607eee280-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "25cebcb5-2e64-4390-9d61-f06607eee280" (UID: "25cebcb5-2e64-4390-9d61-f06607eee280"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:06:49 crc kubenswrapper[4556]: I0218 09:06:49.939851 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kwpbb\" (UniqueName: \"kubernetes.io/projected/25cebcb5-2e64-4390-9d61-f06607eee280-kube-api-access-kwpbb\") on node \"crc\" DevicePath \"\"" Feb 18 09:06:49 crc kubenswrapper[4556]: I0218 09:06:49.939884 4556 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25cebcb5-2e64-4390-9d61-f06607eee280-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 09:06:49 crc kubenswrapper[4556]: I0218 09:06:49.939893 4556 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25cebcb5-2e64-4390-9d61-f06607eee280-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 09:06:49 crc kubenswrapper[4556]: I0218 09:06:49.956032 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-cgwv9" event={"ID":"d2f0e36c-cb29-4b06-bb22-6afd59466cab","Type":"ContainerStarted","Data":"dd1cc741dcb5581519594421cd15d21d21466c03f19744afc914da55cc1b20e9"} Feb 18 09:06:49 crc kubenswrapper[4556]: I0218 09:06:49.957698 4556 generic.go:334] "Generic (PLEG): container finished" podID="25cebcb5-2e64-4390-9d61-f06607eee280" containerID="bc03719a50264fe8b32082004059ef994cf5a497fc0943e571729c8c1de412d4" exitCode=0 Feb 18 09:06:49 crc kubenswrapper[4556]: I0218 09:06:49.957734 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wqmhk" event={"ID":"25cebcb5-2e64-4390-9d61-f06607eee280","Type":"ContainerDied","Data":"bc03719a50264fe8b32082004059ef994cf5a497fc0943e571729c8c1de412d4"} Feb 18 09:06:49 crc kubenswrapper[4556]: I0218 09:06:49.957755 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wqmhk" event={"ID":"25cebcb5-2e64-4390-9d61-f06607eee280","Type":"ContainerDied","Data":"5c554bf507ddc09a88bd7cbc6738f8f0a6481718839f00552ffbfa2bb81ad115"} Feb 18 09:06:49 crc kubenswrapper[4556]: I0218 09:06:49.957768 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wqmhk" Feb 18 09:06:49 crc kubenswrapper[4556]: I0218 09:06:49.957770 4556 scope.go:117] "RemoveContainer" containerID="bc03719a50264fe8b32082004059ef994cf5a497fc0943e571729c8c1de412d4" Feb 18 09:06:49 crc kubenswrapper[4556]: I0218 09:06:49.970066 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-cgwv9" podStartSLOduration=143.970048699 podStartE2EDuration="2m23.970048699s" podCreationTimestamp="2026-02-18 09:04:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:06:49.966765645 +0000 UTC m=+166.983726625" watchObservedRunningTime="2026-02-18 09:06:49.970048699 +0000 UTC m=+166.987009669" Feb 18 09:06:49 crc kubenswrapper[4556]: I0218 09:06:49.973953 4556 scope.go:117] "RemoveContainer" containerID="64dca3deba8f1f18af92dc864b2f5fbb5b624d3afce3d18f9dab3d6f1e538b26" Feb 18 09:06:49 crc kubenswrapper[4556]: I0218 09:06:49.981295 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wqmhk"] Feb 18 09:06:49 crc kubenswrapper[4556]: I0218 09:06:49.983892 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-wqmhk"] Feb 18 09:06:50 crc kubenswrapper[4556]: I0218 09:06:50.002328 4556 scope.go:117] "RemoveContainer" containerID="21164cb24c7ae3c208c4b8151a1fce3c76cc94a1e3dfa324207b42c2c85785f0" Feb 18 09:06:50 crc kubenswrapper[4556]: I0218 09:06:50.015695 4556 scope.go:117] "RemoveContainer" containerID="bc03719a50264fe8b32082004059ef994cf5a497fc0943e571729c8c1de412d4" Feb 18 09:06:50 crc kubenswrapper[4556]: E0218 09:06:50.016072 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc03719a50264fe8b32082004059ef994cf5a497fc0943e571729c8c1de412d4\": container with ID starting with bc03719a50264fe8b32082004059ef994cf5a497fc0943e571729c8c1de412d4 not found: ID does not exist" containerID="bc03719a50264fe8b32082004059ef994cf5a497fc0943e571729c8c1de412d4" Feb 18 09:06:50 crc kubenswrapper[4556]: I0218 09:06:50.016103 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc03719a50264fe8b32082004059ef994cf5a497fc0943e571729c8c1de412d4"} err="failed to get container status \"bc03719a50264fe8b32082004059ef994cf5a497fc0943e571729c8c1de412d4\": rpc error: code = NotFound desc = could not find container \"bc03719a50264fe8b32082004059ef994cf5a497fc0943e571729c8c1de412d4\": container with ID starting with bc03719a50264fe8b32082004059ef994cf5a497fc0943e571729c8c1de412d4 not found: ID does not exist" Feb 18 09:06:50 crc kubenswrapper[4556]: I0218 09:06:50.016128 4556 scope.go:117] "RemoveContainer" containerID="64dca3deba8f1f18af92dc864b2f5fbb5b624d3afce3d18f9dab3d6f1e538b26" Feb 18 09:06:50 crc kubenswrapper[4556]: E0218 09:06:50.016503 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64dca3deba8f1f18af92dc864b2f5fbb5b624d3afce3d18f9dab3d6f1e538b26\": container with ID starting with 64dca3deba8f1f18af92dc864b2f5fbb5b624d3afce3d18f9dab3d6f1e538b26 not found: ID does not exist" containerID="64dca3deba8f1f18af92dc864b2f5fbb5b624d3afce3d18f9dab3d6f1e538b26" Feb 18 09:06:50 crc kubenswrapper[4556]: I0218 09:06:50.016526 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64dca3deba8f1f18af92dc864b2f5fbb5b624d3afce3d18f9dab3d6f1e538b26"} err="failed to get container status \"64dca3deba8f1f18af92dc864b2f5fbb5b624d3afce3d18f9dab3d6f1e538b26\": rpc error: code = NotFound desc = could not find container \"64dca3deba8f1f18af92dc864b2f5fbb5b624d3afce3d18f9dab3d6f1e538b26\": container with ID starting with 64dca3deba8f1f18af92dc864b2f5fbb5b624d3afce3d18f9dab3d6f1e538b26 not found: ID does not exist" Feb 18 09:06:50 crc kubenswrapper[4556]: I0218 09:06:50.016540 4556 scope.go:117] "RemoveContainer" containerID="21164cb24c7ae3c208c4b8151a1fce3c76cc94a1e3dfa324207b42c2c85785f0" Feb 18 09:06:50 crc kubenswrapper[4556]: E0218 09:06:50.016791 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21164cb24c7ae3c208c4b8151a1fce3c76cc94a1e3dfa324207b42c2c85785f0\": container with ID starting with 21164cb24c7ae3c208c4b8151a1fce3c76cc94a1e3dfa324207b42c2c85785f0 not found: ID does not exist" containerID="21164cb24c7ae3c208c4b8151a1fce3c76cc94a1e3dfa324207b42c2c85785f0" Feb 18 09:06:50 crc kubenswrapper[4556]: I0218 09:06:50.016823 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21164cb24c7ae3c208c4b8151a1fce3c76cc94a1e3dfa324207b42c2c85785f0"} err="failed to get container status \"21164cb24c7ae3c208c4b8151a1fce3c76cc94a1e3dfa324207b42c2c85785f0\": rpc error: code = NotFound desc = could not find container \"21164cb24c7ae3c208c4b8151a1fce3c76cc94a1e3dfa324207b42c2c85785f0\": container with ID starting with 21164cb24c7ae3c208c4b8151a1fce3c76cc94a1e3dfa324207b42c2c85785f0 not found: ID does not exist" Feb 18 09:06:51 crc kubenswrapper[4556]: I0218 09:06:51.287922 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25cebcb5-2e64-4390-9d61-f06607eee280" path="/var/lib/kubelet/pods/25cebcb5-2e64-4390-9d61-f06607eee280/volumes" Feb 18 09:06:52 crc kubenswrapper[4556]: I0218 09:06:52.039760 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 18 09:06:52 crc kubenswrapper[4556]: E0218 09:06:52.039933 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc2d3f5c-f685-43ea-bf13-4bec129bf889" containerName="extract-utilities" Feb 18 09:06:52 crc kubenswrapper[4556]: I0218 09:06:52.039945 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc2d3f5c-f685-43ea-bf13-4bec129bf889" containerName="extract-utilities" Feb 18 09:06:52 crc kubenswrapper[4556]: E0218 09:06:52.039955 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25cebcb5-2e64-4390-9d61-f06607eee280" containerName="registry-server" Feb 18 09:06:52 crc kubenswrapper[4556]: I0218 09:06:52.039961 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="25cebcb5-2e64-4390-9d61-f06607eee280" containerName="registry-server" Feb 18 09:06:52 crc kubenswrapper[4556]: E0218 09:06:52.039979 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50482765-d4b6-4bbf-9fd7-6d077e861c98" containerName="pruner" Feb 18 09:06:52 crc kubenswrapper[4556]: I0218 09:06:52.039984 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="50482765-d4b6-4bbf-9fd7-6d077e861c98" containerName="pruner" Feb 18 09:06:52 crc kubenswrapper[4556]: E0218 09:06:52.039994 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25cebcb5-2e64-4390-9d61-f06607eee280" containerName="extract-utilities" Feb 18 09:06:52 crc kubenswrapper[4556]: I0218 09:06:52.039999 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="25cebcb5-2e64-4390-9d61-f06607eee280" containerName="extract-utilities" Feb 18 09:06:52 crc kubenswrapper[4556]: E0218 09:06:52.040006 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="accd4caa-1c67-4430-aa3a-db398194a103" containerName="extract-content" Feb 18 09:06:52 crc kubenswrapper[4556]: I0218 09:06:52.040011 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="accd4caa-1c67-4430-aa3a-db398194a103" containerName="extract-content" Feb 18 09:06:52 crc kubenswrapper[4556]: E0218 09:06:52.040019 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7875d007-bdbc-428a-b291-283e8a1cdad0" containerName="extract-content" Feb 18 09:06:52 crc kubenswrapper[4556]: I0218 09:06:52.040024 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="7875d007-bdbc-428a-b291-283e8a1cdad0" containerName="extract-content" Feb 18 09:06:52 crc kubenswrapper[4556]: E0218 09:06:52.040031 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="accd4caa-1c67-4430-aa3a-db398194a103" containerName="extract-utilities" Feb 18 09:06:52 crc kubenswrapper[4556]: I0218 09:06:52.040036 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="accd4caa-1c67-4430-aa3a-db398194a103" containerName="extract-utilities" Feb 18 09:06:52 crc kubenswrapper[4556]: E0218 09:06:52.040043 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7875d007-bdbc-428a-b291-283e8a1cdad0" containerName="extract-utilities" Feb 18 09:06:52 crc kubenswrapper[4556]: I0218 09:06:52.040049 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="7875d007-bdbc-428a-b291-283e8a1cdad0" containerName="extract-utilities" Feb 18 09:06:52 crc kubenswrapper[4556]: E0218 09:06:52.040058 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7875d007-bdbc-428a-b291-283e8a1cdad0" containerName="registry-server" Feb 18 09:06:52 crc kubenswrapper[4556]: I0218 09:06:52.040064 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="7875d007-bdbc-428a-b291-283e8a1cdad0" containerName="registry-server" Feb 18 09:06:52 crc kubenswrapper[4556]: E0218 09:06:52.040070 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc2d3f5c-f685-43ea-bf13-4bec129bf889" containerName="registry-server" Feb 18 09:06:52 crc kubenswrapper[4556]: I0218 09:06:52.040075 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc2d3f5c-f685-43ea-bf13-4bec129bf889" containerName="registry-server" Feb 18 09:06:52 crc kubenswrapper[4556]: E0218 09:06:52.040082 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="accd4caa-1c67-4430-aa3a-db398194a103" containerName="registry-server" Feb 18 09:06:52 crc kubenswrapper[4556]: I0218 09:06:52.040088 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="accd4caa-1c67-4430-aa3a-db398194a103" containerName="registry-server" Feb 18 09:06:52 crc kubenswrapper[4556]: E0218 09:06:52.040095 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25cebcb5-2e64-4390-9d61-f06607eee280" containerName="extract-content" Feb 18 09:06:52 crc kubenswrapper[4556]: I0218 09:06:52.040100 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="25cebcb5-2e64-4390-9d61-f06607eee280" containerName="extract-content" Feb 18 09:06:52 crc kubenswrapper[4556]: E0218 09:06:52.040112 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc2d3f5c-f685-43ea-bf13-4bec129bf889" containerName="extract-content" Feb 18 09:06:52 crc kubenswrapper[4556]: I0218 09:06:52.040117 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc2d3f5c-f685-43ea-bf13-4bec129bf889" containerName="extract-content" Feb 18 09:06:52 crc kubenswrapper[4556]: I0218 09:06:52.040217 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="50482765-d4b6-4bbf-9fd7-6d077e861c98" containerName="pruner" Feb 18 09:06:52 crc kubenswrapper[4556]: I0218 09:06:52.040227 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="accd4caa-1c67-4430-aa3a-db398194a103" containerName="registry-server" Feb 18 09:06:52 crc kubenswrapper[4556]: I0218 09:06:52.040236 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="25cebcb5-2e64-4390-9d61-f06607eee280" containerName="registry-server" Feb 18 09:06:52 crc kubenswrapper[4556]: I0218 09:06:52.040245 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="7875d007-bdbc-428a-b291-283e8a1cdad0" containerName="registry-server" Feb 18 09:06:52 crc kubenswrapper[4556]: I0218 09:06:52.040252 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc2d3f5c-f685-43ea-bf13-4bec129bf889" containerName="registry-server" Feb 18 09:06:52 crc kubenswrapper[4556]: I0218 09:06:52.040584 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 18 09:06:52 crc kubenswrapper[4556]: I0218 09:06:52.041898 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 18 09:06:52 crc kubenswrapper[4556]: I0218 09:06:52.043205 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 18 09:06:52 crc kubenswrapper[4556]: I0218 09:06:52.046514 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 18 09:06:52 crc kubenswrapper[4556]: I0218 09:06:52.162215 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bcc5d920-dabd-42f2-937c-a27cceeb5874-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"bcc5d920-dabd-42f2-937c-a27cceeb5874\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 18 09:06:52 crc kubenswrapper[4556]: I0218 09:06:52.162299 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bcc5d920-dabd-42f2-937c-a27cceeb5874-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"bcc5d920-dabd-42f2-937c-a27cceeb5874\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 18 09:06:52 crc kubenswrapper[4556]: I0218 09:06:52.262847 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bcc5d920-dabd-42f2-937c-a27cceeb5874-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"bcc5d920-dabd-42f2-937c-a27cceeb5874\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 18 09:06:52 crc kubenswrapper[4556]: I0218 09:06:52.262979 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bcc5d920-dabd-42f2-937c-a27cceeb5874-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"bcc5d920-dabd-42f2-937c-a27cceeb5874\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 18 09:06:52 crc kubenswrapper[4556]: I0218 09:06:52.262992 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bcc5d920-dabd-42f2-937c-a27cceeb5874-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"bcc5d920-dabd-42f2-937c-a27cceeb5874\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 18 09:06:52 crc kubenswrapper[4556]: I0218 09:06:52.277333 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bcc5d920-dabd-42f2-937c-a27cceeb5874-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"bcc5d920-dabd-42f2-937c-a27cceeb5874\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 18 09:06:52 crc kubenswrapper[4556]: I0218 09:06:52.357090 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 18 09:06:52 crc kubenswrapper[4556]: I0218 09:06:52.684783 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 18 09:06:52 crc kubenswrapper[4556]: W0218 09:06:52.688619 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podbcc5d920_dabd_42f2_937c_a27cceeb5874.slice/crio-934e942f8caebc62ff424411c5b929158f1a8340f62a7afebf8ef2352d8348c2 WatchSource:0}: Error finding container 934e942f8caebc62ff424411c5b929158f1a8340f62a7afebf8ef2352d8348c2: Status 404 returned error can't find the container with id 934e942f8caebc62ff424411c5b929158f1a8340f62a7afebf8ef2352d8348c2 Feb 18 09:06:52 crc kubenswrapper[4556]: I0218 09:06:52.976886 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"bcc5d920-dabd-42f2-937c-a27cceeb5874","Type":"ContainerStarted","Data":"992844a549f5d86e99e6cf41392619026257da8e027ba879bd00309ac0e3d168"} Feb 18 09:06:52 crc kubenswrapper[4556]: I0218 09:06:52.977410 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"bcc5d920-dabd-42f2-937c-a27cceeb5874","Type":"ContainerStarted","Data":"934e942f8caebc62ff424411c5b929158f1a8340f62a7afebf8ef2352d8348c2"} Feb 18 09:06:52 crc kubenswrapper[4556]: I0218 09:06:52.987905 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=0.987888365 podStartE2EDuration="987.888365ms" podCreationTimestamp="2026-02-18 09:06:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:06:52.985571479 +0000 UTC m=+170.002532459" watchObservedRunningTime="2026-02-18 09:06:52.987888365 +0000 UTC m=+170.004849334" Feb 18 09:06:53 crc kubenswrapper[4556]: I0218 09:06:53.983338 4556 generic.go:334] "Generic (PLEG): container finished" podID="bcc5d920-dabd-42f2-937c-a27cceeb5874" containerID="992844a549f5d86e99e6cf41392619026257da8e027ba879bd00309ac0e3d168" exitCode=0 Feb 18 09:06:53 crc kubenswrapper[4556]: I0218 09:06:53.983376 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"bcc5d920-dabd-42f2-937c-a27cceeb5874","Type":"ContainerDied","Data":"992844a549f5d86e99e6cf41392619026257da8e027ba879bd00309ac0e3d168"} Feb 18 09:06:54 crc kubenswrapper[4556]: I0218 09:06:54.929038 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-sd4dx" Feb 18 09:06:54 crc kubenswrapper[4556]: I0218 09:06:54.929077 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-sd4dx" Feb 18 09:06:54 crc kubenswrapper[4556]: I0218 09:06:54.953109 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-sd4dx" Feb 18 09:06:55 crc kubenswrapper[4556]: I0218 09:06:55.013347 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-sd4dx" Feb 18 09:06:55 crc kubenswrapper[4556]: I0218 09:06:55.152319 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 18 09:06:55 crc kubenswrapper[4556]: I0218 09:06:55.296131 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bcc5d920-dabd-42f2-937c-a27cceeb5874-kube-api-access\") pod \"bcc5d920-dabd-42f2-937c-a27cceeb5874\" (UID: \"bcc5d920-dabd-42f2-937c-a27cceeb5874\") " Feb 18 09:06:55 crc kubenswrapper[4556]: I0218 09:06:55.296183 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bcc5d920-dabd-42f2-937c-a27cceeb5874-kubelet-dir\") pod \"bcc5d920-dabd-42f2-937c-a27cceeb5874\" (UID: \"bcc5d920-dabd-42f2-937c-a27cceeb5874\") " Feb 18 09:06:55 crc kubenswrapper[4556]: I0218 09:06:55.296226 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bcc5d920-dabd-42f2-937c-a27cceeb5874-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "bcc5d920-dabd-42f2-937c-a27cceeb5874" (UID: "bcc5d920-dabd-42f2-937c-a27cceeb5874"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 09:06:55 crc kubenswrapper[4556]: I0218 09:06:55.296408 4556 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bcc5d920-dabd-42f2-937c-a27cceeb5874-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 18 09:06:55 crc kubenswrapper[4556]: I0218 09:06:55.299922 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcc5d920-dabd-42f2-937c-a27cceeb5874-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "bcc5d920-dabd-42f2-937c-a27cceeb5874" (UID: "bcc5d920-dabd-42f2-937c-a27cceeb5874"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:06:55 crc kubenswrapper[4556]: I0218 09:06:55.397665 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bcc5d920-dabd-42f2-937c-a27cceeb5874-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 18 09:06:55 crc kubenswrapper[4556]: I0218 09:06:55.991727 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 18 09:06:55 crc kubenswrapper[4556]: I0218 09:06:55.991770 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"bcc5d920-dabd-42f2-937c-a27cceeb5874","Type":"ContainerDied","Data":"934e942f8caebc62ff424411c5b929158f1a8340f62a7afebf8ef2352d8348c2"} Feb 18 09:06:55 crc kubenswrapper[4556]: I0218 09:06:55.992070 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="934e942f8caebc62ff424411c5b929158f1a8340f62a7afebf8ef2352d8348c2" Feb 18 09:06:58 crc kubenswrapper[4556]: I0218 09:06:58.037904 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 18 09:06:58 crc kubenswrapper[4556]: E0218 09:06:58.038097 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcc5d920-dabd-42f2-937c-a27cceeb5874" containerName="pruner" Feb 18 09:06:58 crc kubenswrapper[4556]: I0218 09:06:58.038109 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcc5d920-dabd-42f2-937c-a27cceeb5874" containerName="pruner" Feb 18 09:06:58 crc kubenswrapper[4556]: I0218 09:06:58.038213 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcc5d920-dabd-42f2-937c-a27cceeb5874" containerName="pruner" Feb 18 09:06:58 crc kubenswrapper[4556]: I0218 09:06:58.038520 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 18 09:06:58 crc kubenswrapper[4556]: I0218 09:06:58.040220 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 18 09:06:58 crc kubenswrapper[4556]: I0218 09:06:58.040372 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 18 09:06:58 crc kubenswrapper[4556]: I0218 09:06:58.051363 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 18 09:06:58 crc kubenswrapper[4556]: I0218 09:06:58.121617 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/12967c01-f63e-466e-b28f-4828af636833-var-lock\") pod \"installer-9-crc\" (UID: \"12967c01-f63e-466e-b28f-4828af636833\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 18 09:06:58 crc kubenswrapper[4556]: I0218 09:06:58.121670 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/12967c01-f63e-466e-b28f-4828af636833-kubelet-dir\") pod \"installer-9-crc\" (UID: \"12967c01-f63e-466e-b28f-4828af636833\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 18 09:06:58 crc kubenswrapper[4556]: I0218 09:06:58.121723 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/12967c01-f63e-466e-b28f-4828af636833-kube-api-access\") pod \"installer-9-crc\" (UID: \"12967c01-f63e-466e-b28f-4828af636833\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 18 09:06:58 crc kubenswrapper[4556]: I0218 09:06:58.222858 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/12967c01-f63e-466e-b28f-4828af636833-var-lock\") pod \"installer-9-crc\" (UID: \"12967c01-f63e-466e-b28f-4828af636833\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 18 09:06:58 crc kubenswrapper[4556]: I0218 09:06:58.222899 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/12967c01-f63e-466e-b28f-4828af636833-kubelet-dir\") pod \"installer-9-crc\" (UID: \"12967c01-f63e-466e-b28f-4828af636833\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 18 09:06:58 crc kubenswrapper[4556]: I0218 09:06:58.222928 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/12967c01-f63e-466e-b28f-4828af636833-kube-api-access\") pod \"installer-9-crc\" (UID: \"12967c01-f63e-466e-b28f-4828af636833\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 18 09:06:58 crc kubenswrapper[4556]: I0218 09:06:58.222946 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/12967c01-f63e-466e-b28f-4828af636833-var-lock\") pod \"installer-9-crc\" (UID: \"12967c01-f63e-466e-b28f-4828af636833\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 18 09:06:58 crc kubenswrapper[4556]: I0218 09:06:58.223048 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/12967c01-f63e-466e-b28f-4828af636833-kubelet-dir\") pod \"installer-9-crc\" (UID: \"12967c01-f63e-466e-b28f-4828af636833\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 18 09:06:58 crc kubenswrapper[4556]: I0218 09:06:58.237283 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/12967c01-f63e-466e-b28f-4828af636833-kube-api-access\") pod \"installer-9-crc\" (UID: \"12967c01-f63e-466e-b28f-4828af636833\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 18 09:06:58 crc kubenswrapper[4556]: I0218 09:06:58.356392 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 18 09:06:58 crc kubenswrapper[4556]: I0218 09:06:58.683216 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 18 09:06:59 crc kubenswrapper[4556]: I0218 09:06:59.008103 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"12967c01-f63e-466e-b28f-4828af636833","Type":"ContainerStarted","Data":"606ca898b3cfd55868817996f930826fa3c99883e206b760ac8f6965d3593a2e"} Feb 18 09:06:59 crc kubenswrapper[4556]: I0218 09:06:59.008323 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"12967c01-f63e-466e-b28f-4828af636833","Type":"ContainerStarted","Data":"7f7c44c9cd7c42a3a328479b76c8c9732a99646ccd2b103743ae485bf0c02092"} Feb 18 09:06:59 crc kubenswrapper[4556]: I0218 09:06:59.019942 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=1.019926623 podStartE2EDuration="1.019926623s" podCreationTimestamp="2026-02-18 09:06:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:06:59.018858772 +0000 UTC m=+176.035819752" watchObservedRunningTime="2026-02-18 09:06:59.019926623 +0000 UTC m=+176.036887604" Feb 18 09:07:01 crc kubenswrapper[4556]: I0218 09:07:01.728249 4556 patch_prober.go:28] interesting pod/machine-config-daemon-f76hs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 09:07:01 crc kubenswrapper[4556]: I0218 09:07:01.728314 4556 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 09:07:02 crc kubenswrapper[4556]: I0218 09:07:02.148420 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-vxdmn" podUID="9db33ada-017a-4281-85c1-0c1530bc3bbc" containerName="oauth-openshift" containerID="cri-o://f95e8072a9929cfc1b3f58f632a3d8a2bdbaf2dafc7a7f7a6cb653d94674a089" gracePeriod=15 Feb 18 09:07:02 crc kubenswrapper[4556]: I0218 09:07:02.423200 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-vxdmn" Feb 18 09:07:02 crc kubenswrapper[4556]: I0218 09:07:02.566923 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-user-idp-0-file-data\") pod \"9db33ada-017a-4281-85c1-0c1530bc3bbc\" (UID: \"9db33ada-017a-4281-85c1-0c1530bc3bbc\") " Feb 18 09:07:02 crc kubenswrapper[4556]: I0218 09:07:02.566975 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-user-template-error\") pod \"9db33ada-017a-4281-85c1-0c1530bc3bbc\" (UID: \"9db33ada-017a-4281-85c1-0c1530bc3bbc\") " Feb 18 09:07:02 crc kubenswrapper[4556]: I0218 09:07:02.567029 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-system-session\") pod \"9db33ada-017a-4281-85c1-0c1530bc3bbc\" (UID: \"9db33ada-017a-4281-85c1-0c1530bc3bbc\") " Feb 18 09:07:02 crc kubenswrapper[4556]: I0218 09:07:02.567052 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-system-router-certs\") pod \"9db33ada-017a-4281-85c1-0c1530bc3bbc\" (UID: \"9db33ada-017a-4281-85c1-0c1530bc3bbc\") " Feb 18 09:07:02 crc kubenswrapper[4556]: I0218 09:07:02.567078 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kn29l\" (UniqueName: \"kubernetes.io/projected/9db33ada-017a-4281-85c1-0c1530bc3bbc-kube-api-access-kn29l\") pod \"9db33ada-017a-4281-85c1-0c1530bc3bbc\" (UID: \"9db33ada-017a-4281-85c1-0c1530bc3bbc\") " Feb 18 09:07:02 crc kubenswrapper[4556]: I0218 09:07:02.567096 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-system-cliconfig\") pod \"9db33ada-017a-4281-85c1-0c1530bc3bbc\" (UID: \"9db33ada-017a-4281-85c1-0c1530bc3bbc\") " Feb 18 09:07:02 crc kubenswrapper[4556]: I0218 09:07:02.567110 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9db33ada-017a-4281-85c1-0c1530bc3bbc-audit-policies\") pod \"9db33ada-017a-4281-85c1-0c1530bc3bbc\" (UID: \"9db33ada-017a-4281-85c1-0c1530bc3bbc\") " Feb 18 09:07:02 crc kubenswrapper[4556]: I0218 09:07:02.567169 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-system-service-ca\") pod \"9db33ada-017a-4281-85c1-0c1530bc3bbc\" (UID: \"9db33ada-017a-4281-85c1-0c1530bc3bbc\") " Feb 18 09:07:02 crc kubenswrapper[4556]: I0218 09:07:02.567186 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-system-trusted-ca-bundle\") pod \"9db33ada-017a-4281-85c1-0c1530bc3bbc\" (UID: \"9db33ada-017a-4281-85c1-0c1530bc3bbc\") " Feb 18 09:07:02 crc kubenswrapper[4556]: I0218 09:07:02.567202 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-user-template-login\") pod \"9db33ada-017a-4281-85c1-0c1530bc3bbc\" (UID: \"9db33ada-017a-4281-85c1-0c1530bc3bbc\") " Feb 18 09:07:02 crc kubenswrapper[4556]: I0218 09:07:02.567226 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-user-template-provider-selection\") pod \"9db33ada-017a-4281-85c1-0c1530bc3bbc\" (UID: \"9db33ada-017a-4281-85c1-0c1530bc3bbc\") " Feb 18 09:07:02 crc kubenswrapper[4556]: I0218 09:07:02.567251 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-system-ocp-branding-template\") pod \"9db33ada-017a-4281-85c1-0c1530bc3bbc\" (UID: \"9db33ada-017a-4281-85c1-0c1530bc3bbc\") " Feb 18 09:07:02 crc kubenswrapper[4556]: I0218 09:07:02.567289 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-system-serving-cert\") pod \"9db33ada-017a-4281-85c1-0c1530bc3bbc\" (UID: \"9db33ada-017a-4281-85c1-0c1530bc3bbc\") " Feb 18 09:07:02 crc kubenswrapper[4556]: I0218 09:07:02.567308 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9db33ada-017a-4281-85c1-0c1530bc3bbc-audit-dir\") pod \"9db33ada-017a-4281-85c1-0c1530bc3bbc\" (UID: \"9db33ada-017a-4281-85c1-0c1530bc3bbc\") " Feb 18 09:07:02 crc kubenswrapper[4556]: I0218 09:07:02.567890 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9db33ada-017a-4281-85c1-0c1530bc3bbc-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "9db33ada-017a-4281-85c1-0c1530bc3bbc" (UID: "9db33ada-017a-4281-85c1-0c1530bc3bbc"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 09:07:02 crc kubenswrapper[4556]: I0218 09:07:02.568096 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "9db33ada-017a-4281-85c1-0c1530bc3bbc" (UID: "9db33ada-017a-4281-85c1-0c1530bc3bbc"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:07:02 crc kubenswrapper[4556]: I0218 09:07:02.568112 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9db33ada-017a-4281-85c1-0c1530bc3bbc-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "9db33ada-017a-4281-85c1-0c1530bc3bbc" (UID: "9db33ada-017a-4281-85c1-0c1530bc3bbc"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:07:02 crc kubenswrapper[4556]: I0218 09:07:02.568119 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "9db33ada-017a-4281-85c1-0c1530bc3bbc" (UID: "9db33ada-017a-4281-85c1-0c1530bc3bbc"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:07:02 crc kubenswrapper[4556]: I0218 09:07:02.568361 4556 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9db33ada-017a-4281-85c1-0c1530bc3bbc-audit-dir\") on node \"crc\" DevicePath \"\"" Feb 18 09:07:02 crc kubenswrapper[4556]: I0218 09:07:02.568383 4556 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9db33ada-017a-4281-85c1-0c1530bc3bbc-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 18 09:07:02 crc kubenswrapper[4556]: I0218 09:07:02.568394 4556 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Feb 18 09:07:02 crc kubenswrapper[4556]: I0218 09:07:02.568402 4556 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:07:02 crc kubenswrapper[4556]: I0218 09:07:02.568488 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "9db33ada-017a-4281-85c1-0c1530bc3bbc" (UID: "9db33ada-017a-4281-85c1-0c1530bc3bbc"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:07:02 crc kubenswrapper[4556]: I0218 09:07:02.571707 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "9db33ada-017a-4281-85c1-0c1530bc3bbc" (UID: "9db33ada-017a-4281-85c1-0c1530bc3bbc"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:07:02 crc kubenswrapper[4556]: I0218 09:07:02.571888 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "9db33ada-017a-4281-85c1-0c1530bc3bbc" (UID: "9db33ada-017a-4281-85c1-0c1530bc3bbc"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:07:02 crc kubenswrapper[4556]: I0218 09:07:02.572090 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "9db33ada-017a-4281-85c1-0c1530bc3bbc" (UID: "9db33ada-017a-4281-85c1-0c1530bc3bbc"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:07:02 crc kubenswrapper[4556]: I0218 09:07:02.572353 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "9db33ada-017a-4281-85c1-0c1530bc3bbc" (UID: "9db33ada-017a-4281-85c1-0c1530bc3bbc"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:07:02 crc kubenswrapper[4556]: I0218 09:07:02.572521 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "9db33ada-017a-4281-85c1-0c1530bc3bbc" (UID: "9db33ada-017a-4281-85c1-0c1530bc3bbc"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:07:02 crc kubenswrapper[4556]: I0218 09:07:02.572714 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "9db33ada-017a-4281-85c1-0c1530bc3bbc" (UID: "9db33ada-017a-4281-85c1-0c1530bc3bbc"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:07:02 crc kubenswrapper[4556]: I0218 09:07:02.572786 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9db33ada-017a-4281-85c1-0c1530bc3bbc-kube-api-access-kn29l" (OuterVolumeSpecName: "kube-api-access-kn29l") pod "9db33ada-017a-4281-85c1-0c1530bc3bbc" (UID: "9db33ada-017a-4281-85c1-0c1530bc3bbc"). InnerVolumeSpecName "kube-api-access-kn29l". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:07:02 crc kubenswrapper[4556]: I0218 09:07:02.572828 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "9db33ada-017a-4281-85c1-0c1530bc3bbc" (UID: "9db33ada-017a-4281-85c1-0c1530bc3bbc"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:07:02 crc kubenswrapper[4556]: I0218 09:07:02.573428 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "9db33ada-017a-4281-85c1-0c1530bc3bbc" (UID: "9db33ada-017a-4281-85c1-0c1530bc3bbc"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:07:02 crc kubenswrapper[4556]: I0218 09:07:02.669229 4556 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Feb 18 09:07:02 crc kubenswrapper[4556]: I0218 09:07:02.669256 4556 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 09:07:02 crc kubenswrapper[4556]: I0218 09:07:02.669267 4556 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Feb 18 09:07:02 crc kubenswrapper[4556]: I0218 09:07:02.669278 4556 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Feb 18 09:07:02 crc kubenswrapper[4556]: I0218 09:07:02.669287 4556 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Feb 18 09:07:02 crc kubenswrapper[4556]: I0218 09:07:02.669296 4556 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Feb 18 09:07:02 crc kubenswrapper[4556]: I0218 09:07:02.669304 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kn29l\" (UniqueName: \"kubernetes.io/projected/9db33ada-017a-4281-85c1-0c1530bc3bbc-kube-api-access-kn29l\") on node \"crc\" DevicePath \"\"" Feb 18 09:07:02 crc kubenswrapper[4556]: I0218 09:07:02.669313 4556 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Feb 18 09:07:02 crc kubenswrapper[4556]: I0218 09:07:02.669324 4556 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Feb 18 09:07:02 crc kubenswrapper[4556]: I0218 09:07:02.669333 4556 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/9db33ada-017a-4281-85c1-0c1530bc3bbc-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Feb 18 09:07:03 crc kubenswrapper[4556]: I0218 09:07:03.024447 4556 generic.go:334] "Generic (PLEG): container finished" podID="9db33ada-017a-4281-85c1-0c1530bc3bbc" containerID="f95e8072a9929cfc1b3f58f632a3d8a2bdbaf2dafc7a7f7a6cb653d94674a089" exitCode=0 Feb 18 09:07:03 crc kubenswrapper[4556]: I0218 09:07:03.024504 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-vxdmn" Feb 18 09:07:03 crc kubenswrapper[4556]: I0218 09:07:03.024492 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-vxdmn" event={"ID":"9db33ada-017a-4281-85c1-0c1530bc3bbc","Type":"ContainerDied","Data":"f95e8072a9929cfc1b3f58f632a3d8a2bdbaf2dafc7a7f7a6cb653d94674a089"} Feb 18 09:07:03 crc kubenswrapper[4556]: I0218 09:07:03.024620 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-vxdmn" event={"ID":"9db33ada-017a-4281-85c1-0c1530bc3bbc","Type":"ContainerDied","Data":"8aa857eb2ce6837468a5a1ffec5a01a532f05112cc366ffa17fda96412a25bd1"} Feb 18 09:07:03 crc kubenswrapper[4556]: I0218 09:07:03.024639 4556 scope.go:117] "RemoveContainer" containerID="f95e8072a9929cfc1b3f58f632a3d8a2bdbaf2dafc7a7f7a6cb653d94674a089" Feb 18 09:07:03 crc kubenswrapper[4556]: I0218 09:07:03.037946 4556 scope.go:117] "RemoveContainer" containerID="f95e8072a9929cfc1b3f58f632a3d8a2bdbaf2dafc7a7f7a6cb653d94674a089" Feb 18 09:07:03 crc kubenswrapper[4556]: E0218 09:07:03.038246 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f95e8072a9929cfc1b3f58f632a3d8a2bdbaf2dafc7a7f7a6cb653d94674a089\": container with ID starting with f95e8072a9929cfc1b3f58f632a3d8a2bdbaf2dafc7a7f7a6cb653d94674a089 not found: ID does not exist" containerID="f95e8072a9929cfc1b3f58f632a3d8a2bdbaf2dafc7a7f7a6cb653d94674a089" Feb 18 09:07:03 crc kubenswrapper[4556]: I0218 09:07:03.038272 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f95e8072a9929cfc1b3f58f632a3d8a2bdbaf2dafc7a7f7a6cb653d94674a089"} err="failed to get container status \"f95e8072a9929cfc1b3f58f632a3d8a2bdbaf2dafc7a7f7a6cb653d94674a089\": rpc error: code = NotFound desc = could not find container \"f95e8072a9929cfc1b3f58f632a3d8a2bdbaf2dafc7a7f7a6cb653d94674a089\": container with ID starting with f95e8072a9929cfc1b3f58f632a3d8a2bdbaf2dafc7a7f7a6cb653d94674a089 not found: ID does not exist" Feb 18 09:07:03 crc kubenswrapper[4556]: I0218 09:07:03.042965 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-vxdmn"] Feb 18 09:07:03 crc kubenswrapper[4556]: I0218 09:07:03.045307 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-vxdmn"] Feb 18 09:07:03 crc kubenswrapper[4556]: I0218 09:07:03.287415 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9db33ada-017a-4281-85c1-0c1530bc3bbc" path="/var/lib/kubelet/pods/9db33ada-017a-4281-85c1-0c1530bc3bbc/volumes" Feb 18 09:07:05 crc kubenswrapper[4556]: I0218 09:07:05.909834 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-6b68866bbb-df7kt"] Feb 18 09:07:05 crc kubenswrapper[4556]: E0218 09:07:05.910212 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9db33ada-017a-4281-85c1-0c1530bc3bbc" containerName="oauth-openshift" Feb 18 09:07:05 crc kubenswrapper[4556]: I0218 09:07:05.910225 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="9db33ada-017a-4281-85c1-0c1530bc3bbc" containerName="oauth-openshift" Feb 18 09:07:05 crc kubenswrapper[4556]: I0218 09:07:05.910324 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="9db33ada-017a-4281-85c1-0c1530bc3bbc" containerName="oauth-openshift" Feb 18 09:07:05 crc kubenswrapper[4556]: I0218 09:07:05.910656 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-6b68866bbb-df7kt" Feb 18 09:07:05 crc kubenswrapper[4556]: I0218 09:07:05.912534 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Feb 18 09:07:05 crc kubenswrapper[4556]: I0218 09:07:05.912937 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Feb 18 09:07:05 crc kubenswrapper[4556]: I0218 09:07:05.912941 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Feb 18 09:07:05 crc kubenswrapper[4556]: I0218 09:07:05.913715 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Feb 18 09:07:05 crc kubenswrapper[4556]: I0218 09:07:05.913812 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Feb 18 09:07:05 crc kubenswrapper[4556]: I0218 09:07:05.913814 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Feb 18 09:07:05 crc kubenswrapper[4556]: I0218 09:07:05.914046 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Feb 18 09:07:05 crc kubenswrapper[4556]: I0218 09:07:05.914395 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Feb 18 09:07:05 crc kubenswrapper[4556]: I0218 09:07:05.915123 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Feb 18 09:07:05 crc kubenswrapper[4556]: I0218 09:07:05.915277 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Feb 18 09:07:05 crc kubenswrapper[4556]: I0218 09:07:05.916219 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 18 09:07:05 crc kubenswrapper[4556]: I0218 09:07:05.916655 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Feb 18 09:07:05 crc kubenswrapper[4556]: I0218 09:07:05.917819 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-6b68866bbb-df7kt"] Feb 18 09:07:05 crc kubenswrapper[4556]: I0218 09:07:05.922725 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Feb 18 09:07:05 crc kubenswrapper[4556]: I0218 09:07:05.925026 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Feb 18 09:07:05 crc kubenswrapper[4556]: I0218 09:07:05.930850 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Feb 18 09:07:05 crc kubenswrapper[4556]: I0218 09:07:05.999022 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e-v4-0-config-system-session\") pod \"oauth-openshift-6b68866bbb-df7kt\" (UID: \"ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e\") " pod="openshift-authentication/oauth-openshift-6b68866bbb-df7kt" Feb 18 09:07:05 crc kubenswrapper[4556]: I0218 09:07:05.999065 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e-v4-0-config-system-router-certs\") pod \"oauth-openshift-6b68866bbb-df7kt\" (UID: \"ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e\") " pod="openshift-authentication/oauth-openshift-6b68866bbb-df7kt" Feb 18 09:07:05 crc kubenswrapper[4556]: I0218 09:07:05.999085 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e-audit-dir\") pod \"oauth-openshift-6b68866bbb-df7kt\" (UID: \"ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e\") " pod="openshift-authentication/oauth-openshift-6b68866bbb-df7kt" Feb 18 09:07:05 crc kubenswrapper[4556]: I0218 09:07:05.999104 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kh4pp\" (UniqueName: \"kubernetes.io/projected/ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e-kube-api-access-kh4pp\") pod \"oauth-openshift-6b68866bbb-df7kt\" (UID: \"ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e\") " pod="openshift-authentication/oauth-openshift-6b68866bbb-df7kt" Feb 18 09:07:05 crc kubenswrapper[4556]: I0218 09:07:05.999204 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6b68866bbb-df7kt\" (UID: \"ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e\") " pod="openshift-authentication/oauth-openshift-6b68866bbb-df7kt" Feb 18 09:07:05 crc kubenswrapper[4556]: I0218 09:07:05.999310 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6b68866bbb-df7kt\" (UID: \"ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e\") " pod="openshift-authentication/oauth-openshift-6b68866bbb-df7kt" Feb 18 09:07:05 crc kubenswrapper[4556]: I0218 09:07:05.999334 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e-v4-0-config-system-service-ca\") pod \"oauth-openshift-6b68866bbb-df7kt\" (UID: \"ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e\") " pod="openshift-authentication/oauth-openshift-6b68866bbb-df7kt" Feb 18 09:07:05 crc kubenswrapper[4556]: I0218 09:07:05.999357 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6b68866bbb-df7kt\" (UID: \"ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e\") " pod="openshift-authentication/oauth-openshift-6b68866bbb-df7kt" Feb 18 09:07:05 crc kubenswrapper[4556]: I0218 09:07:05.999375 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6b68866bbb-df7kt\" (UID: \"ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e\") " pod="openshift-authentication/oauth-openshift-6b68866bbb-df7kt" Feb 18 09:07:05 crc kubenswrapper[4556]: I0218 09:07:05.999399 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6b68866bbb-df7kt\" (UID: \"ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e\") " pod="openshift-authentication/oauth-openshift-6b68866bbb-df7kt" Feb 18 09:07:05 crc kubenswrapper[4556]: I0218 09:07:05.999436 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e-v4-0-config-user-template-login\") pod \"oauth-openshift-6b68866bbb-df7kt\" (UID: \"ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e\") " pod="openshift-authentication/oauth-openshift-6b68866bbb-df7kt" Feb 18 09:07:05 crc kubenswrapper[4556]: I0218 09:07:05.999465 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e-v4-0-config-user-template-error\") pod \"oauth-openshift-6b68866bbb-df7kt\" (UID: \"ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e\") " pod="openshift-authentication/oauth-openshift-6b68866bbb-df7kt" Feb 18 09:07:05 crc kubenswrapper[4556]: I0218 09:07:05.999487 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e-audit-policies\") pod \"oauth-openshift-6b68866bbb-df7kt\" (UID: \"ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e\") " pod="openshift-authentication/oauth-openshift-6b68866bbb-df7kt" Feb 18 09:07:05 crc kubenswrapper[4556]: I0218 09:07:05.999502 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6b68866bbb-df7kt\" (UID: \"ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e\") " pod="openshift-authentication/oauth-openshift-6b68866bbb-df7kt" Feb 18 09:07:06 crc kubenswrapper[4556]: I0218 09:07:06.100008 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6b68866bbb-df7kt\" (UID: \"ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e\") " pod="openshift-authentication/oauth-openshift-6b68866bbb-df7kt" Feb 18 09:07:06 crc kubenswrapper[4556]: I0218 09:07:06.100043 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e-v4-0-config-system-service-ca\") pod \"oauth-openshift-6b68866bbb-df7kt\" (UID: \"ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e\") " pod="openshift-authentication/oauth-openshift-6b68866bbb-df7kt" Feb 18 09:07:06 crc kubenswrapper[4556]: I0218 09:07:06.100062 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6b68866bbb-df7kt\" (UID: \"ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e\") " pod="openshift-authentication/oauth-openshift-6b68866bbb-df7kt" Feb 18 09:07:06 crc kubenswrapper[4556]: I0218 09:07:06.100076 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6b68866bbb-df7kt\" (UID: \"ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e\") " pod="openshift-authentication/oauth-openshift-6b68866bbb-df7kt" Feb 18 09:07:06 crc kubenswrapper[4556]: I0218 09:07:06.100098 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6b68866bbb-df7kt\" (UID: \"ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e\") " pod="openshift-authentication/oauth-openshift-6b68866bbb-df7kt" Feb 18 09:07:06 crc kubenswrapper[4556]: I0218 09:07:06.100116 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e-v4-0-config-user-template-login\") pod \"oauth-openshift-6b68866bbb-df7kt\" (UID: \"ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e\") " pod="openshift-authentication/oauth-openshift-6b68866bbb-df7kt" Feb 18 09:07:06 crc kubenswrapper[4556]: I0218 09:07:06.100141 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e-v4-0-config-user-template-error\") pod \"oauth-openshift-6b68866bbb-df7kt\" (UID: \"ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e\") " pod="openshift-authentication/oauth-openshift-6b68866bbb-df7kt" Feb 18 09:07:06 crc kubenswrapper[4556]: I0218 09:07:06.100177 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e-audit-policies\") pod \"oauth-openshift-6b68866bbb-df7kt\" (UID: \"ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e\") " pod="openshift-authentication/oauth-openshift-6b68866bbb-df7kt" Feb 18 09:07:06 crc kubenswrapper[4556]: I0218 09:07:06.100191 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6b68866bbb-df7kt\" (UID: \"ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e\") " pod="openshift-authentication/oauth-openshift-6b68866bbb-df7kt" Feb 18 09:07:06 crc kubenswrapper[4556]: I0218 09:07:06.100208 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e-v4-0-config-system-session\") pod \"oauth-openshift-6b68866bbb-df7kt\" (UID: \"ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e\") " pod="openshift-authentication/oauth-openshift-6b68866bbb-df7kt" Feb 18 09:07:06 crc kubenswrapper[4556]: I0218 09:07:06.100229 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e-v4-0-config-system-router-certs\") pod \"oauth-openshift-6b68866bbb-df7kt\" (UID: \"ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e\") " pod="openshift-authentication/oauth-openshift-6b68866bbb-df7kt" Feb 18 09:07:06 crc kubenswrapper[4556]: I0218 09:07:06.100245 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e-audit-dir\") pod \"oauth-openshift-6b68866bbb-df7kt\" (UID: \"ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e\") " pod="openshift-authentication/oauth-openshift-6b68866bbb-df7kt" Feb 18 09:07:06 crc kubenswrapper[4556]: I0218 09:07:06.100263 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kh4pp\" (UniqueName: \"kubernetes.io/projected/ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e-kube-api-access-kh4pp\") pod \"oauth-openshift-6b68866bbb-df7kt\" (UID: \"ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e\") " pod="openshift-authentication/oauth-openshift-6b68866bbb-df7kt" Feb 18 09:07:06 crc kubenswrapper[4556]: I0218 09:07:06.100283 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6b68866bbb-df7kt\" (UID: \"ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e\") " pod="openshift-authentication/oauth-openshift-6b68866bbb-df7kt" Feb 18 09:07:06 crc kubenswrapper[4556]: I0218 09:07:06.100390 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e-audit-dir\") pod \"oauth-openshift-6b68866bbb-df7kt\" (UID: \"ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e\") " pod="openshift-authentication/oauth-openshift-6b68866bbb-df7kt" Feb 18 09:07:06 crc kubenswrapper[4556]: I0218 09:07:06.101022 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e-v4-0-config-system-service-ca\") pod \"oauth-openshift-6b68866bbb-df7kt\" (UID: \"ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e\") " pod="openshift-authentication/oauth-openshift-6b68866bbb-df7kt" Feb 18 09:07:06 crc kubenswrapper[4556]: I0218 09:07:06.101024 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6b68866bbb-df7kt\" (UID: \"ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e\") " pod="openshift-authentication/oauth-openshift-6b68866bbb-df7kt" Feb 18 09:07:06 crc kubenswrapper[4556]: I0218 09:07:06.101077 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6b68866bbb-df7kt\" (UID: \"ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e\") " pod="openshift-authentication/oauth-openshift-6b68866bbb-df7kt" Feb 18 09:07:06 crc kubenswrapper[4556]: I0218 09:07:06.101126 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e-audit-policies\") pod \"oauth-openshift-6b68866bbb-df7kt\" (UID: \"ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e\") " pod="openshift-authentication/oauth-openshift-6b68866bbb-df7kt" Feb 18 09:07:06 crc kubenswrapper[4556]: I0218 09:07:06.105012 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6b68866bbb-df7kt\" (UID: \"ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e\") " pod="openshift-authentication/oauth-openshift-6b68866bbb-df7kt" Feb 18 09:07:06 crc kubenswrapper[4556]: I0218 09:07:06.105043 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6b68866bbb-df7kt\" (UID: \"ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e\") " pod="openshift-authentication/oauth-openshift-6b68866bbb-df7kt" Feb 18 09:07:06 crc kubenswrapper[4556]: I0218 09:07:06.105053 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e-v4-0-config-system-session\") pod \"oauth-openshift-6b68866bbb-df7kt\" (UID: \"ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e\") " pod="openshift-authentication/oauth-openshift-6b68866bbb-df7kt" Feb 18 09:07:06 crc kubenswrapper[4556]: I0218 09:07:06.105269 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e-v4-0-config-user-template-error\") pod \"oauth-openshift-6b68866bbb-df7kt\" (UID: \"ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e\") " pod="openshift-authentication/oauth-openshift-6b68866bbb-df7kt" Feb 18 09:07:06 crc kubenswrapper[4556]: I0218 09:07:06.105373 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6b68866bbb-df7kt\" (UID: \"ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e\") " pod="openshift-authentication/oauth-openshift-6b68866bbb-df7kt" Feb 18 09:07:06 crc kubenswrapper[4556]: I0218 09:07:06.105430 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6b68866bbb-df7kt\" (UID: \"ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e\") " pod="openshift-authentication/oauth-openshift-6b68866bbb-df7kt" Feb 18 09:07:06 crc kubenswrapper[4556]: I0218 09:07:06.105500 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e-v4-0-config-user-template-login\") pod \"oauth-openshift-6b68866bbb-df7kt\" (UID: \"ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e\") " pod="openshift-authentication/oauth-openshift-6b68866bbb-df7kt" Feb 18 09:07:06 crc kubenswrapper[4556]: I0218 09:07:06.105883 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e-v4-0-config-system-router-certs\") pod \"oauth-openshift-6b68866bbb-df7kt\" (UID: \"ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e\") " pod="openshift-authentication/oauth-openshift-6b68866bbb-df7kt" Feb 18 09:07:06 crc kubenswrapper[4556]: I0218 09:07:06.112024 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kh4pp\" (UniqueName: \"kubernetes.io/projected/ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e-kube-api-access-kh4pp\") pod \"oauth-openshift-6b68866bbb-df7kt\" (UID: \"ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e\") " pod="openshift-authentication/oauth-openshift-6b68866bbb-df7kt" Feb 18 09:07:06 crc kubenswrapper[4556]: I0218 09:07:06.221977 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-6b68866bbb-df7kt" Feb 18 09:07:06 crc kubenswrapper[4556]: I0218 09:07:06.593732 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-6b68866bbb-df7kt"] Feb 18 09:07:07 crc kubenswrapper[4556]: I0218 09:07:07.041846 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-6b68866bbb-df7kt" event={"ID":"ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e","Type":"ContainerStarted","Data":"47853d6a5ed9fe8dc6fc6dc29e067e308d4c3e5ef93758e4fe8ecd97f4047ef7"} Feb 18 09:07:07 crc kubenswrapper[4556]: I0218 09:07:07.042082 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-6b68866bbb-df7kt" event={"ID":"ab09b298-c5b3-4ceb-b5a9-cef4399d0c1e","Type":"ContainerStarted","Data":"97e4a2f38ac6ee3550a52eba65b78e45211d1c089cd464f1623846cd41a350a1"} Feb 18 09:07:07 crc kubenswrapper[4556]: I0218 09:07:07.042921 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-6b68866bbb-df7kt" Feb 18 09:07:07 crc kubenswrapper[4556]: I0218 09:07:07.194641 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-6b68866bbb-df7kt" Feb 18 09:07:07 crc kubenswrapper[4556]: I0218 09:07:07.209800 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-6b68866bbb-df7kt" podStartSLOduration=30.209782719 podStartE2EDuration="30.209782719s" podCreationTimestamp="2026-02-18 09:06:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:07:07.058111803 +0000 UTC m=+184.075072783" watchObservedRunningTime="2026-02-18 09:07:07.209782719 +0000 UTC m=+184.226743699" Feb 18 09:07:09 crc kubenswrapper[4556]: I0218 09:07:09.501873 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.276488 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-brf28"] Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.277280 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-brf28" podUID="eac84811-d54a-47c2-ae5e-77f3a531fc4f" containerName="registry-server" containerID="cri-o://5bb7c88c3a0aaf429e291c99988ed6c53f910bd094b7e378bae706c0b21db144" gracePeriod=30 Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.287966 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-sd4dx"] Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.288464 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-sd4dx" podUID="90477a11-cc7c-48cf-a210-9d1d802360ce" containerName="registry-server" containerID="cri-o://832cdf04bbfa34d0e4091a6792478f2d3cbe62887d49c17406613a2b92772956" gracePeriod=30 Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.295242 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vc7vq"] Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.295364 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-vc7vq" podUID="7fb132fe-af75-4b1a-828f-5b1dcd074fd6" containerName="marketplace-operator" containerID="cri-o://e01ec1bf5ac6396edac9a10179c99bb15ddd5777cc18550b56abeffcd1e6e39d" gracePeriod=30 Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.309224 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ndnpr"] Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.309518 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-ndnpr" podUID="e23bd1bf-79ec-4cc0-95c4-16959eab3aa7" containerName="registry-server" containerID="cri-o://2b7ded25b1a59a7579dc6d99250e77bc4deae52ae4a074b1393c9ac3324ab332" gracePeriod=30 Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.313439 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vh264"] Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.314176 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-vh264" Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.322984 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bwwg2"] Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.323282 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bwwg2" podUID="fdd9dfd4-872b-42fc-960f-5d4c51e6e66e" containerName="registry-server" containerID="cri-o://aeff40b439e6e8751290c8681fc2c889491f6e822c6c530d08c1d651305a72f6" gracePeriod=30 Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.328749 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vh264"] Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.411908 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/070f724e-944f-48d2-bb39-67b98f7667ce-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-vh264\" (UID: \"070f724e-944f-48d2-bb39-67b98f7667ce\") " pod="openshift-marketplace/marketplace-operator-79b997595-vh264" Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.411967 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7lntv\" (UniqueName: \"kubernetes.io/projected/070f724e-944f-48d2-bb39-67b98f7667ce-kube-api-access-7lntv\") pod \"marketplace-operator-79b997595-vh264\" (UID: \"070f724e-944f-48d2-bb39-67b98f7667ce\") " pod="openshift-marketplace/marketplace-operator-79b997595-vh264" Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.411995 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/070f724e-944f-48d2-bb39-67b98f7667ce-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-vh264\" (UID: \"070f724e-944f-48d2-bb39-67b98f7667ce\") " pod="openshift-marketplace/marketplace-operator-79b997595-vh264" Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.513004 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/070f724e-944f-48d2-bb39-67b98f7667ce-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-vh264\" (UID: \"070f724e-944f-48d2-bb39-67b98f7667ce\") " pod="openshift-marketplace/marketplace-operator-79b997595-vh264" Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.513072 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7lntv\" (UniqueName: \"kubernetes.io/projected/070f724e-944f-48d2-bb39-67b98f7667ce-kube-api-access-7lntv\") pod \"marketplace-operator-79b997595-vh264\" (UID: \"070f724e-944f-48d2-bb39-67b98f7667ce\") " pod="openshift-marketplace/marketplace-operator-79b997595-vh264" Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.513098 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/070f724e-944f-48d2-bb39-67b98f7667ce-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-vh264\" (UID: \"070f724e-944f-48d2-bb39-67b98f7667ce\") " pod="openshift-marketplace/marketplace-operator-79b997595-vh264" Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.514097 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/070f724e-944f-48d2-bb39-67b98f7667ce-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-vh264\" (UID: \"070f724e-944f-48d2-bb39-67b98f7667ce\") " pod="openshift-marketplace/marketplace-operator-79b997595-vh264" Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.522269 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/070f724e-944f-48d2-bb39-67b98f7667ce-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-vh264\" (UID: \"070f724e-944f-48d2-bb39-67b98f7667ce\") " pod="openshift-marketplace/marketplace-operator-79b997595-vh264" Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.528916 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7lntv\" (UniqueName: \"kubernetes.io/projected/070f724e-944f-48d2-bb39-67b98f7667ce-kube-api-access-7lntv\") pod \"marketplace-operator-79b997595-vh264\" (UID: \"070f724e-944f-48d2-bb39-67b98f7667ce\") " pod="openshift-marketplace/marketplace-operator-79b997595-vh264" Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.630384 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-vh264" Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.713293 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-brf28" Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.722242 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-vc7vq" Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.722293 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sd4dx" Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.727257 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ndnpr" Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.733648 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bwwg2" Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.812709 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vh264"] Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.816399 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h4pnl\" (UniqueName: \"kubernetes.io/projected/7fb132fe-af75-4b1a-828f-5b1dcd074fd6-kube-api-access-h4pnl\") pod \"7fb132fe-af75-4b1a-828f-5b1dcd074fd6\" (UID: \"7fb132fe-af75-4b1a-828f-5b1dcd074fd6\") " Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.816470 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jmc75\" (UniqueName: \"kubernetes.io/projected/eac84811-d54a-47c2-ae5e-77f3a531fc4f-kube-api-access-jmc75\") pod \"eac84811-d54a-47c2-ae5e-77f3a531fc4f\" (UID: \"eac84811-d54a-47c2-ae5e-77f3a531fc4f\") " Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.816497 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90477a11-cc7c-48cf-a210-9d1d802360ce-catalog-content\") pod \"90477a11-cc7c-48cf-a210-9d1d802360ce\" (UID: \"90477a11-cc7c-48cf-a210-9d1d802360ce\") " Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.816531 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eac84811-d54a-47c2-ae5e-77f3a531fc4f-catalog-content\") pod \"eac84811-d54a-47c2-ae5e-77f3a531fc4f\" (UID: \"eac84811-d54a-47c2-ae5e-77f3a531fc4f\") " Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.816555 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90477a11-cc7c-48cf-a210-9d1d802360ce-utilities\") pod \"90477a11-cc7c-48cf-a210-9d1d802360ce\" (UID: \"90477a11-cc7c-48cf-a210-9d1d802360ce\") " Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.816569 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fdd9dfd4-872b-42fc-960f-5d4c51e6e66e-utilities\") pod \"fdd9dfd4-872b-42fc-960f-5d4c51e6e66e\" (UID: \"fdd9dfd4-872b-42fc-960f-5d4c51e6e66e\") " Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.816599 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7fb132fe-af75-4b1a-828f-5b1dcd074fd6-marketplace-operator-metrics\") pod \"7fb132fe-af75-4b1a-828f-5b1dcd074fd6\" (UID: \"7fb132fe-af75-4b1a-828f-5b1dcd074fd6\") " Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.816616 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2xgm2\" (UniqueName: \"kubernetes.io/projected/fdd9dfd4-872b-42fc-960f-5d4c51e6e66e-kube-api-access-2xgm2\") pod \"fdd9dfd4-872b-42fc-960f-5d4c51e6e66e\" (UID: \"fdd9dfd4-872b-42fc-960f-5d4c51e6e66e\") " Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.816653 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e23bd1bf-79ec-4cc0-95c4-16959eab3aa7-catalog-content\") pod \"e23bd1bf-79ec-4cc0-95c4-16959eab3aa7\" (UID: \"e23bd1bf-79ec-4cc0-95c4-16959eab3aa7\") " Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.816677 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngl5n\" (UniqueName: \"kubernetes.io/projected/e23bd1bf-79ec-4cc0-95c4-16959eab3aa7-kube-api-access-ngl5n\") pod \"e23bd1bf-79ec-4cc0-95c4-16959eab3aa7\" (UID: \"e23bd1bf-79ec-4cc0-95c4-16959eab3aa7\") " Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.816697 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rcdq4\" (UniqueName: \"kubernetes.io/projected/90477a11-cc7c-48cf-a210-9d1d802360ce-kube-api-access-rcdq4\") pod \"90477a11-cc7c-48cf-a210-9d1d802360ce\" (UID: \"90477a11-cc7c-48cf-a210-9d1d802360ce\") " Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.816716 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7fb132fe-af75-4b1a-828f-5b1dcd074fd6-marketplace-trusted-ca\") pod \"7fb132fe-af75-4b1a-828f-5b1dcd074fd6\" (UID: \"7fb132fe-af75-4b1a-828f-5b1dcd074fd6\") " Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.816735 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eac84811-d54a-47c2-ae5e-77f3a531fc4f-utilities\") pod \"eac84811-d54a-47c2-ae5e-77f3a531fc4f\" (UID: \"eac84811-d54a-47c2-ae5e-77f3a531fc4f\") " Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.816749 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fdd9dfd4-872b-42fc-960f-5d4c51e6e66e-catalog-content\") pod \"fdd9dfd4-872b-42fc-960f-5d4c51e6e66e\" (UID: \"fdd9dfd4-872b-42fc-960f-5d4c51e6e66e\") " Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.816765 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e23bd1bf-79ec-4cc0-95c4-16959eab3aa7-utilities\") pod \"e23bd1bf-79ec-4cc0-95c4-16959eab3aa7\" (UID: \"e23bd1bf-79ec-4cc0-95c4-16959eab3aa7\") " Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.818057 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fdd9dfd4-872b-42fc-960f-5d4c51e6e66e-utilities" (OuterVolumeSpecName: "utilities") pod "fdd9dfd4-872b-42fc-960f-5d4c51e6e66e" (UID: "fdd9dfd4-872b-42fc-960f-5d4c51e6e66e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.818626 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e23bd1bf-79ec-4cc0-95c4-16959eab3aa7-utilities" (OuterVolumeSpecName: "utilities") pod "e23bd1bf-79ec-4cc0-95c4-16959eab3aa7" (UID: "e23bd1bf-79ec-4cc0-95c4-16959eab3aa7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.818865 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7fb132fe-af75-4b1a-828f-5b1dcd074fd6-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "7fb132fe-af75-4b1a-828f-5b1dcd074fd6" (UID: "7fb132fe-af75-4b1a-828f-5b1dcd074fd6"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.818969 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eac84811-d54a-47c2-ae5e-77f3a531fc4f-utilities" (OuterVolumeSpecName: "utilities") pod "eac84811-d54a-47c2-ae5e-77f3a531fc4f" (UID: "eac84811-d54a-47c2-ae5e-77f3a531fc4f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.821054 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eac84811-d54a-47c2-ae5e-77f3a531fc4f-kube-api-access-jmc75" (OuterVolumeSpecName: "kube-api-access-jmc75") pod "eac84811-d54a-47c2-ae5e-77f3a531fc4f" (UID: "eac84811-d54a-47c2-ae5e-77f3a531fc4f"). InnerVolumeSpecName "kube-api-access-jmc75". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.821547 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fb132fe-af75-4b1a-828f-5b1dcd074fd6-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "7fb132fe-af75-4b1a-828f-5b1dcd074fd6" (UID: "7fb132fe-af75-4b1a-828f-5b1dcd074fd6"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.823937 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e23bd1bf-79ec-4cc0-95c4-16959eab3aa7-kube-api-access-ngl5n" (OuterVolumeSpecName: "kube-api-access-ngl5n") pod "e23bd1bf-79ec-4cc0-95c4-16959eab3aa7" (UID: "e23bd1bf-79ec-4cc0-95c4-16959eab3aa7"). InnerVolumeSpecName "kube-api-access-ngl5n". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.827545 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fb132fe-af75-4b1a-828f-5b1dcd074fd6-kube-api-access-h4pnl" (OuterVolumeSpecName: "kube-api-access-h4pnl") pod "7fb132fe-af75-4b1a-828f-5b1dcd074fd6" (UID: "7fb132fe-af75-4b1a-828f-5b1dcd074fd6"). InnerVolumeSpecName "kube-api-access-h4pnl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.827732 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/90477a11-cc7c-48cf-a210-9d1d802360ce-utilities" (OuterVolumeSpecName: "utilities") pod "90477a11-cc7c-48cf-a210-9d1d802360ce" (UID: "90477a11-cc7c-48cf-a210-9d1d802360ce"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.830316 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fdd9dfd4-872b-42fc-960f-5d4c51e6e66e-kube-api-access-2xgm2" (OuterVolumeSpecName: "kube-api-access-2xgm2") pod "fdd9dfd4-872b-42fc-960f-5d4c51e6e66e" (UID: "fdd9dfd4-872b-42fc-960f-5d4c51e6e66e"). InnerVolumeSpecName "kube-api-access-2xgm2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.832223 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90477a11-cc7c-48cf-a210-9d1d802360ce-kube-api-access-rcdq4" (OuterVolumeSpecName: "kube-api-access-rcdq4") pod "90477a11-cc7c-48cf-a210-9d1d802360ce" (UID: "90477a11-cc7c-48cf-a210-9d1d802360ce"). InnerVolumeSpecName "kube-api-access-rcdq4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.848188 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e23bd1bf-79ec-4cc0-95c4-16959eab3aa7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e23bd1bf-79ec-4cc0-95c4-16959eab3aa7" (UID: "e23bd1bf-79ec-4cc0-95c4-16959eab3aa7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.869413 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/90477a11-cc7c-48cf-a210-9d1d802360ce-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "90477a11-cc7c-48cf-a210-9d1d802360ce" (UID: "90477a11-cc7c-48cf-a210-9d1d802360ce"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.875689 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eac84811-d54a-47c2-ae5e-77f3a531fc4f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "eac84811-d54a-47c2-ae5e-77f3a531fc4f" (UID: "eac84811-d54a-47c2-ae5e-77f3a531fc4f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.918318 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jmc75\" (UniqueName: \"kubernetes.io/projected/eac84811-d54a-47c2-ae5e-77f3a531fc4f-kube-api-access-jmc75\") on node \"crc\" DevicePath \"\"" Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.918454 4556 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90477a11-cc7c-48cf-a210-9d1d802360ce-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.918533 4556 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eac84811-d54a-47c2-ae5e-77f3a531fc4f-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.918623 4556 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90477a11-cc7c-48cf-a210-9d1d802360ce-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.918700 4556 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fdd9dfd4-872b-42fc-960f-5d4c51e6e66e-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.918860 4556 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7fb132fe-af75-4b1a-828f-5b1dcd074fd6-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.918991 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2xgm2\" (UniqueName: \"kubernetes.io/projected/fdd9dfd4-872b-42fc-960f-5d4c51e6e66e-kube-api-access-2xgm2\") on node \"crc\" DevicePath \"\"" Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.919076 4556 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e23bd1bf-79ec-4cc0-95c4-16959eab3aa7-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.919147 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngl5n\" (UniqueName: \"kubernetes.io/projected/e23bd1bf-79ec-4cc0-95c4-16959eab3aa7-kube-api-access-ngl5n\") on node \"crc\" DevicePath \"\"" Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.919235 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rcdq4\" (UniqueName: \"kubernetes.io/projected/90477a11-cc7c-48cf-a210-9d1d802360ce-kube-api-access-rcdq4\") on node \"crc\" DevicePath \"\"" Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.919318 4556 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7fb132fe-af75-4b1a-828f-5b1dcd074fd6-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.919395 4556 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eac84811-d54a-47c2-ae5e-77f3a531fc4f-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.919472 4556 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e23bd1bf-79ec-4cc0-95c4-16959eab3aa7-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.919551 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h4pnl\" (UniqueName: \"kubernetes.io/projected/7fb132fe-af75-4b1a-828f-5b1dcd074fd6-kube-api-access-h4pnl\") on node \"crc\" DevicePath \"\"" Feb 18 09:07:16 crc kubenswrapper[4556]: I0218 09:07:16.936856 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fdd9dfd4-872b-42fc-960f-5d4c51e6e66e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fdd9dfd4-872b-42fc-960f-5d4c51e6e66e" (UID: "fdd9dfd4-872b-42fc-960f-5d4c51e6e66e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.020637 4556 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fdd9dfd4-872b-42fc-960f-5d4c51e6e66e-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.090042 4556 generic.go:334] "Generic (PLEG): container finished" podID="7fb132fe-af75-4b1a-828f-5b1dcd074fd6" containerID="e01ec1bf5ac6396edac9a10179c99bb15ddd5777cc18550b56abeffcd1e6e39d" exitCode=0 Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.090147 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-vc7vq" Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.090167 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-vc7vq" event={"ID":"7fb132fe-af75-4b1a-828f-5b1dcd074fd6","Type":"ContainerDied","Data":"e01ec1bf5ac6396edac9a10179c99bb15ddd5777cc18550b56abeffcd1e6e39d"} Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.090299 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-vc7vq" event={"ID":"7fb132fe-af75-4b1a-828f-5b1dcd074fd6","Type":"ContainerDied","Data":"8b09bc452f5187c3acd8beed22526cbb2d46907fea2cb19de628c721e12705ed"} Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.090322 4556 scope.go:117] "RemoveContainer" containerID="e01ec1bf5ac6396edac9a10179c99bb15ddd5777cc18550b56abeffcd1e6e39d" Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.093677 4556 generic.go:334] "Generic (PLEG): container finished" podID="e23bd1bf-79ec-4cc0-95c4-16959eab3aa7" containerID="2b7ded25b1a59a7579dc6d99250e77bc4deae52ae4a074b1393c9ac3324ab332" exitCode=0 Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.093743 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ndnpr" Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.093748 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ndnpr" event={"ID":"e23bd1bf-79ec-4cc0-95c4-16959eab3aa7","Type":"ContainerDied","Data":"2b7ded25b1a59a7579dc6d99250e77bc4deae52ae4a074b1393c9ac3324ab332"} Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.094104 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ndnpr" event={"ID":"e23bd1bf-79ec-4cc0-95c4-16959eab3aa7","Type":"ContainerDied","Data":"4804b76870038a935a002ae8e9b4d53a56df4542126f2424cd995637a12631b8"} Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.095360 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-vh264" event={"ID":"070f724e-944f-48d2-bb39-67b98f7667ce","Type":"ContainerStarted","Data":"3852d490c4a1b8eec371fc67f9290731be561ed140504657231e9ef34daa62c8"} Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.095387 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-vh264" event={"ID":"070f724e-944f-48d2-bb39-67b98f7667ce","Type":"ContainerStarted","Data":"5ee83e7e223a121186858b4969a117e81937feb601893154c8b9ad0627c0fcc3"} Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.097967 4556 generic.go:334] "Generic (PLEG): container finished" podID="90477a11-cc7c-48cf-a210-9d1d802360ce" containerID="832cdf04bbfa34d0e4091a6792478f2d3cbe62887d49c17406613a2b92772956" exitCode=0 Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.098030 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sd4dx" event={"ID":"90477a11-cc7c-48cf-a210-9d1d802360ce","Type":"ContainerDied","Data":"832cdf04bbfa34d0e4091a6792478f2d3cbe62887d49c17406613a2b92772956"} Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.098056 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sd4dx" event={"ID":"90477a11-cc7c-48cf-a210-9d1d802360ce","Type":"ContainerDied","Data":"1606cfc7ff8627098dd6180d64cf31ed84f3331507f250eaa8b250f3995e57a1"} Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.098118 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sd4dx" Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.102091 4556 generic.go:334] "Generic (PLEG): container finished" podID="eac84811-d54a-47c2-ae5e-77f3a531fc4f" containerID="5bb7c88c3a0aaf429e291c99988ed6c53f910bd094b7e378bae706c0b21db144" exitCode=0 Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.102147 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-brf28" event={"ID":"eac84811-d54a-47c2-ae5e-77f3a531fc4f","Type":"ContainerDied","Data":"5bb7c88c3a0aaf429e291c99988ed6c53f910bd094b7e378bae706c0b21db144"} Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.102184 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-brf28" event={"ID":"eac84811-d54a-47c2-ae5e-77f3a531fc4f","Type":"ContainerDied","Data":"8630aa61bf54c6289842987772ae6787bdc62980758815c13ed806b02bc25265"} Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.102252 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-brf28" Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.107678 4556 generic.go:334] "Generic (PLEG): container finished" podID="fdd9dfd4-872b-42fc-960f-5d4c51e6e66e" containerID="aeff40b439e6e8751290c8681fc2c889491f6e822c6c530d08c1d651305a72f6" exitCode=0 Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.107703 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bwwg2" event={"ID":"fdd9dfd4-872b-42fc-960f-5d4c51e6e66e","Type":"ContainerDied","Data":"aeff40b439e6e8751290c8681fc2c889491f6e822c6c530d08c1d651305a72f6"} Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.107721 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bwwg2" event={"ID":"fdd9dfd4-872b-42fc-960f-5d4c51e6e66e","Type":"ContainerDied","Data":"887663eb39f115ab07d951379bc7310153d46c4bf8928ed2dc23a27df0654b1e"} Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.107729 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bwwg2" Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.108480 4556 scope.go:117] "RemoveContainer" containerID="e01ec1bf5ac6396edac9a10179c99bb15ddd5777cc18550b56abeffcd1e6e39d" Feb 18 09:07:17 crc kubenswrapper[4556]: E0218 09:07:17.108819 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e01ec1bf5ac6396edac9a10179c99bb15ddd5777cc18550b56abeffcd1e6e39d\": container with ID starting with e01ec1bf5ac6396edac9a10179c99bb15ddd5777cc18550b56abeffcd1e6e39d not found: ID does not exist" containerID="e01ec1bf5ac6396edac9a10179c99bb15ddd5777cc18550b56abeffcd1e6e39d" Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.108844 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e01ec1bf5ac6396edac9a10179c99bb15ddd5777cc18550b56abeffcd1e6e39d"} err="failed to get container status \"e01ec1bf5ac6396edac9a10179c99bb15ddd5777cc18550b56abeffcd1e6e39d\": rpc error: code = NotFound desc = could not find container \"e01ec1bf5ac6396edac9a10179c99bb15ddd5777cc18550b56abeffcd1e6e39d\": container with ID starting with e01ec1bf5ac6396edac9a10179c99bb15ddd5777cc18550b56abeffcd1e6e39d not found: ID does not exist" Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.108865 4556 scope.go:117] "RemoveContainer" containerID="2b7ded25b1a59a7579dc6d99250e77bc4deae52ae4a074b1393c9ac3324ab332" Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.130967 4556 scope.go:117] "RemoveContainer" containerID="737dda86681f77545e2f061e569a8489f2e9b178bdbffe9bc502e44f548eea16" Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.141389 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-vh264" podStartSLOduration=1.141328688 podStartE2EDuration="1.141328688s" podCreationTimestamp="2026-02-18 09:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:07:17.115714225 +0000 UTC m=+194.132675205" watchObservedRunningTime="2026-02-18 09:07:17.141328688 +0000 UTC m=+194.158289668" Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.149904 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-sd4dx"] Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.156621 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-sd4dx"] Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.158889 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bwwg2"] Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.162399 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bwwg2"] Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.167452 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ndnpr"] Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.168793 4556 scope.go:117] "RemoveContainer" containerID="fb51382c701dac9eff637f394b33ff49ba492534b3ec98a703fe8770b728aab1" Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.174242 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-ndnpr"] Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.182004 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vc7vq"] Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.182763 4556 scope.go:117] "RemoveContainer" containerID="2b7ded25b1a59a7579dc6d99250e77bc4deae52ae4a074b1393c9ac3324ab332" Feb 18 09:07:17 crc kubenswrapper[4556]: E0218 09:07:17.183130 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b7ded25b1a59a7579dc6d99250e77bc4deae52ae4a074b1393c9ac3324ab332\": container with ID starting with 2b7ded25b1a59a7579dc6d99250e77bc4deae52ae4a074b1393c9ac3324ab332 not found: ID does not exist" containerID="2b7ded25b1a59a7579dc6d99250e77bc4deae52ae4a074b1393c9ac3324ab332" Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.183180 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b7ded25b1a59a7579dc6d99250e77bc4deae52ae4a074b1393c9ac3324ab332"} err="failed to get container status \"2b7ded25b1a59a7579dc6d99250e77bc4deae52ae4a074b1393c9ac3324ab332\": rpc error: code = NotFound desc = could not find container \"2b7ded25b1a59a7579dc6d99250e77bc4deae52ae4a074b1393c9ac3324ab332\": container with ID starting with 2b7ded25b1a59a7579dc6d99250e77bc4deae52ae4a074b1393c9ac3324ab332 not found: ID does not exist" Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.183210 4556 scope.go:117] "RemoveContainer" containerID="737dda86681f77545e2f061e569a8489f2e9b178bdbffe9bc502e44f548eea16" Feb 18 09:07:17 crc kubenswrapper[4556]: E0218 09:07:17.183569 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"737dda86681f77545e2f061e569a8489f2e9b178bdbffe9bc502e44f548eea16\": container with ID starting with 737dda86681f77545e2f061e569a8489f2e9b178bdbffe9bc502e44f548eea16 not found: ID does not exist" containerID="737dda86681f77545e2f061e569a8489f2e9b178bdbffe9bc502e44f548eea16" Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.183614 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"737dda86681f77545e2f061e569a8489f2e9b178bdbffe9bc502e44f548eea16"} err="failed to get container status \"737dda86681f77545e2f061e569a8489f2e9b178bdbffe9bc502e44f548eea16\": rpc error: code = NotFound desc = could not find container \"737dda86681f77545e2f061e569a8489f2e9b178bdbffe9bc502e44f548eea16\": container with ID starting with 737dda86681f77545e2f061e569a8489f2e9b178bdbffe9bc502e44f548eea16 not found: ID does not exist" Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.183646 4556 scope.go:117] "RemoveContainer" containerID="fb51382c701dac9eff637f394b33ff49ba492534b3ec98a703fe8770b728aab1" Feb 18 09:07:17 crc kubenswrapper[4556]: E0218 09:07:17.183905 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb51382c701dac9eff637f394b33ff49ba492534b3ec98a703fe8770b728aab1\": container with ID starting with fb51382c701dac9eff637f394b33ff49ba492534b3ec98a703fe8770b728aab1 not found: ID does not exist" containerID="fb51382c701dac9eff637f394b33ff49ba492534b3ec98a703fe8770b728aab1" Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.183937 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb51382c701dac9eff637f394b33ff49ba492534b3ec98a703fe8770b728aab1"} err="failed to get container status \"fb51382c701dac9eff637f394b33ff49ba492534b3ec98a703fe8770b728aab1\": rpc error: code = NotFound desc = could not find container \"fb51382c701dac9eff637f394b33ff49ba492534b3ec98a703fe8770b728aab1\": container with ID starting with fb51382c701dac9eff637f394b33ff49ba492534b3ec98a703fe8770b728aab1 not found: ID does not exist" Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.183963 4556 scope.go:117] "RemoveContainer" containerID="832cdf04bbfa34d0e4091a6792478f2d3cbe62887d49c17406613a2b92772956" Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.187276 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vc7vq"] Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.188961 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-brf28"] Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.190838 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-brf28"] Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.197098 4556 scope.go:117] "RemoveContainer" containerID="c0208355264ca4b2cb097ab6e495c05ecd84248a539dc45f09174336766c649a" Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.211249 4556 scope.go:117] "RemoveContainer" containerID="c623b3371772b683d554cf6cc7a299bd1973b82f4a7504608f10972d19a633e9" Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.224578 4556 scope.go:117] "RemoveContainer" containerID="832cdf04bbfa34d0e4091a6792478f2d3cbe62887d49c17406613a2b92772956" Feb 18 09:07:17 crc kubenswrapper[4556]: E0218 09:07:17.224862 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"832cdf04bbfa34d0e4091a6792478f2d3cbe62887d49c17406613a2b92772956\": container with ID starting with 832cdf04bbfa34d0e4091a6792478f2d3cbe62887d49c17406613a2b92772956 not found: ID does not exist" containerID="832cdf04bbfa34d0e4091a6792478f2d3cbe62887d49c17406613a2b92772956" Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.224896 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"832cdf04bbfa34d0e4091a6792478f2d3cbe62887d49c17406613a2b92772956"} err="failed to get container status \"832cdf04bbfa34d0e4091a6792478f2d3cbe62887d49c17406613a2b92772956\": rpc error: code = NotFound desc = could not find container \"832cdf04bbfa34d0e4091a6792478f2d3cbe62887d49c17406613a2b92772956\": container with ID starting with 832cdf04bbfa34d0e4091a6792478f2d3cbe62887d49c17406613a2b92772956 not found: ID does not exist" Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.224921 4556 scope.go:117] "RemoveContainer" containerID="c0208355264ca4b2cb097ab6e495c05ecd84248a539dc45f09174336766c649a" Feb 18 09:07:17 crc kubenswrapper[4556]: E0218 09:07:17.225410 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0208355264ca4b2cb097ab6e495c05ecd84248a539dc45f09174336766c649a\": container with ID starting with c0208355264ca4b2cb097ab6e495c05ecd84248a539dc45f09174336766c649a not found: ID does not exist" containerID="c0208355264ca4b2cb097ab6e495c05ecd84248a539dc45f09174336766c649a" Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.225437 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0208355264ca4b2cb097ab6e495c05ecd84248a539dc45f09174336766c649a"} err="failed to get container status \"c0208355264ca4b2cb097ab6e495c05ecd84248a539dc45f09174336766c649a\": rpc error: code = NotFound desc = could not find container \"c0208355264ca4b2cb097ab6e495c05ecd84248a539dc45f09174336766c649a\": container with ID starting with c0208355264ca4b2cb097ab6e495c05ecd84248a539dc45f09174336766c649a not found: ID does not exist" Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.225452 4556 scope.go:117] "RemoveContainer" containerID="c623b3371772b683d554cf6cc7a299bd1973b82f4a7504608f10972d19a633e9" Feb 18 09:07:17 crc kubenswrapper[4556]: E0218 09:07:17.225750 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c623b3371772b683d554cf6cc7a299bd1973b82f4a7504608f10972d19a633e9\": container with ID starting with c623b3371772b683d554cf6cc7a299bd1973b82f4a7504608f10972d19a633e9 not found: ID does not exist" containerID="c623b3371772b683d554cf6cc7a299bd1973b82f4a7504608f10972d19a633e9" Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.225780 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c623b3371772b683d554cf6cc7a299bd1973b82f4a7504608f10972d19a633e9"} err="failed to get container status \"c623b3371772b683d554cf6cc7a299bd1973b82f4a7504608f10972d19a633e9\": rpc error: code = NotFound desc = could not find container \"c623b3371772b683d554cf6cc7a299bd1973b82f4a7504608f10972d19a633e9\": container with ID starting with c623b3371772b683d554cf6cc7a299bd1973b82f4a7504608f10972d19a633e9 not found: ID does not exist" Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.225800 4556 scope.go:117] "RemoveContainer" containerID="5bb7c88c3a0aaf429e291c99988ed6c53f910bd094b7e378bae706c0b21db144" Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.237050 4556 scope.go:117] "RemoveContainer" containerID="0fce5558b046231804fca44b8ec9931210abf6bfc754bd3033f4e3a85819c0f1" Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.252656 4556 scope.go:117] "RemoveContainer" containerID="6f3e016b9541f7fb845611ae98130a4b33004793082ce622b1c15b7c7e02297d" Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.265623 4556 scope.go:117] "RemoveContainer" containerID="5bb7c88c3a0aaf429e291c99988ed6c53f910bd094b7e378bae706c0b21db144" Feb 18 09:07:17 crc kubenswrapper[4556]: E0218 09:07:17.266479 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5bb7c88c3a0aaf429e291c99988ed6c53f910bd094b7e378bae706c0b21db144\": container with ID starting with 5bb7c88c3a0aaf429e291c99988ed6c53f910bd094b7e378bae706c0b21db144 not found: ID does not exist" containerID="5bb7c88c3a0aaf429e291c99988ed6c53f910bd094b7e378bae706c0b21db144" Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.267549 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5bb7c88c3a0aaf429e291c99988ed6c53f910bd094b7e378bae706c0b21db144"} err="failed to get container status \"5bb7c88c3a0aaf429e291c99988ed6c53f910bd094b7e378bae706c0b21db144\": rpc error: code = NotFound desc = could not find container \"5bb7c88c3a0aaf429e291c99988ed6c53f910bd094b7e378bae706c0b21db144\": container with ID starting with 5bb7c88c3a0aaf429e291c99988ed6c53f910bd094b7e378bae706c0b21db144 not found: ID does not exist" Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.267574 4556 scope.go:117] "RemoveContainer" containerID="0fce5558b046231804fca44b8ec9931210abf6bfc754bd3033f4e3a85819c0f1" Feb 18 09:07:17 crc kubenswrapper[4556]: E0218 09:07:17.268179 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0fce5558b046231804fca44b8ec9931210abf6bfc754bd3033f4e3a85819c0f1\": container with ID starting with 0fce5558b046231804fca44b8ec9931210abf6bfc754bd3033f4e3a85819c0f1 not found: ID does not exist" containerID="0fce5558b046231804fca44b8ec9931210abf6bfc754bd3033f4e3a85819c0f1" Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.268220 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0fce5558b046231804fca44b8ec9931210abf6bfc754bd3033f4e3a85819c0f1"} err="failed to get container status \"0fce5558b046231804fca44b8ec9931210abf6bfc754bd3033f4e3a85819c0f1\": rpc error: code = NotFound desc = could not find container \"0fce5558b046231804fca44b8ec9931210abf6bfc754bd3033f4e3a85819c0f1\": container with ID starting with 0fce5558b046231804fca44b8ec9931210abf6bfc754bd3033f4e3a85819c0f1 not found: ID does not exist" Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.268247 4556 scope.go:117] "RemoveContainer" containerID="6f3e016b9541f7fb845611ae98130a4b33004793082ce622b1c15b7c7e02297d" Feb 18 09:07:17 crc kubenswrapper[4556]: E0218 09:07:17.268589 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f3e016b9541f7fb845611ae98130a4b33004793082ce622b1c15b7c7e02297d\": container with ID starting with 6f3e016b9541f7fb845611ae98130a4b33004793082ce622b1c15b7c7e02297d not found: ID does not exist" containerID="6f3e016b9541f7fb845611ae98130a4b33004793082ce622b1c15b7c7e02297d" Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.268614 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f3e016b9541f7fb845611ae98130a4b33004793082ce622b1c15b7c7e02297d"} err="failed to get container status \"6f3e016b9541f7fb845611ae98130a4b33004793082ce622b1c15b7c7e02297d\": rpc error: code = NotFound desc = could not find container \"6f3e016b9541f7fb845611ae98130a4b33004793082ce622b1c15b7c7e02297d\": container with ID starting with 6f3e016b9541f7fb845611ae98130a4b33004793082ce622b1c15b7c7e02297d not found: ID does not exist" Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.268630 4556 scope.go:117] "RemoveContainer" containerID="aeff40b439e6e8751290c8681fc2c889491f6e822c6c530d08c1d651305a72f6" Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.285062 4556 scope.go:117] "RemoveContainer" containerID="b1e538edcfe7d69af3ccabf373f71ff4f59d9eed02244cc39e055c7904706da5" Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.287682 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fb132fe-af75-4b1a-828f-5b1dcd074fd6" path="/var/lib/kubelet/pods/7fb132fe-af75-4b1a-828f-5b1dcd074fd6/volumes" Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.288129 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90477a11-cc7c-48cf-a210-9d1d802360ce" path="/var/lib/kubelet/pods/90477a11-cc7c-48cf-a210-9d1d802360ce/volumes" Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.289018 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e23bd1bf-79ec-4cc0-95c4-16959eab3aa7" path="/var/lib/kubelet/pods/e23bd1bf-79ec-4cc0-95c4-16959eab3aa7/volumes" Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.289562 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eac84811-d54a-47c2-ae5e-77f3a531fc4f" path="/var/lib/kubelet/pods/eac84811-d54a-47c2-ae5e-77f3a531fc4f/volumes" Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.290170 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fdd9dfd4-872b-42fc-960f-5d4c51e6e66e" path="/var/lib/kubelet/pods/fdd9dfd4-872b-42fc-960f-5d4c51e6e66e/volumes" Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.311118 4556 scope.go:117] "RemoveContainer" containerID="071fb0569134209452182b6f81b2cb0454f0201b5b5056bb3fdb2e29f34c768d" Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.324140 4556 scope.go:117] "RemoveContainer" containerID="aeff40b439e6e8751290c8681fc2c889491f6e822c6c530d08c1d651305a72f6" Feb 18 09:07:17 crc kubenswrapper[4556]: E0218 09:07:17.324514 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aeff40b439e6e8751290c8681fc2c889491f6e822c6c530d08c1d651305a72f6\": container with ID starting with aeff40b439e6e8751290c8681fc2c889491f6e822c6c530d08c1d651305a72f6 not found: ID does not exist" containerID="aeff40b439e6e8751290c8681fc2c889491f6e822c6c530d08c1d651305a72f6" Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.324553 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aeff40b439e6e8751290c8681fc2c889491f6e822c6c530d08c1d651305a72f6"} err="failed to get container status \"aeff40b439e6e8751290c8681fc2c889491f6e822c6c530d08c1d651305a72f6\": rpc error: code = NotFound desc = could not find container \"aeff40b439e6e8751290c8681fc2c889491f6e822c6c530d08c1d651305a72f6\": container with ID starting with aeff40b439e6e8751290c8681fc2c889491f6e822c6c530d08c1d651305a72f6 not found: ID does not exist" Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.324581 4556 scope.go:117] "RemoveContainer" containerID="b1e538edcfe7d69af3ccabf373f71ff4f59d9eed02244cc39e055c7904706da5" Feb 18 09:07:17 crc kubenswrapper[4556]: E0218 09:07:17.324885 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1e538edcfe7d69af3ccabf373f71ff4f59d9eed02244cc39e055c7904706da5\": container with ID starting with b1e538edcfe7d69af3ccabf373f71ff4f59d9eed02244cc39e055c7904706da5 not found: ID does not exist" containerID="b1e538edcfe7d69af3ccabf373f71ff4f59d9eed02244cc39e055c7904706da5" Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.324913 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1e538edcfe7d69af3ccabf373f71ff4f59d9eed02244cc39e055c7904706da5"} err="failed to get container status \"b1e538edcfe7d69af3ccabf373f71ff4f59d9eed02244cc39e055c7904706da5\": rpc error: code = NotFound desc = could not find container \"b1e538edcfe7d69af3ccabf373f71ff4f59d9eed02244cc39e055c7904706da5\": container with ID starting with b1e538edcfe7d69af3ccabf373f71ff4f59d9eed02244cc39e055c7904706da5 not found: ID does not exist" Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.324943 4556 scope.go:117] "RemoveContainer" containerID="071fb0569134209452182b6f81b2cb0454f0201b5b5056bb3fdb2e29f34c768d" Feb 18 09:07:17 crc kubenswrapper[4556]: E0218 09:07:17.325196 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"071fb0569134209452182b6f81b2cb0454f0201b5b5056bb3fdb2e29f34c768d\": container with ID starting with 071fb0569134209452182b6f81b2cb0454f0201b5b5056bb3fdb2e29f34c768d not found: ID does not exist" containerID="071fb0569134209452182b6f81b2cb0454f0201b5b5056bb3fdb2e29f34c768d" Feb 18 09:07:17 crc kubenswrapper[4556]: I0218 09:07:17.325213 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"071fb0569134209452182b6f81b2cb0454f0201b5b5056bb3fdb2e29f34c768d"} err="failed to get container status \"071fb0569134209452182b6f81b2cb0454f0201b5b5056bb3fdb2e29f34c768d\": rpc error: code = NotFound desc = could not find container \"071fb0569134209452182b6f81b2cb0454f0201b5b5056bb3fdb2e29f34c768d\": container with ID starting with 071fb0569134209452182b6f81b2cb0454f0201b5b5056bb3fdb2e29f34c768d not found: ID does not exist" Feb 18 09:07:18 crc kubenswrapper[4556]: I0218 09:07:18.087966 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-bfrcb"] Feb 18 09:07:18 crc kubenswrapper[4556]: E0218 09:07:18.088198 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eac84811-d54a-47c2-ae5e-77f3a531fc4f" containerName="extract-content" Feb 18 09:07:18 crc kubenswrapper[4556]: I0218 09:07:18.088213 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="eac84811-d54a-47c2-ae5e-77f3a531fc4f" containerName="extract-content" Feb 18 09:07:18 crc kubenswrapper[4556]: E0218 09:07:18.088224 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e23bd1bf-79ec-4cc0-95c4-16959eab3aa7" containerName="registry-server" Feb 18 09:07:18 crc kubenswrapper[4556]: I0218 09:07:18.088231 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="e23bd1bf-79ec-4cc0-95c4-16959eab3aa7" containerName="registry-server" Feb 18 09:07:18 crc kubenswrapper[4556]: E0218 09:07:18.088243 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eac84811-d54a-47c2-ae5e-77f3a531fc4f" containerName="extract-utilities" Feb 18 09:07:18 crc kubenswrapper[4556]: I0218 09:07:18.088249 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="eac84811-d54a-47c2-ae5e-77f3a531fc4f" containerName="extract-utilities" Feb 18 09:07:18 crc kubenswrapper[4556]: E0218 09:07:18.088257 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e23bd1bf-79ec-4cc0-95c4-16959eab3aa7" containerName="extract-utilities" Feb 18 09:07:18 crc kubenswrapper[4556]: I0218 09:07:18.088263 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="e23bd1bf-79ec-4cc0-95c4-16959eab3aa7" containerName="extract-utilities" Feb 18 09:07:18 crc kubenswrapper[4556]: E0218 09:07:18.088272 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fb132fe-af75-4b1a-828f-5b1dcd074fd6" containerName="marketplace-operator" Feb 18 09:07:18 crc kubenswrapper[4556]: I0218 09:07:18.088278 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fb132fe-af75-4b1a-828f-5b1dcd074fd6" containerName="marketplace-operator" Feb 18 09:07:18 crc kubenswrapper[4556]: E0218 09:07:18.088287 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90477a11-cc7c-48cf-a210-9d1d802360ce" containerName="extract-utilities" Feb 18 09:07:18 crc kubenswrapper[4556]: I0218 09:07:18.088293 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="90477a11-cc7c-48cf-a210-9d1d802360ce" containerName="extract-utilities" Feb 18 09:07:18 crc kubenswrapper[4556]: E0218 09:07:18.088300 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdd9dfd4-872b-42fc-960f-5d4c51e6e66e" containerName="extract-utilities" Feb 18 09:07:18 crc kubenswrapper[4556]: I0218 09:07:18.088307 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdd9dfd4-872b-42fc-960f-5d4c51e6e66e" containerName="extract-utilities" Feb 18 09:07:18 crc kubenswrapper[4556]: E0218 09:07:18.088316 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90477a11-cc7c-48cf-a210-9d1d802360ce" containerName="registry-server" Feb 18 09:07:18 crc kubenswrapper[4556]: I0218 09:07:18.088322 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="90477a11-cc7c-48cf-a210-9d1d802360ce" containerName="registry-server" Feb 18 09:07:18 crc kubenswrapper[4556]: E0218 09:07:18.088330 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdd9dfd4-872b-42fc-960f-5d4c51e6e66e" containerName="extract-content" Feb 18 09:07:18 crc kubenswrapper[4556]: I0218 09:07:18.088335 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdd9dfd4-872b-42fc-960f-5d4c51e6e66e" containerName="extract-content" Feb 18 09:07:18 crc kubenswrapper[4556]: E0218 09:07:18.088343 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90477a11-cc7c-48cf-a210-9d1d802360ce" containerName="extract-content" Feb 18 09:07:18 crc kubenswrapper[4556]: I0218 09:07:18.088348 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="90477a11-cc7c-48cf-a210-9d1d802360ce" containerName="extract-content" Feb 18 09:07:18 crc kubenswrapper[4556]: E0218 09:07:18.088358 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e23bd1bf-79ec-4cc0-95c4-16959eab3aa7" containerName="extract-content" Feb 18 09:07:18 crc kubenswrapper[4556]: I0218 09:07:18.088363 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="e23bd1bf-79ec-4cc0-95c4-16959eab3aa7" containerName="extract-content" Feb 18 09:07:18 crc kubenswrapper[4556]: E0218 09:07:18.088371 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eac84811-d54a-47c2-ae5e-77f3a531fc4f" containerName="registry-server" Feb 18 09:07:18 crc kubenswrapper[4556]: I0218 09:07:18.088376 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="eac84811-d54a-47c2-ae5e-77f3a531fc4f" containerName="registry-server" Feb 18 09:07:18 crc kubenswrapper[4556]: E0218 09:07:18.088382 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdd9dfd4-872b-42fc-960f-5d4c51e6e66e" containerName="registry-server" Feb 18 09:07:18 crc kubenswrapper[4556]: I0218 09:07:18.088387 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdd9dfd4-872b-42fc-960f-5d4c51e6e66e" containerName="registry-server" Feb 18 09:07:18 crc kubenswrapper[4556]: I0218 09:07:18.088468 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="e23bd1bf-79ec-4cc0-95c4-16959eab3aa7" containerName="registry-server" Feb 18 09:07:18 crc kubenswrapper[4556]: I0218 09:07:18.088480 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="eac84811-d54a-47c2-ae5e-77f3a531fc4f" containerName="registry-server" Feb 18 09:07:18 crc kubenswrapper[4556]: I0218 09:07:18.088487 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fb132fe-af75-4b1a-828f-5b1dcd074fd6" containerName="marketplace-operator" Feb 18 09:07:18 crc kubenswrapper[4556]: I0218 09:07:18.088493 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="fdd9dfd4-872b-42fc-960f-5d4c51e6e66e" containerName="registry-server" Feb 18 09:07:18 crc kubenswrapper[4556]: I0218 09:07:18.088500 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="90477a11-cc7c-48cf-a210-9d1d802360ce" containerName="registry-server" Feb 18 09:07:18 crc kubenswrapper[4556]: I0218 09:07:18.089146 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bfrcb" Feb 18 09:07:18 crc kubenswrapper[4556]: I0218 09:07:18.091484 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 18 09:07:18 crc kubenswrapper[4556]: I0218 09:07:18.098425 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bfrcb"] Feb 18 09:07:18 crc kubenswrapper[4556]: I0218 09:07:18.120647 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-vh264" Feb 18 09:07:18 crc kubenswrapper[4556]: I0218 09:07:18.124240 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-vh264" Feb 18 09:07:18 crc kubenswrapper[4556]: I0218 09:07:18.244617 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ba6b57c-a8a5-4692-955d-e851c1014fa4-catalog-content\") pod \"certified-operators-bfrcb\" (UID: \"6ba6b57c-a8a5-4692-955d-e851c1014fa4\") " pod="openshift-marketplace/certified-operators-bfrcb" Feb 18 09:07:18 crc kubenswrapper[4556]: I0218 09:07:18.244662 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmbb7\" (UniqueName: \"kubernetes.io/projected/6ba6b57c-a8a5-4692-955d-e851c1014fa4-kube-api-access-lmbb7\") pod \"certified-operators-bfrcb\" (UID: \"6ba6b57c-a8a5-4692-955d-e851c1014fa4\") " pod="openshift-marketplace/certified-operators-bfrcb" Feb 18 09:07:18 crc kubenswrapper[4556]: I0218 09:07:18.244732 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ba6b57c-a8a5-4692-955d-e851c1014fa4-utilities\") pod \"certified-operators-bfrcb\" (UID: \"6ba6b57c-a8a5-4692-955d-e851c1014fa4\") " pod="openshift-marketplace/certified-operators-bfrcb" Feb 18 09:07:18 crc kubenswrapper[4556]: I0218 09:07:18.346025 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ba6b57c-a8a5-4692-955d-e851c1014fa4-catalog-content\") pod \"certified-operators-bfrcb\" (UID: \"6ba6b57c-a8a5-4692-955d-e851c1014fa4\") " pod="openshift-marketplace/certified-operators-bfrcb" Feb 18 09:07:18 crc kubenswrapper[4556]: I0218 09:07:18.346075 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmbb7\" (UniqueName: \"kubernetes.io/projected/6ba6b57c-a8a5-4692-955d-e851c1014fa4-kube-api-access-lmbb7\") pod \"certified-operators-bfrcb\" (UID: \"6ba6b57c-a8a5-4692-955d-e851c1014fa4\") " pod="openshift-marketplace/certified-operators-bfrcb" Feb 18 09:07:18 crc kubenswrapper[4556]: I0218 09:07:18.346131 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ba6b57c-a8a5-4692-955d-e851c1014fa4-utilities\") pod \"certified-operators-bfrcb\" (UID: \"6ba6b57c-a8a5-4692-955d-e851c1014fa4\") " pod="openshift-marketplace/certified-operators-bfrcb" Feb 18 09:07:18 crc kubenswrapper[4556]: I0218 09:07:18.346502 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ba6b57c-a8a5-4692-955d-e851c1014fa4-catalog-content\") pod \"certified-operators-bfrcb\" (UID: \"6ba6b57c-a8a5-4692-955d-e851c1014fa4\") " pod="openshift-marketplace/certified-operators-bfrcb" Feb 18 09:07:18 crc kubenswrapper[4556]: I0218 09:07:18.346838 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ba6b57c-a8a5-4692-955d-e851c1014fa4-utilities\") pod \"certified-operators-bfrcb\" (UID: \"6ba6b57c-a8a5-4692-955d-e851c1014fa4\") " pod="openshift-marketplace/certified-operators-bfrcb" Feb 18 09:07:18 crc kubenswrapper[4556]: I0218 09:07:18.366834 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmbb7\" (UniqueName: \"kubernetes.io/projected/6ba6b57c-a8a5-4692-955d-e851c1014fa4-kube-api-access-lmbb7\") pod \"certified-operators-bfrcb\" (UID: \"6ba6b57c-a8a5-4692-955d-e851c1014fa4\") " pod="openshift-marketplace/certified-operators-bfrcb" Feb 18 09:07:18 crc kubenswrapper[4556]: I0218 09:07:18.409342 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bfrcb" Feb 18 09:07:18 crc kubenswrapper[4556]: I0218 09:07:18.766359 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bfrcb"] Feb 18 09:07:18 crc kubenswrapper[4556]: W0218 09:07:18.772471 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ba6b57c_a8a5_4692_955d_e851c1014fa4.slice/crio-22a36a33030cd94bb7d0684d85dc545879d4d9e5d5957b63a85222c7dabf214c WatchSource:0}: Error finding container 22a36a33030cd94bb7d0684d85dc545879d4d9e5d5957b63a85222c7dabf214c: Status 404 returned error can't find the container with id 22a36a33030cd94bb7d0684d85dc545879d4d9e5d5957b63a85222c7dabf214c Feb 18 09:07:19 crc kubenswrapper[4556]: I0218 09:07:19.092702 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-m8g6v"] Feb 18 09:07:19 crc kubenswrapper[4556]: I0218 09:07:19.093659 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m8g6v" Feb 18 09:07:19 crc kubenswrapper[4556]: I0218 09:07:19.096049 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 18 09:07:19 crc kubenswrapper[4556]: I0218 09:07:19.101080 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-m8g6v"] Feb 18 09:07:19 crc kubenswrapper[4556]: I0218 09:07:19.126930 4556 generic.go:334] "Generic (PLEG): container finished" podID="6ba6b57c-a8a5-4692-955d-e851c1014fa4" containerID="7aa607233a25a65d5c4e3927f45bba7660464f235544c3b28a6c6b05dce5d110" exitCode=0 Feb 18 09:07:19 crc kubenswrapper[4556]: I0218 09:07:19.126992 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bfrcb" event={"ID":"6ba6b57c-a8a5-4692-955d-e851c1014fa4","Type":"ContainerDied","Data":"7aa607233a25a65d5c4e3927f45bba7660464f235544c3b28a6c6b05dce5d110"} Feb 18 09:07:19 crc kubenswrapper[4556]: I0218 09:07:19.127042 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bfrcb" event={"ID":"6ba6b57c-a8a5-4692-955d-e851c1014fa4","Type":"ContainerStarted","Data":"22a36a33030cd94bb7d0684d85dc545879d4d9e5d5957b63a85222c7dabf214c"} Feb 18 09:07:19 crc kubenswrapper[4556]: I0218 09:07:19.260011 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4k528\" (UniqueName: \"kubernetes.io/projected/7e5017d0-70be-4177-9c9c-827f97ed470b-kube-api-access-4k528\") pod \"redhat-marketplace-m8g6v\" (UID: \"7e5017d0-70be-4177-9c9c-827f97ed470b\") " pod="openshift-marketplace/redhat-marketplace-m8g6v" Feb 18 09:07:19 crc kubenswrapper[4556]: I0218 09:07:19.260077 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e5017d0-70be-4177-9c9c-827f97ed470b-catalog-content\") pod \"redhat-marketplace-m8g6v\" (UID: \"7e5017d0-70be-4177-9c9c-827f97ed470b\") " pod="openshift-marketplace/redhat-marketplace-m8g6v" Feb 18 09:07:19 crc kubenswrapper[4556]: I0218 09:07:19.260149 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e5017d0-70be-4177-9c9c-827f97ed470b-utilities\") pod \"redhat-marketplace-m8g6v\" (UID: \"7e5017d0-70be-4177-9c9c-827f97ed470b\") " pod="openshift-marketplace/redhat-marketplace-m8g6v" Feb 18 09:07:19 crc kubenswrapper[4556]: I0218 09:07:19.361359 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4k528\" (UniqueName: \"kubernetes.io/projected/7e5017d0-70be-4177-9c9c-827f97ed470b-kube-api-access-4k528\") pod \"redhat-marketplace-m8g6v\" (UID: \"7e5017d0-70be-4177-9c9c-827f97ed470b\") " pod="openshift-marketplace/redhat-marketplace-m8g6v" Feb 18 09:07:19 crc kubenswrapper[4556]: I0218 09:07:19.361414 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e5017d0-70be-4177-9c9c-827f97ed470b-catalog-content\") pod \"redhat-marketplace-m8g6v\" (UID: \"7e5017d0-70be-4177-9c9c-827f97ed470b\") " pod="openshift-marketplace/redhat-marketplace-m8g6v" Feb 18 09:07:19 crc kubenswrapper[4556]: I0218 09:07:19.361441 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e5017d0-70be-4177-9c9c-827f97ed470b-utilities\") pod \"redhat-marketplace-m8g6v\" (UID: \"7e5017d0-70be-4177-9c9c-827f97ed470b\") " pod="openshift-marketplace/redhat-marketplace-m8g6v" Feb 18 09:07:19 crc kubenswrapper[4556]: I0218 09:07:19.361985 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e5017d0-70be-4177-9c9c-827f97ed470b-utilities\") pod \"redhat-marketplace-m8g6v\" (UID: \"7e5017d0-70be-4177-9c9c-827f97ed470b\") " pod="openshift-marketplace/redhat-marketplace-m8g6v" Feb 18 09:07:19 crc kubenswrapper[4556]: I0218 09:07:19.362181 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e5017d0-70be-4177-9c9c-827f97ed470b-catalog-content\") pod \"redhat-marketplace-m8g6v\" (UID: \"7e5017d0-70be-4177-9c9c-827f97ed470b\") " pod="openshift-marketplace/redhat-marketplace-m8g6v" Feb 18 09:07:19 crc kubenswrapper[4556]: I0218 09:07:19.377993 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4k528\" (UniqueName: \"kubernetes.io/projected/7e5017d0-70be-4177-9c9c-827f97ed470b-kube-api-access-4k528\") pod \"redhat-marketplace-m8g6v\" (UID: \"7e5017d0-70be-4177-9c9c-827f97ed470b\") " pod="openshift-marketplace/redhat-marketplace-m8g6v" Feb 18 09:07:19 crc kubenswrapper[4556]: I0218 09:07:19.413068 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m8g6v" Feb 18 09:07:19 crc kubenswrapper[4556]: I0218 09:07:19.777545 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-m8g6v"] Feb 18 09:07:20 crc kubenswrapper[4556]: I0218 09:07:20.133987 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bfrcb" event={"ID":"6ba6b57c-a8a5-4692-955d-e851c1014fa4","Type":"ContainerStarted","Data":"ddfac033853d31551b9a3e927dec877515c5f05c4bad1842b7cc27d4c69b5526"} Feb 18 09:07:20 crc kubenswrapper[4556]: I0218 09:07:20.135441 4556 generic.go:334] "Generic (PLEG): container finished" podID="7e5017d0-70be-4177-9c9c-827f97ed470b" containerID="966400146bac531347356467e21b701f26fce82abefd1f8164ea03f241d3e271" exitCode=0 Feb 18 09:07:20 crc kubenswrapper[4556]: I0218 09:07:20.135521 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m8g6v" event={"ID":"7e5017d0-70be-4177-9c9c-827f97ed470b","Type":"ContainerDied","Data":"966400146bac531347356467e21b701f26fce82abefd1f8164ea03f241d3e271"} Feb 18 09:07:20 crc kubenswrapper[4556]: I0218 09:07:20.135569 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m8g6v" event={"ID":"7e5017d0-70be-4177-9c9c-827f97ed470b","Type":"ContainerStarted","Data":"577b5fd06f8590070b994c8ffb8fedd8be3e5dfd38b0c265e4080d8bdc14b19f"} Feb 18 09:07:20 crc kubenswrapper[4556]: I0218 09:07:20.489910 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-w6td5"] Feb 18 09:07:20 crc kubenswrapper[4556]: I0218 09:07:20.490958 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w6td5" Feb 18 09:07:20 crc kubenswrapper[4556]: I0218 09:07:20.492979 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 18 09:07:20 crc kubenswrapper[4556]: I0218 09:07:20.499577 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-w6td5"] Feb 18 09:07:20 crc kubenswrapper[4556]: I0218 09:07:20.678787 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e82c70e2-ba17-4919-a836-a6e252af6380-utilities\") pod \"redhat-operators-w6td5\" (UID: \"e82c70e2-ba17-4919-a836-a6e252af6380\") " pod="openshift-marketplace/redhat-operators-w6td5" Feb 18 09:07:20 crc kubenswrapper[4556]: I0218 09:07:20.678870 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e82c70e2-ba17-4919-a836-a6e252af6380-catalog-content\") pod \"redhat-operators-w6td5\" (UID: \"e82c70e2-ba17-4919-a836-a6e252af6380\") " pod="openshift-marketplace/redhat-operators-w6td5" Feb 18 09:07:20 crc kubenswrapper[4556]: I0218 09:07:20.678897 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fcz9p\" (UniqueName: \"kubernetes.io/projected/e82c70e2-ba17-4919-a836-a6e252af6380-kube-api-access-fcz9p\") pod \"redhat-operators-w6td5\" (UID: \"e82c70e2-ba17-4919-a836-a6e252af6380\") " pod="openshift-marketplace/redhat-operators-w6td5" Feb 18 09:07:20 crc kubenswrapper[4556]: I0218 09:07:20.779303 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e82c70e2-ba17-4919-a836-a6e252af6380-utilities\") pod \"redhat-operators-w6td5\" (UID: \"e82c70e2-ba17-4919-a836-a6e252af6380\") " pod="openshift-marketplace/redhat-operators-w6td5" Feb 18 09:07:20 crc kubenswrapper[4556]: I0218 09:07:20.779368 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e82c70e2-ba17-4919-a836-a6e252af6380-catalog-content\") pod \"redhat-operators-w6td5\" (UID: \"e82c70e2-ba17-4919-a836-a6e252af6380\") " pod="openshift-marketplace/redhat-operators-w6td5" Feb 18 09:07:20 crc kubenswrapper[4556]: I0218 09:07:20.779387 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fcz9p\" (UniqueName: \"kubernetes.io/projected/e82c70e2-ba17-4919-a836-a6e252af6380-kube-api-access-fcz9p\") pod \"redhat-operators-w6td5\" (UID: \"e82c70e2-ba17-4919-a836-a6e252af6380\") " pod="openshift-marketplace/redhat-operators-w6td5" Feb 18 09:07:20 crc kubenswrapper[4556]: I0218 09:07:20.779729 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e82c70e2-ba17-4919-a836-a6e252af6380-utilities\") pod \"redhat-operators-w6td5\" (UID: \"e82c70e2-ba17-4919-a836-a6e252af6380\") " pod="openshift-marketplace/redhat-operators-w6td5" Feb 18 09:07:20 crc kubenswrapper[4556]: I0218 09:07:20.779974 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e82c70e2-ba17-4919-a836-a6e252af6380-catalog-content\") pod \"redhat-operators-w6td5\" (UID: \"e82c70e2-ba17-4919-a836-a6e252af6380\") " pod="openshift-marketplace/redhat-operators-w6td5" Feb 18 09:07:20 crc kubenswrapper[4556]: I0218 09:07:20.794620 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fcz9p\" (UniqueName: \"kubernetes.io/projected/e82c70e2-ba17-4919-a836-a6e252af6380-kube-api-access-fcz9p\") pod \"redhat-operators-w6td5\" (UID: \"e82c70e2-ba17-4919-a836-a6e252af6380\") " pod="openshift-marketplace/redhat-operators-w6td5" Feb 18 09:07:20 crc kubenswrapper[4556]: I0218 09:07:20.805938 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w6td5" Feb 18 09:07:20 crc kubenswrapper[4556]: I0218 09:07:20.959931 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-w6td5"] Feb 18 09:07:20 crc kubenswrapper[4556]: W0218 09:07:20.967693 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode82c70e2_ba17_4919_a836_a6e252af6380.slice/crio-67dc79861dc806bb86953422bd8346e70d64ad9278f52e2bc16e898e8a4f039e WatchSource:0}: Error finding container 67dc79861dc806bb86953422bd8346e70d64ad9278f52e2bc16e898e8a4f039e: Status 404 returned error can't find the container with id 67dc79861dc806bb86953422bd8346e70d64ad9278f52e2bc16e898e8a4f039e Feb 18 09:07:21 crc kubenswrapper[4556]: I0218 09:07:21.142879 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w6td5" event={"ID":"e82c70e2-ba17-4919-a836-a6e252af6380","Type":"ContainerStarted","Data":"67dc79861dc806bb86953422bd8346e70d64ad9278f52e2bc16e898e8a4f039e"} Feb 18 09:07:21 crc kubenswrapper[4556]: I0218 09:07:21.145084 4556 generic.go:334] "Generic (PLEG): container finished" podID="7e5017d0-70be-4177-9c9c-827f97ed470b" containerID="513c544eaa541edecc311b1330e2324fcbe3be148b1ebf4682909a25d9997e2a" exitCode=0 Feb 18 09:07:21 crc kubenswrapper[4556]: I0218 09:07:21.145162 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m8g6v" event={"ID":"7e5017d0-70be-4177-9c9c-827f97ed470b","Type":"ContainerDied","Data":"513c544eaa541edecc311b1330e2324fcbe3be148b1ebf4682909a25d9997e2a"} Feb 18 09:07:21 crc kubenswrapper[4556]: I0218 09:07:21.146970 4556 generic.go:334] "Generic (PLEG): container finished" podID="6ba6b57c-a8a5-4692-955d-e851c1014fa4" containerID="ddfac033853d31551b9a3e927dec877515c5f05c4bad1842b7cc27d4c69b5526" exitCode=0 Feb 18 09:07:21 crc kubenswrapper[4556]: I0218 09:07:21.147001 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bfrcb" event={"ID":"6ba6b57c-a8a5-4692-955d-e851c1014fa4","Type":"ContainerDied","Data":"ddfac033853d31551b9a3e927dec877515c5f05c4bad1842b7cc27d4c69b5526"} Feb 18 09:07:21 crc kubenswrapper[4556]: I0218 09:07:21.489999 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ldrzh"] Feb 18 09:07:21 crc kubenswrapper[4556]: I0218 09:07:21.491358 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ldrzh" Feb 18 09:07:21 crc kubenswrapper[4556]: I0218 09:07:21.493616 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 18 09:07:21 crc kubenswrapper[4556]: I0218 09:07:21.495634 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ldrzh"] Feb 18 09:07:21 crc kubenswrapper[4556]: I0218 09:07:21.591792 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8dns\" (UniqueName: \"kubernetes.io/projected/745b3a46-b0b9-443b-95d2-6bfd635d1243-kube-api-access-z8dns\") pod \"community-operators-ldrzh\" (UID: \"745b3a46-b0b9-443b-95d2-6bfd635d1243\") " pod="openshift-marketplace/community-operators-ldrzh" Feb 18 09:07:21 crc kubenswrapper[4556]: I0218 09:07:21.591863 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/745b3a46-b0b9-443b-95d2-6bfd635d1243-catalog-content\") pod \"community-operators-ldrzh\" (UID: \"745b3a46-b0b9-443b-95d2-6bfd635d1243\") " pod="openshift-marketplace/community-operators-ldrzh" Feb 18 09:07:21 crc kubenswrapper[4556]: I0218 09:07:21.592049 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/745b3a46-b0b9-443b-95d2-6bfd635d1243-utilities\") pod \"community-operators-ldrzh\" (UID: \"745b3a46-b0b9-443b-95d2-6bfd635d1243\") " pod="openshift-marketplace/community-operators-ldrzh" Feb 18 09:07:21 crc kubenswrapper[4556]: I0218 09:07:21.692825 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/745b3a46-b0b9-443b-95d2-6bfd635d1243-utilities\") pod \"community-operators-ldrzh\" (UID: \"745b3a46-b0b9-443b-95d2-6bfd635d1243\") " pod="openshift-marketplace/community-operators-ldrzh" Feb 18 09:07:21 crc kubenswrapper[4556]: I0218 09:07:21.692927 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8dns\" (UniqueName: \"kubernetes.io/projected/745b3a46-b0b9-443b-95d2-6bfd635d1243-kube-api-access-z8dns\") pod \"community-operators-ldrzh\" (UID: \"745b3a46-b0b9-443b-95d2-6bfd635d1243\") " pod="openshift-marketplace/community-operators-ldrzh" Feb 18 09:07:21 crc kubenswrapper[4556]: I0218 09:07:21.692950 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/745b3a46-b0b9-443b-95d2-6bfd635d1243-catalog-content\") pod \"community-operators-ldrzh\" (UID: \"745b3a46-b0b9-443b-95d2-6bfd635d1243\") " pod="openshift-marketplace/community-operators-ldrzh" Feb 18 09:07:21 crc kubenswrapper[4556]: I0218 09:07:21.693494 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/745b3a46-b0b9-443b-95d2-6bfd635d1243-utilities\") pod \"community-operators-ldrzh\" (UID: \"745b3a46-b0b9-443b-95d2-6bfd635d1243\") " pod="openshift-marketplace/community-operators-ldrzh" Feb 18 09:07:21 crc kubenswrapper[4556]: I0218 09:07:21.693580 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/745b3a46-b0b9-443b-95d2-6bfd635d1243-catalog-content\") pod \"community-operators-ldrzh\" (UID: \"745b3a46-b0b9-443b-95d2-6bfd635d1243\") " pod="openshift-marketplace/community-operators-ldrzh" Feb 18 09:07:21 crc kubenswrapper[4556]: I0218 09:07:21.708746 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8dns\" (UniqueName: \"kubernetes.io/projected/745b3a46-b0b9-443b-95d2-6bfd635d1243-kube-api-access-z8dns\") pod \"community-operators-ldrzh\" (UID: \"745b3a46-b0b9-443b-95d2-6bfd635d1243\") " pod="openshift-marketplace/community-operators-ldrzh" Feb 18 09:07:21 crc kubenswrapper[4556]: I0218 09:07:21.807107 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ldrzh" Feb 18 09:07:21 crc kubenswrapper[4556]: I0218 09:07:21.981655 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ldrzh"] Feb 18 09:07:21 crc kubenswrapper[4556]: W0218 09:07:21.991161 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod745b3a46_b0b9_443b_95d2_6bfd635d1243.slice/crio-ab803441c7d7915024cec406e4aa5368cc5f003b22af71878580efe43a5335bb WatchSource:0}: Error finding container ab803441c7d7915024cec406e4aa5368cc5f003b22af71878580efe43a5335bb: Status 404 returned error can't find the container with id ab803441c7d7915024cec406e4aa5368cc5f003b22af71878580efe43a5335bb Feb 18 09:07:22 crc kubenswrapper[4556]: I0218 09:07:22.154911 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bfrcb" event={"ID":"6ba6b57c-a8a5-4692-955d-e851c1014fa4","Type":"ContainerStarted","Data":"157218c86b7301d736cd4cecafa1c68e9353e4fc88d5f88c7c48b2ba6266a638"} Feb 18 09:07:22 crc kubenswrapper[4556]: I0218 09:07:22.157105 4556 generic.go:334] "Generic (PLEG): container finished" podID="e82c70e2-ba17-4919-a836-a6e252af6380" containerID="f68201182bee17548b0edff08cf1aaa22c2dd45e49fb84036e04faeb7a19741e" exitCode=0 Feb 18 09:07:22 crc kubenswrapper[4556]: I0218 09:07:22.157177 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w6td5" event={"ID":"e82c70e2-ba17-4919-a836-a6e252af6380","Type":"ContainerDied","Data":"f68201182bee17548b0edff08cf1aaa22c2dd45e49fb84036e04faeb7a19741e"} Feb 18 09:07:22 crc kubenswrapper[4556]: I0218 09:07:22.161520 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m8g6v" event={"ID":"7e5017d0-70be-4177-9c9c-827f97ed470b","Type":"ContainerStarted","Data":"446a5bd7d372c6e7894c32842eb645f561ed37b3d0e17ae6801eba7595f91d91"} Feb 18 09:07:22 crc kubenswrapper[4556]: I0218 09:07:22.163974 4556 generic.go:334] "Generic (PLEG): container finished" podID="745b3a46-b0b9-443b-95d2-6bfd635d1243" containerID="a0d4022c8da57efcf45358e0d2be4e5421e3e04ab2bac89e46b9b54d48d64142" exitCode=0 Feb 18 09:07:22 crc kubenswrapper[4556]: I0218 09:07:22.164011 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ldrzh" event={"ID":"745b3a46-b0b9-443b-95d2-6bfd635d1243","Type":"ContainerDied","Data":"a0d4022c8da57efcf45358e0d2be4e5421e3e04ab2bac89e46b9b54d48d64142"} Feb 18 09:07:22 crc kubenswrapper[4556]: I0218 09:07:22.164034 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ldrzh" event={"ID":"745b3a46-b0b9-443b-95d2-6bfd635d1243","Type":"ContainerStarted","Data":"ab803441c7d7915024cec406e4aa5368cc5f003b22af71878580efe43a5335bb"} Feb 18 09:07:22 crc kubenswrapper[4556]: I0218 09:07:22.171968 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-bfrcb" podStartSLOduration=1.657618574 podStartE2EDuration="4.171948532s" podCreationTimestamp="2026-02-18 09:07:18 +0000 UTC" firstStartedPulling="2026-02-18 09:07:19.129414482 +0000 UTC m=+196.146375462" lastFinishedPulling="2026-02-18 09:07:21.643744441 +0000 UTC m=+198.660705420" observedRunningTime="2026-02-18 09:07:22.168318785 +0000 UTC m=+199.185279765" watchObservedRunningTime="2026-02-18 09:07:22.171948532 +0000 UTC m=+199.188909512" Feb 18 09:07:22 crc kubenswrapper[4556]: I0218 09:07:22.215364 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-m8g6v" podStartSLOduration=1.715953421 podStartE2EDuration="3.215346672s" podCreationTimestamp="2026-02-18 09:07:19 +0000 UTC" firstStartedPulling="2026-02-18 09:07:20.136775919 +0000 UTC m=+197.153736899" lastFinishedPulling="2026-02-18 09:07:21.63616917 +0000 UTC m=+198.653130150" observedRunningTime="2026-02-18 09:07:22.214332498 +0000 UTC m=+199.231293498" watchObservedRunningTime="2026-02-18 09:07:22.215346672 +0000 UTC m=+199.232307651" Feb 18 09:07:23 crc kubenswrapper[4556]: I0218 09:07:23.171339 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ldrzh" event={"ID":"745b3a46-b0b9-443b-95d2-6bfd635d1243","Type":"ContainerStarted","Data":"aa30531576612ea9e8b91b859c2251b132c838b41ca367343cad4125d7c3015f"} Feb 18 09:07:23 crc kubenswrapper[4556]: I0218 09:07:23.175039 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w6td5" event={"ID":"e82c70e2-ba17-4919-a836-a6e252af6380","Type":"ContainerStarted","Data":"4b16c8cdb035a134a56828081c02d397baeb957c64cae5a8b0f72ea470361789"} Feb 18 09:07:24 crc kubenswrapper[4556]: I0218 09:07:24.194372 4556 generic.go:334] "Generic (PLEG): container finished" podID="745b3a46-b0b9-443b-95d2-6bfd635d1243" containerID="aa30531576612ea9e8b91b859c2251b132c838b41ca367343cad4125d7c3015f" exitCode=0 Feb 18 09:07:24 crc kubenswrapper[4556]: I0218 09:07:24.194461 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ldrzh" event={"ID":"745b3a46-b0b9-443b-95d2-6bfd635d1243","Type":"ContainerDied","Data":"aa30531576612ea9e8b91b859c2251b132c838b41ca367343cad4125d7c3015f"} Feb 18 09:07:24 crc kubenswrapper[4556]: I0218 09:07:24.198224 4556 generic.go:334] "Generic (PLEG): container finished" podID="e82c70e2-ba17-4919-a836-a6e252af6380" containerID="4b16c8cdb035a134a56828081c02d397baeb957c64cae5a8b0f72ea470361789" exitCode=0 Feb 18 09:07:24 crc kubenswrapper[4556]: I0218 09:07:24.198270 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w6td5" event={"ID":"e82c70e2-ba17-4919-a836-a6e252af6380","Type":"ContainerDied","Data":"4b16c8cdb035a134a56828081c02d397baeb957c64cae5a8b0f72ea470361789"} Feb 18 09:07:25 crc kubenswrapper[4556]: I0218 09:07:25.207021 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ldrzh" event={"ID":"745b3a46-b0b9-443b-95d2-6bfd635d1243","Type":"ContainerStarted","Data":"c84e0440ee4ebba67ef2a0411e760b933ba5f5df84297ad696488df6aa1691cd"} Feb 18 09:07:25 crc kubenswrapper[4556]: I0218 09:07:25.210243 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w6td5" event={"ID":"e82c70e2-ba17-4919-a836-a6e252af6380","Type":"ContainerStarted","Data":"5726d7291fd59dbd3199c342c0b23c4f26ebc8eeeacb2e4ef608beb12da5b88d"} Feb 18 09:07:25 crc kubenswrapper[4556]: I0218 09:07:25.222479 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ldrzh" podStartSLOduration=1.6071277990000001 podStartE2EDuration="4.222459846s" podCreationTimestamp="2026-02-18 09:07:21 +0000 UTC" firstStartedPulling="2026-02-18 09:07:22.165005656 +0000 UTC m=+199.181966637" lastFinishedPulling="2026-02-18 09:07:24.780337704 +0000 UTC m=+201.797298684" observedRunningTime="2026-02-18 09:07:25.220938866 +0000 UTC m=+202.237899847" watchObservedRunningTime="2026-02-18 09:07:25.222459846 +0000 UTC m=+202.239420826" Feb 18 09:07:25 crc kubenswrapper[4556]: I0218 09:07:25.237062 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-w6td5" podStartSLOduration=2.745518536 podStartE2EDuration="5.237045167s" podCreationTimestamp="2026-02-18 09:07:20 +0000 UTC" firstStartedPulling="2026-02-18 09:07:22.158454038 +0000 UTC m=+199.175415018" lastFinishedPulling="2026-02-18 09:07:24.64998067 +0000 UTC m=+201.666941649" observedRunningTime="2026-02-18 09:07:25.236008221 +0000 UTC m=+202.252969221" watchObservedRunningTime="2026-02-18 09:07:25.237045167 +0000 UTC m=+202.254006147" Feb 18 09:07:28 crc kubenswrapper[4556]: I0218 09:07:28.409835 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-bfrcb" Feb 18 09:07:28 crc kubenswrapper[4556]: I0218 09:07:28.409886 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-bfrcb" Feb 18 09:07:28 crc kubenswrapper[4556]: I0218 09:07:28.443299 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-bfrcb" Feb 18 09:07:29 crc kubenswrapper[4556]: I0218 09:07:29.256895 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-bfrcb" Feb 18 09:07:29 crc kubenswrapper[4556]: I0218 09:07:29.413550 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-m8g6v" Feb 18 09:07:29 crc kubenswrapper[4556]: I0218 09:07:29.413587 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-m8g6v" Feb 18 09:07:29 crc kubenswrapper[4556]: I0218 09:07:29.440123 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-m8g6v" Feb 18 09:07:30 crc kubenswrapper[4556]: I0218 09:07:30.258111 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-m8g6v" Feb 18 09:07:30 crc kubenswrapper[4556]: I0218 09:07:30.806943 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-w6td5" Feb 18 09:07:30 crc kubenswrapper[4556]: I0218 09:07:30.806998 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-w6td5" Feb 18 09:07:30 crc kubenswrapper[4556]: I0218 09:07:30.832633 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-w6td5" Feb 18 09:07:31 crc kubenswrapper[4556]: I0218 09:07:31.274839 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-w6td5" Feb 18 09:07:31 crc kubenswrapper[4556]: I0218 09:07:31.727421 4556 patch_prober.go:28] interesting pod/machine-config-daemon-f76hs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 09:07:31 crc kubenswrapper[4556]: I0218 09:07:31.727627 4556 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 09:07:31 crc kubenswrapper[4556]: I0218 09:07:31.727673 4556 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" Feb 18 09:07:31 crc kubenswrapper[4556]: I0218 09:07:31.728280 4556 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6d41b85076e695bbc51a41fce200e0e47c099d88b4d825b4d133d33ed96237e2"} pod="openshift-machine-config-operator/machine-config-daemon-f76hs" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 18 09:07:31 crc kubenswrapper[4556]: I0218 09:07:31.728338 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerName="machine-config-daemon" containerID="cri-o://6d41b85076e695bbc51a41fce200e0e47c099d88b4d825b4d133d33ed96237e2" gracePeriod=600 Feb 18 09:07:31 crc kubenswrapper[4556]: I0218 09:07:31.808087 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ldrzh" Feb 18 09:07:31 crc kubenswrapper[4556]: I0218 09:07:31.808137 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ldrzh" Feb 18 09:07:31 crc kubenswrapper[4556]: I0218 09:07:31.834723 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ldrzh" Feb 18 09:07:32 crc kubenswrapper[4556]: I0218 09:07:32.243601 4556 generic.go:334] "Generic (PLEG): container finished" podID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerID="6d41b85076e695bbc51a41fce200e0e47c099d88b4d825b4d133d33ed96237e2" exitCode=0 Feb 18 09:07:32 crc kubenswrapper[4556]: I0218 09:07:32.243647 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" event={"ID":"8dac7f27-d3d1-4778-9e54-f273035a1d37","Type":"ContainerDied","Data":"6d41b85076e695bbc51a41fce200e0e47c099d88b4d825b4d133d33ed96237e2"} Feb 18 09:07:32 crc kubenswrapper[4556]: I0218 09:07:32.243690 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" event={"ID":"8dac7f27-d3d1-4778-9e54-f273035a1d37","Type":"ContainerStarted","Data":"8b54607784a82541a6af0430808ac511ecebb51bf04484e5e4e70f755a71305e"} Feb 18 09:07:32 crc kubenswrapper[4556]: I0218 09:07:32.270915 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ldrzh" Feb 18 09:07:36 crc kubenswrapper[4556]: I0218 09:07:36.340621 4556 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 18 09:07:36 crc kubenswrapper[4556]: I0218 09:07:36.341274 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://c3a99dc228d1c4cb8aeab2140ec5f37358b65684739f833fbde45822782430a8" gracePeriod=15 Feb 18 09:07:36 crc kubenswrapper[4556]: I0218 09:07:36.341302 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://b42d13c1767c29d22e52537ff45561a0fbf8f2069b2c3612e903e8c127e12485" gracePeriod=15 Feb 18 09:07:36 crc kubenswrapper[4556]: I0218 09:07:36.341405 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://974e17ae7082631df2d1bcee60c8a5f142bae2659e9d200d537a4d650f401b4b" gracePeriod=15 Feb 18 09:07:36 crc kubenswrapper[4556]: I0218 09:07:36.341446 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://14f794063db9d33b82b79acc7134b5a2a3435e09d7cffe49d49251d1f7d20402" gracePeriod=15 Feb 18 09:07:36 crc kubenswrapper[4556]: I0218 09:07:36.341464 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://5eb3d3728da247903bece56bc2481547317e386e9e97b6c69926a3b55dc89b8e" gracePeriod=15 Feb 18 09:07:36 crc kubenswrapper[4556]: I0218 09:07:36.342077 4556 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 18 09:07:36 crc kubenswrapper[4556]: E0218 09:07:36.342391 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 18 09:07:36 crc kubenswrapper[4556]: I0218 09:07:36.342424 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 18 09:07:36 crc kubenswrapper[4556]: E0218 09:07:36.342433 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 18 09:07:36 crc kubenswrapper[4556]: I0218 09:07:36.342439 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 18 09:07:36 crc kubenswrapper[4556]: E0218 09:07:36.342446 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 18 09:07:36 crc kubenswrapper[4556]: I0218 09:07:36.342452 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 18 09:07:36 crc kubenswrapper[4556]: E0218 09:07:36.343436 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 18 09:07:36 crc kubenswrapper[4556]: I0218 09:07:36.343449 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 18 09:07:36 crc kubenswrapper[4556]: E0218 09:07:36.343459 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 18 09:07:36 crc kubenswrapper[4556]: I0218 09:07:36.343464 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 18 09:07:36 crc kubenswrapper[4556]: E0218 09:07:36.343473 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 18 09:07:36 crc kubenswrapper[4556]: I0218 09:07:36.343480 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 18 09:07:36 crc kubenswrapper[4556]: E0218 09:07:36.343492 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Feb 18 09:07:36 crc kubenswrapper[4556]: I0218 09:07:36.343497 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Feb 18 09:07:36 crc kubenswrapper[4556]: I0218 09:07:36.343645 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 18 09:07:36 crc kubenswrapper[4556]: I0218 09:07:36.343660 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 18 09:07:36 crc kubenswrapper[4556]: I0218 09:07:36.343667 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 18 09:07:36 crc kubenswrapper[4556]: I0218 09:07:36.343675 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 18 09:07:36 crc kubenswrapper[4556]: I0218 09:07:36.343681 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 18 09:07:36 crc kubenswrapper[4556]: I0218 09:07:36.343687 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 18 09:07:36 crc kubenswrapper[4556]: I0218 09:07:36.344916 4556 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 18 09:07:36 crc kubenswrapper[4556]: I0218 09:07:36.345426 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 09:07:36 crc kubenswrapper[4556]: I0218 09:07:36.348630 4556 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="f4b27818a5e8e43d0dc095d08835c792" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" Feb 18 09:07:36 crc kubenswrapper[4556]: I0218 09:07:36.391560 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 09:07:36 crc kubenswrapper[4556]: I0218 09:07:36.391613 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 09:07:36 crc kubenswrapper[4556]: I0218 09:07:36.391655 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 09:07:36 crc kubenswrapper[4556]: I0218 09:07:36.391674 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 09:07:36 crc kubenswrapper[4556]: I0218 09:07:36.391752 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 09:07:36 crc kubenswrapper[4556]: I0218 09:07:36.391813 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 09:07:36 crc kubenswrapper[4556]: I0218 09:07:36.391971 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 09:07:36 crc kubenswrapper[4556]: I0218 09:07:36.392018 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 09:07:36 crc kubenswrapper[4556]: I0218 09:07:36.493294 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 09:07:36 crc kubenswrapper[4556]: I0218 09:07:36.493348 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 09:07:36 crc kubenswrapper[4556]: I0218 09:07:36.493382 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 09:07:36 crc kubenswrapper[4556]: I0218 09:07:36.493400 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 09:07:36 crc kubenswrapper[4556]: I0218 09:07:36.493419 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 09:07:36 crc kubenswrapper[4556]: I0218 09:07:36.493439 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 09:07:36 crc kubenswrapper[4556]: I0218 09:07:36.493466 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 09:07:36 crc kubenswrapper[4556]: I0218 09:07:36.493510 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 09:07:36 crc kubenswrapper[4556]: I0218 09:07:36.493514 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 09:07:36 crc kubenswrapper[4556]: I0218 09:07:36.493524 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 09:07:36 crc kubenswrapper[4556]: I0218 09:07:36.493436 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 09:07:36 crc kubenswrapper[4556]: I0218 09:07:36.493479 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 09:07:36 crc kubenswrapper[4556]: I0218 09:07:36.493595 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 09:07:36 crc kubenswrapper[4556]: I0218 09:07:36.493539 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 09:07:36 crc kubenswrapper[4556]: I0218 09:07:36.493642 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 09:07:36 crc kubenswrapper[4556]: I0218 09:07:36.493668 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 09:07:36 crc kubenswrapper[4556]: E0218 09:07:36.617905 4556 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.217:6443: connect: connection refused" Feb 18 09:07:36 crc kubenswrapper[4556]: E0218 09:07:36.618313 4556 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.217:6443: connect: connection refused" Feb 18 09:07:36 crc kubenswrapper[4556]: E0218 09:07:36.618542 4556 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.217:6443: connect: connection refused" Feb 18 09:07:36 crc kubenswrapper[4556]: E0218 09:07:36.618767 4556 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.217:6443: connect: connection refused" Feb 18 09:07:36 crc kubenswrapper[4556]: E0218 09:07:36.619021 4556 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.217:6443: connect: connection refused" Feb 18 09:07:36 crc kubenswrapper[4556]: I0218 09:07:36.619050 4556 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Feb 18 09:07:36 crc kubenswrapper[4556]: E0218 09:07:36.619282 4556 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.217:6443: connect: connection refused" interval="200ms" Feb 18 09:07:36 crc kubenswrapper[4556]: E0218 09:07:36.820406 4556 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.217:6443: connect: connection refused" interval="400ms" Feb 18 09:07:37 crc kubenswrapper[4556]: E0218 09:07:37.221625 4556 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.217:6443: connect: connection refused" interval="800ms" Feb 18 09:07:37 crc kubenswrapper[4556]: I0218 09:07:37.265520 4556 generic.go:334] "Generic (PLEG): container finished" podID="12967c01-f63e-466e-b28f-4828af636833" containerID="606ca898b3cfd55868817996f930826fa3c99883e206b760ac8f6965d3593a2e" exitCode=0 Feb 18 09:07:37 crc kubenswrapper[4556]: I0218 09:07:37.265593 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"12967c01-f63e-466e-b28f-4828af636833","Type":"ContainerDied","Data":"606ca898b3cfd55868817996f930826fa3c99883e206b760ac8f6965d3593a2e"} Feb 18 09:07:37 crc kubenswrapper[4556]: I0218 09:07:37.266122 4556 status_manager.go:851] "Failed to get status for pod" podUID="12967c01-f63e-466e-b28f-4828af636833" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 192.168.26.217:6443: connect: connection refused" Feb 18 09:07:37 crc kubenswrapper[4556]: I0218 09:07:37.267804 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 18 09:07:37 crc kubenswrapper[4556]: I0218 09:07:37.269038 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 18 09:07:37 crc kubenswrapper[4556]: I0218 09:07:37.269665 4556 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="b42d13c1767c29d22e52537ff45561a0fbf8f2069b2c3612e903e8c127e12485" exitCode=0 Feb 18 09:07:37 crc kubenswrapper[4556]: I0218 09:07:37.269682 4556 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="974e17ae7082631df2d1bcee60c8a5f142bae2659e9d200d537a4d650f401b4b" exitCode=0 Feb 18 09:07:37 crc kubenswrapper[4556]: I0218 09:07:37.269691 4556 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="14f794063db9d33b82b79acc7134b5a2a3435e09d7cffe49d49251d1f7d20402" exitCode=0 Feb 18 09:07:37 crc kubenswrapper[4556]: I0218 09:07:37.269697 4556 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="5eb3d3728da247903bece56bc2481547317e386e9e97b6c69926a3b55dc89b8e" exitCode=2 Feb 18 09:07:37 crc kubenswrapper[4556]: I0218 09:07:37.269725 4556 scope.go:117] "RemoveContainer" containerID="072e865e6a50c03519ce22958ac4fcbb911fc140c330b6d7ffe5c9d7dd381e96" Feb 18 09:07:38 crc kubenswrapper[4556]: E0218 09:07:38.023176 4556 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.217:6443: connect: connection refused" interval="1.6s" Feb 18 09:07:38 crc kubenswrapper[4556]: I0218 09:07:38.277441 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 18 09:07:38 crc kubenswrapper[4556]: I0218 09:07:38.484041 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 18 09:07:38 crc kubenswrapper[4556]: I0218 09:07:38.485501 4556 status_manager.go:851] "Failed to get status for pod" podUID="12967c01-f63e-466e-b28f-4828af636833" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 192.168.26.217:6443: connect: connection refused" Feb 18 09:07:38 crc kubenswrapper[4556]: I0218 09:07:38.512828 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/12967c01-f63e-466e-b28f-4828af636833-kubelet-dir\") pod \"12967c01-f63e-466e-b28f-4828af636833\" (UID: \"12967c01-f63e-466e-b28f-4828af636833\") " Feb 18 09:07:38 crc kubenswrapper[4556]: I0218 09:07:38.512863 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/12967c01-f63e-466e-b28f-4828af636833-var-lock\") pod \"12967c01-f63e-466e-b28f-4828af636833\" (UID: \"12967c01-f63e-466e-b28f-4828af636833\") " Feb 18 09:07:38 crc kubenswrapper[4556]: I0218 09:07:38.512890 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/12967c01-f63e-466e-b28f-4828af636833-kube-api-access\") pod \"12967c01-f63e-466e-b28f-4828af636833\" (UID: \"12967c01-f63e-466e-b28f-4828af636833\") " Feb 18 09:07:38 crc kubenswrapper[4556]: I0218 09:07:38.512956 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/12967c01-f63e-466e-b28f-4828af636833-var-lock" (OuterVolumeSpecName: "var-lock") pod "12967c01-f63e-466e-b28f-4828af636833" (UID: "12967c01-f63e-466e-b28f-4828af636833"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 09:07:38 crc kubenswrapper[4556]: I0218 09:07:38.512952 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/12967c01-f63e-466e-b28f-4828af636833-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "12967c01-f63e-466e-b28f-4828af636833" (UID: "12967c01-f63e-466e-b28f-4828af636833"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 09:07:38 crc kubenswrapper[4556]: I0218 09:07:38.513278 4556 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/12967c01-f63e-466e-b28f-4828af636833-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 18 09:07:38 crc kubenswrapper[4556]: I0218 09:07:38.513294 4556 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/12967c01-f63e-466e-b28f-4828af636833-var-lock\") on node \"crc\" DevicePath \"\"" Feb 18 09:07:38 crc kubenswrapper[4556]: I0218 09:07:38.517479 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12967c01-f63e-466e-b28f-4828af636833-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "12967c01-f63e-466e-b28f-4828af636833" (UID: "12967c01-f63e-466e-b28f-4828af636833"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:07:38 crc kubenswrapper[4556]: I0218 09:07:38.614185 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/12967c01-f63e-466e-b28f-4828af636833-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 18 09:07:38 crc kubenswrapper[4556]: I0218 09:07:38.615926 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 18 09:07:38 crc kubenswrapper[4556]: I0218 09:07:38.616539 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 09:07:38 crc kubenswrapper[4556]: I0218 09:07:38.617021 4556 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 192.168.26.217:6443: connect: connection refused" Feb 18 09:07:38 crc kubenswrapper[4556]: I0218 09:07:38.617270 4556 status_manager.go:851] "Failed to get status for pod" podUID="12967c01-f63e-466e-b28f-4828af636833" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 192.168.26.217:6443: connect: connection refused" Feb 18 09:07:38 crc kubenswrapper[4556]: I0218 09:07:38.714588 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 18 09:07:38 crc kubenswrapper[4556]: I0218 09:07:38.714631 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 18 09:07:38 crc kubenswrapper[4556]: I0218 09:07:38.714651 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 18 09:07:38 crc kubenswrapper[4556]: I0218 09:07:38.714848 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 09:07:38 crc kubenswrapper[4556]: I0218 09:07:38.714877 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 09:07:38 crc kubenswrapper[4556]: I0218 09:07:38.714893 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 09:07:38 crc kubenswrapper[4556]: I0218 09:07:38.815213 4556 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 18 09:07:38 crc kubenswrapper[4556]: I0218 09:07:38.815239 4556 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Feb 18 09:07:38 crc kubenswrapper[4556]: I0218 09:07:38.815249 4556 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Feb 18 09:07:39 crc kubenswrapper[4556]: I0218 09:07:39.284558 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 18 09:07:39 crc kubenswrapper[4556]: I0218 09:07:39.288049 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Feb 18 09:07:39 crc kubenswrapper[4556]: I0218 09:07:39.288274 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 18 09:07:39 crc kubenswrapper[4556]: I0218 09:07:39.288803 4556 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="c3a99dc228d1c4cb8aeab2140ec5f37358b65684739f833fbde45822782430a8" exitCode=0 Feb 18 09:07:39 crc kubenswrapper[4556]: I0218 09:07:39.288873 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 09:07:39 crc kubenswrapper[4556]: I0218 09:07:39.289283 4556 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 192.168.26.217:6443: connect: connection refused" Feb 18 09:07:39 crc kubenswrapper[4556]: I0218 09:07:39.289517 4556 status_manager.go:851] "Failed to get status for pod" podUID="12967c01-f63e-466e-b28f-4828af636833" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 192.168.26.217:6443: connect: connection refused" Feb 18 09:07:39 crc kubenswrapper[4556]: I0218 09:07:39.290220 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"12967c01-f63e-466e-b28f-4828af636833","Type":"ContainerDied","Data":"7f7c44c9cd7c42a3a328479b76c8c9732a99646ccd2b103743ae485bf0c02092"} Feb 18 09:07:39 crc kubenswrapper[4556]: I0218 09:07:39.290252 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f7c44c9cd7c42a3a328479b76c8c9732a99646ccd2b103743ae485bf0c02092" Feb 18 09:07:39 crc kubenswrapper[4556]: I0218 09:07:39.290280 4556 scope.go:117] "RemoveContainer" containerID="b42d13c1767c29d22e52537ff45561a0fbf8f2069b2c3612e903e8c127e12485" Feb 18 09:07:39 crc kubenswrapper[4556]: I0218 09:07:39.303868 4556 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 192.168.26.217:6443: connect: connection refused" Feb 18 09:07:39 crc kubenswrapper[4556]: I0218 09:07:39.304041 4556 status_manager.go:851] "Failed to get status for pod" podUID="12967c01-f63e-466e-b28f-4828af636833" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 192.168.26.217:6443: connect: connection refused" Feb 18 09:07:39 crc kubenswrapper[4556]: I0218 09:07:39.304921 4556 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 192.168.26.217:6443: connect: connection refused" Feb 18 09:07:39 crc kubenswrapper[4556]: I0218 09:07:39.305181 4556 status_manager.go:851] "Failed to get status for pod" podUID="12967c01-f63e-466e-b28f-4828af636833" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 192.168.26.217:6443: connect: connection refused" Feb 18 09:07:39 crc kubenswrapper[4556]: I0218 09:07:39.308003 4556 scope.go:117] "RemoveContainer" containerID="974e17ae7082631df2d1bcee60c8a5f142bae2659e9d200d537a4d650f401b4b" Feb 18 09:07:39 crc kubenswrapper[4556]: I0218 09:07:39.317529 4556 scope.go:117] "RemoveContainer" containerID="14f794063db9d33b82b79acc7134b5a2a3435e09d7cffe49d49251d1f7d20402" Feb 18 09:07:39 crc kubenswrapper[4556]: I0218 09:07:39.326631 4556 scope.go:117] "RemoveContainer" containerID="5eb3d3728da247903bece56bc2481547317e386e9e97b6c69926a3b55dc89b8e" Feb 18 09:07:39 crc kubenswrapper[4556]: I0218 09:07:39.336720 4556 scope.go:117] "RemoveContainer" containerID="c3a99dc228d1c4cb8aeab2140ec5f37358b65684739f833fbde45822782430a8" Feb 18 09:07:39 crc kubenswrapper[4556]: I0218 09:07:39.347180 4556 scope.go:117] "RemoveContainer" containerID="6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e" Feb 18 09:07:39 crc kubenswrapper[4556]: I0218 09:07:39.366969 4556 scope.go:117] "RemoveContainer" containerID="b42d13c1767c29d22e52537ff45561a0fbf8f2069b2c3612e903e8c127e12485" Feb 18 09:07:39 crc kubenswrapper[4556]: E0218 09:07:39.367263 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b42d13c1767c29d22e52537ff45561a0fbf8f2069b2c3612e903e8c127e12485\": container with ID starting with b42d13c1767c29d22e52537ff45561a0fbf8f2069b2c3612e903e8c127e12485 not found: ID does not exist" containerID="b42d13c1767c29d22e52537ff45561a0fbf8f2069b2c3612e903e8c127e12485" Feb 18 09:07:39 crc kubenswrapper[4556]: I0218 09:07:39.367300 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b42d13c1767c29d22e52537ff45561a0fbf8f2069b2c3612e903e8c127e12485"} err="failed to get container status \"b42d13c1767c29d22e52537ff45561a0fbf8f2069b2c3612e903e8c127e12485\": rpc error: code = NotFound desc = could not find container \"b42d13c1767c29d22e52537ff45561a0fbf8f2069b2c3612e903e8c127e12485\": container with ID starting with b42d13c1767c29d22e52537ff45561a0fbf8f2069b2c3612e903e8c127e12485 not found: ID does not exist" Feb 18 09:07:39 crc kubenswrapper[4556]: I0218 09:07:39.367323 4556 scope.go:117] "RemoveContainer" containerID="974e17ae7082631df2d1bcee60c8a5f142bae2659e9d200d537a4d650f401b4b" Feb 18 09:07:39 crc kubenswrapper[4556]: E0218 09:07:39.367547 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"974e17ae7082631df2d1bcee60c8a5f142bae2659e9d200d537a4d650f401b4b\": container with ID starting with 974e17ae7082631df2d1bcee60c8a5f142bae2659e9d200d537a4d650f401b4b not found: ID does not exist" containerID="974e17ae7082631df2d1bcee60c8a5f142bae2659e9d200d537a4d650f401b4b" Feb 18 09:07:39 crc kubenswrapper[4556]: I0218 09:07:39.367578 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"974e17ae7082631df2d1bcee60c8a5f142bae2659e9d200d537a4d650f401b4b"} err="failed to get container status \"974e17ae7082631df2d1bcee60c8a5f142bae2659e9d200d537a4d650f401b4b\": rpc error: code = NotFound desc = could not find container \"974e17ae7082631df2d1bcee60c8a5f142bae2659e9d200d537a4d650f401b4b\": container with ID starting with 974e17ae7082631df2d1bcee60c8a5f142bae2659e9d200d537a4d650f401b4b not found: ID does not exist" Feb 18 09:07:39 crc kubenswrapper[4556]: I0218 09:07:39.367598 4556 scope.go:117] "RemoveContainer" containerID="14f794063db9d33b82b79acc7134b5a2a3435e09d7cffe49d49251d1f7d20402" Feb 18 09:07:39 crc kubenswrapper[4556]: E0218 09:07:39.369276 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14f794063db9d33b82b79acc7134b5a2a3435e09d7cffe49d49251d1f7d20402\": container with ID starting with 14f794063db9d33b82b79acc7134b5a2a3435e09d7cffe49d49251d1f7d20402 not found: ID does not exist" containerID="14f794063db9d33b82b79acc7134b5a2a3435e09d7cffe49d49251d1f7d20402" Feb 18 09:07:39 crc kubenswrapper[4556]: I0218 09:07:39.369307 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14f794063db9d33b82b79acc7134b5a2a3435e09d7cffe49d49251d1f7d20402"} err="failed to get container status \"14f794063db9d33b82b79acc7134b5a2a3435e09d7cffe49d49251d1f7d20402\": rpc error: code = NotFound desc = could not find container \"14f794063db9d33b82b79acc7134b5a2a3435e09d7cffe49d49251d1f7d20402\": container with ID starting with 14f794063db9d33b82b79acc7134b5a2a3435e09d7cffe49d49251d1f7d20402 not found: ID does not exist" Feb 18 09:07:39 crc kubenswrapper[4556]: I0218 09:07:39.369467 4556 scope.go:117] "RemoveContainer" containerID="5eb3d3728da247903bece56bc2481547317e386e9e97b6c69926a3b55dc89b8e" Feb 18 09:07:39 crc kubenswrapper[4556]: E0218 09:07:39.369776 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5eb3d3728da247903bece56bc2481547317e386e9e97b6c69926a3b55dc89b8e\": container with ID starting with 5eb3d3728da247903bece56bc2481547317e386e9e97b6c69926a3b55dc89b8e not found: ID does not exist" containerID="5eb3d3728da247903bece56bc2481547317e386e9e97b6c69926a3b55dc89b8e" Feb 18 09:07:39 crc kubenswrapper[4556]: I0218 09:07:39.369802 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5eb3d3728da247903bece56bc2481547317e386e9e97b6c69926a3b55dc89b8e"} err="failed to get container status \"5eb3d3728da247903bece56bc2481547317e386e9e97b6c69926a3b55dc89b8e\": rpc error: code = NotFound desc = could not find container \"5eb3d3728da247903bece56bc2481547317e386e9e97b6c69926a3b55dc89b8e\": container with ID starting with 5eb3d3728da247903bece56bc2481547317e386e9e97b6c69926a3b55dc89b8e not found: ID does not exist" Feb 18 09:07:39 crc kubenswrapper[4556]: I0218 09:07:39.369818 4556 scope.go:117] "RemoveContainer" containerID="c3a99dc228d1c4cb8aeab2140ec5f37358b65684739f833fbde45822782430a8" Feb 18 09:07:39 crc kubenswrapper[4556]: E0218 09:07:39.370039 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3a99dc228d1c4cb8aeab2140ec5f37358b65684739f833fbde45822782430a8\": container with ID starting with c3a99dc228d1c4cb8aeab2140ec5f37358b65684739f833fbde45822782430a8 not found: ID does not exist" containerID="c3a99dc228d1c4cb8aeab2140ec5f37358b65684739f833fbde45822782430a8" Feb 18 09:07:39 crc kubenswrapper[4556]: I0218 09:07:39.370064 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3a99dc228d1c4cb8aeab2140ec5f37358b65684739f833fbde45822782430a8"} err="failed to get container status \"c3a99dc228d1c4cb8aeab2140ec5f37358b65684739f833fbde45822782430a8\": rpc error: code = NotFound desc = could not find container \"c3a99dc228d1c4cb8aeab2140ec5f37358b65684739f833fbde45822782430a8\": container with ID starting with c3a99dc228d1c4cb8aeab2140ec5f37358b65684739f833fbde45822782430a8 not found: ID does not exist" Feb 18 09:07:39 crc kubenswrapper[4556]: I0218 09:07:39.370082 4556 scope.go:117] "RemoveContainer" containerID="6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e" Feb 18 09:07:39 crc kubenswrapper[4556]: E0218 09:07:39.370280 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\": container with ID starting with 6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e not found: ID does not exist" containerID="6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e" Feb 18 09:07:39 crc kubenswrapper[4556]: I0218 09:07:39.370305 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e"} err="failed to get container status \"6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\": rpc error: code = NotFound desc = could not find container \"6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e\": container with ID starting with 6761e2cf5b64775c8500379ef1b0e466421cad7bf20241b62886d8345f96122e not found: ID does not exist" Feb 18 09:07:39 crc kubenswrapper[4556]: E0218 09:07:39.623884 4556 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.217:6443: connect: connection refused" interval="3.2s" Feb 18 09:07:41 crc kubenswrapper[4556]: E0218 09:07:41.361956 4556 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 192.168.26.217:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 09:07:41 crc kubenswrapper[4556]: I0218 09:07:41.362662 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 09:07:41 crc kubenswrapper[4556]: W0218 09:07:41.378999 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-0178b16043e279ee18fbdd838083b1c841e598a6209358a7b4e72a74490b1c0c WatchSource:0}: Error finding container 0178b16043e279ee18fbdd838083b1c841e598a6209358a7b4e72a74490b1c0c: Status 404 returned error can't find the container with id 0178b16043e279ee18fbdd838083b1c841e598a6209358a7b4e72a74490b1c0c Feb 18 09:07:41 crc kubenswrapper[4556]: E0218 09:07:41.381009 4556 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 192.168.26.217:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.18954c104fb4e9e7 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-18 09:07:41.380602343 +0000 UTC m=+218.397563323,LastTimestamp:2026-02-18 09:07:41.380602343 +0000 UTC m=+218.397563323,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 18 09:07:42 crc kubenswrapper[4556]: I0218 09:07:42.302876 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"00d23fad26f6025c9d80e4438c939e4c205d81d1edb047fc31f1c19450a6bf26"} Feb 18 09:07:42 crc kubenswrapper[4556]: I0218 09:07:42.303513 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"0178b16043e279ee18fbdd838083b1c841e598a6209358a7b4e72a74490b1c0c"} Feb 18 09:07:42 crc kubenswrapper[4556]: I0218 09:07:42.303994 4556 status_manager.go:851] "Failed to get status for pod" podUID="12967c01-f63e-466e-b28f-4828af636833" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 192.168.26.217:6443: connect: connection refused" Feb 18 09:07:42 crc kubenswrapper[4556]: E0218 09:07:42.304036 4556 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 192.168.26.217:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 09:07:42 crc kubenswrapper[4556]: E0218 09:07:42.495921 4556 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 192.168.26.217:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.18954c104fb4e9e7 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-18 09:07:41.380602343 +0000 UTC m=+218.397563323,LastTimestamp:2026-02-18 09:07:41.380602343 +0000 UTC m=+218.397563323,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 18 09:07:42 crc kubenswrapper[4556]: E0218 09:07:42.824728 4556 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.217:6443: connect: connection refused" interval="6.4s" Feb 18 09:07:43 crc kubenswrapper[4556]: I0218 09:07:43.283919 4556 status_manager.go:851] "Failed to get status for pod" podUID="12967c01-f63e-466e-b28f-4828af636833" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 192.168.26.217:6443: connect: connection refused" Feb 18 09:07:47 crc kubenswrapper[4556]: I0218 09:07:47.281808 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 09:07:47 crc kubenswrapper[4556]: I0218 09:07:47.282819 4556 status_manager.go:851] "Failed to get status for pod" podUID="12967c01-f63e-466e-b28f-4828af636833" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 192.168.26.217:6443: connect: connection refused" Feb 18 09:07:47 crc kubenswrapper[4556]: I0218 09:07:47.294028 4556 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="302eb215-a4c2-4324-836d-f37debfd620f" Feb 18 09:07:47 crc kubenswrapper[4556]: I0218 09:07:47.294062 4556 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="302eb215-a4c2-4324-836d-f37debfd620f" Feb 18 09:07:47 crc kubenswrapper[4556]: E0218 09:07:47.294358 4556 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 192.168.26.217:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 09:07:47 crc kubenswrapper[4556]: I0218 09:07:47.294877 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 09:07:47 crc kubenswrapper[4556]: W0218 09:07:47.313982 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-fd14a8303d8be937e783443a5e9bf5527b3fc24eaa3893c954d13ab2bd1d2419 WatchSource:0}: Error finding container fd14a8303d8be937e783443a5e9bf5527b3fc24eaa3893c954d13ab2bd1d2419: Status 404 returned error can't find the container with id fd14a8303d8be937e783443a5e9bf5527b3fc24eaa3893c954d13ab2bd1d2419 Feb 18 09:07:47 crc kubenswrapper[4556]: I0218 09:07:47.328868 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"fd14a8303d8be937e783443a5e9bf5527b3fc24eaa3893c954d13ab2bd1d2419"} Feb 18 09:07:48 crc kubenswrapper[4556]: I0218 09:07:48.334196 4556 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="b090ecba6a0e5dcb21067cd3e026e86b67dc624b3c0353a661d983f31a026601" exitCode=0 Feb 18 09:07:48 crc kubenswrapper[4556]: I0218 09:07:48.334459 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"b090ecba6a0e5dcb21067cd3e026e86b67dc624b3c0353a661d983f31a026601"} Feb 18 09:07:48 crc kubenswrapper[4556]: I0218 09:07:48.334648 4556 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="302eb215-a4c2-4324-836d-f37debfd620f" Feb 18 09:07:48 crc kubenswrapper[4556]: I0218 09:07:48.334661 4556 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="302eb215-a4c2-4324-836d-f37debfd620f" Feb 18 09:07:48 crc kubenswrapper[4556]: E0218 09:07:48.334904 4556 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 192.168.26.217:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 09:07:48 crc kubenswrapper[4556]: I0218 09:07:48.335181 4556 status_manager.go:851] "Failed to get status for pod" podUID="12967c01-f63e-466e-b28f-4828af636833" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 192.168.26.217:6443: connect: connection refused" Feb 18 09:07:49 crc kubenswrapper[4556]: I0218 09:07:49.340346 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"e2e5a2a992fcbdf18db7be158c282f9c9b37f5c5a5bd51c7d6a90681be8cddf4"} Feb 18 09:07:49 crc kubenswrapper[4556]: I0218 09:07:49.340608 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 09:07:49 crc kubenswrapper[4556]: I0218 09:07:49.340606 4556 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="302eb215-a4c2-4324-836d-f37debfd620f" Feb 18 09:07:49 crc kubenswrapper[4556]: I0218 09:07:49.340631 4556 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="302eb215-a4c2-4324-836d-f37debfd620f" Feb 18 09:07:49 crc kubenswrapper[4556]: I0218 09:07:49.340619 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"c6181c5e37248c75acb32107076703e4e9cf21a786d9b4a2b9016b55d8a91c4a"} Feb 18 09:07:49 crc kubenswrapper[4556]: I0218 09:07:49.341340 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"86a8cebf265d21f2a577891ed89be2762ba90df0d4d6193cd75252e5a0826d1b"} Feb 18 09:07:49 crc kubenswrapper[4556]: I0218 09:07:49.341361 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"d2c8be5526c35e431b08397b1b76fa45337ee49bf9745837942377fbd963442e"} Feb 18 09:07:49 crc kubenswrapper[4556]: I0218 09:07:49.341371 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"e9def0886bd372a663f777112b6b2f28ebef28a15061015fe721cdddebeed82b"} Feb 18 09:07:50 crc kubenswrapper[4556]: I0218 09:07:50.346773 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Feb 18 09:07:50 crc kubenswrapper[4556]: I0218 09:07:50.346985 4556 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="ac09c421b3101887bfb83107dbe26f2b5856d9b32a9cea92f6233fb6ae218cc1" exitCode=1 Feb 18 09:07:50 crc kubenswrapper[4556]: I0218 09:07:50.347012 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"ac09c421b3101887bfb83107dbe26f2b5856d9b32a9cea92f6233fb6ae218cc1"} Feb 18 09:07:50 crc kubenswrapper[4556]: I0218 09:07:50.347464 4556 scope.go:117] "RemoveContainer" containerID="ac09c421b3101887bfb83107dbe26f2b5856d9b32a9cea92f6233fb6ae218cc1" Feb 18 09:07:51 crc kubenswrapper[4556]: I0218 09:07:51.358519 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Feb 18 09:07:51 crc kubenswrapper[4556]: I0218 09:07:51.358580 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"1584e7c278ac85aa588550b48023c0521f6ab9956180aa015cb386d7f33c93d8"} Feb 18 09:07:52 crc kubenswrapper[4556]: I0218 09:07:52.296038 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 09:07:52 crc kubenswrapper[4556]: I0218 09:07:52.296073 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 09:07:52 crc kubenswrapper[4556]: I0218 09:07:52.300266 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 09:07:53 crc kubenswrapper[4556]: I0218 09:07:53.211124 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 18 09:07:54 crc kubenswrapper[4556]: I0218 09:07:54.478669 4556 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 09:07:54 crc kubenswrapper[4556]: I0218 09:07:54.519062 4556 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="2bcb60c6-7363-415c-a873-2c2dd00f88b7" Feb 18 09:07:55 crc kubenswrapper[4556]: I0218 09:07:55.376881 4556 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="302eb215-a4c2-4324-836d-f37debfd620f" Feb 18 09:07:55 crc kubenswrapper[4556]: I0218 09:07:55.376909 4556 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="302eb215-a4c2-4324-836d-f37debfd620f" Feb 18 09:07:55 crc kubenswrapper[4556]: I0218 09:07:55.381775 4556 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="2bcb60c6-7363-415c-a873-2c2dd00f88b7" Feb 18 09:07:55 crc kubenswrapper[4556]: I0218 09:07:55.382724 4556 status_manager.go:308] "Container readiness changed before pod has synced" pod="openshift-kube-apiserver/kube-apiserver-crc" containerID="cri-o://e9def0886bd372a663f777112b6b2f28ebef28a15061015fe721cdddebeed82b" Feb 18 09:07:55 crc kubenswrapper[4556]: I0218 09:07:55.382747 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 09:07:56 crc kubenswrapper[4556]: I0218 09:07:56.380953 4556 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="302eb215-a4c2-4324-836d-f37debfd620f" Feb 18 09:07:56 crc kubenswrapper[4556]: I0218 09:07:56.380982 4556 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="302eb215-a4c2-4324-836d-f37debfd620f" Feb 18 09:07:56 crc kubenswrapper[4556]: I0218 09:07:56.382592 4556 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="2bcb60c6-7363-415c-a873-2c2dd00f88b7" Feb 18 09:07:59 crc kubenswrapper[4556]: I0218 09:07:59.318889 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 18 09:07:59 crc kubenswrapper[4556]: I0218 09:07:59.322083 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 18 09:08:00 crc kubenswrapper[4556]: I0218 09:08:00.938255 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Feb 18 09:08:00 crc kubenswrapper[4556]: I0218 09:08:00.984872 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Feb 18 09:08:02 crc kubenswrapper[4556]: I0218 09:08:02.080088 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Feb 18 09:08:02 crc kubenswrapper[4556]: I0218 09:08:02.438808 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Feb 18 09:08:03 crc kubenswrapper[4556]: I0218 09:08:03.214037 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 18 09:08:03 crc kubenswrapper[4556]: I0218 09:08:03.358646 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Feb 18 09:08:03 crc kubenswrapper[4556]: I0218 09:08:03.566557 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Feb 18 09:08:03 crc kubenswrapper[4556]: I0218 09:08:03.997350 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 18 09:08:04 crc kubenswrapper[4556]: I0218 09:08:04.260662 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Feb 18 09:08:04 crc kubenswrapper[4556]: I0218 09:08:04.333567 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Feb 18 09:08:04 crc kubenswrapper[4556]: I0218 09:08:04.426796 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 18 09:08:04 crc kubenswrapper[4556]: I0218 09:08:04.733374 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Feb 18 09:08:05 crc kubenswrapper[4556]: I0218 09:08:05.186905 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Feb 18 09:08:06 crc kubenswrapper[4556]: I0218 09:08:06.255995 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Feb 18 09:08:06 crc kubenswrapper[4556]: I0218 09:08:06.299903 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Feb 18 09:08:06 crc kubenswrapper[4556]: I0218 09:08:06.399478 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Feb 18 09:08:06 crc kubenswrapper[4556]: I0218 09:08:06.451007 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Feb 18 09:08:06 crc kubenswrapper[4556]: I0218 09:08:06.566550 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Feb 18 09:08:06 crc kubenswrapper[4556]: I0218 09:08:06.740998 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Feb 18 09:08:07 crc kubenswrapper[4556]: I0218 09:08:07.243528 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Feb 18 09:08:07 crc kubenswrapper[4556]: I0218 09:08:07.491222 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Feb 18 09:08:07 crc kubenswrapper[4556]: I0218 09:08:07.564538 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Feb 18 09:08:07 crc kubenswrapper[4556]: I0218 09:08:07.660985 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Feb 18 09:08:07 crc kubenswrapper[4556]: I0218 09:08:07.688377 4556 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Feb 18 09:08:08 crc kubenswrapper[4556]: I0218 09:08:08.285660 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 18 09:08:08 crc kubenswrapper[4556]: I0218 09:08:08.313263 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Feb 18 09:08:08 crc kubenswrapper[4556]: I0218 09:08:08.318134 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Feb 18 09:08:08 crc kubenswrapper[4556]: I0218 09:08:08.330870 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Feb 18 09:08:08 crc kubenswrapper[4556]: I0218 09:08:08.925286 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Feb 18 09:08:08 crc kubenswrapper[4556]: I0218 09:08:08.937723 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Feb 18 09:08:08 crc kubenswrapper[4556]: I0218 09:08:08.981368 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Feb 18 09:08:09 crc kubenswrapper[4556]: I0218 09:08:09.028985 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Feb 18 09:08:09 crc kubenswrapper[4556]: I0218 09:08:09.054826 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Feb 18 09:08:09 crc kubenswrapper[4556]: I0218 09:08:09.075910 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Feb 18 09:08:09 crc kubenswrapper[4556]: I0218 09:08:09.128902 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Feb 18 09:08:09 crc kubenswrapper[4556]: I0218 09:08:09.204622 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Feb 18 09:08:09 crc kubenswrapper[4556]: I0218 09:08:09.354482 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Feb 18 09:08:09 crc kubenswrapper[4556]: I0218 09:08:09.358666 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Feb 18 09:08:09 crc kubenswrapper[4556]: I0218 09:08:09.392020 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Feb 18 09:08:09 crc kubenswrapper[4556]: I0218 09:08:09.629192 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Feb 18 09:08:09 crc kubenswrapper[4556]: I0218 09:08:09.772236 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Feb 18 09:08:09 crc kubenswrapper[4556]: I0218 09:08:09.779724 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Feb 18 09:08:09 crc kubenswrapper[4556]: I0218 09:08:09.859202 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Feb 18 09:08:09 crc kubenswrapper[4556]: I0218 09:08:09.896474 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Feb 18 09:08:10 crc kubenswrapper[4556]: I0218 09:08:10.000355 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Feb 18 09:08:10 crc kubenswrapper[4556]: I0218 09:08:10.208303 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Feb 18 09:08:10 crc kubenswrapper[4556]: I0218 09:08:10.227147 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Feb 18 09:08:10 crc kubenswrapper[4556]: I0218 09:08:10.239867 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Feb 18 09:08:10 crc kubenswrapper[4556]: I0218 09:08:10.334246 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Feb 18 09:08:10 crc kubenswrapper[4556]: I0218 09:08:10.338514 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Feb 18 09:08:10 crc kubenswrapper[4556]: I0218 09:08:10.393664 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Feb 18 09:08:10 crc kubenswrapper[4556]: I0218 09:08:10.460280 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Feb 18 09:08:10 crc kubenswrapper[4556]: I0218 09:08:10.482222 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Feb 18 09:08:10 crc kubenswrapper[4556]: I0218 09:08:10.559227 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Feb 18 09:08:10 crc kubenswrapper[4556]: I0218 09:08:10.575756 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Feb 18 09:08:10 crc kubenswrapper[4556]: I0218 09:08:10.675779 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 18 09:08:10 crc kubenswrapper[4556]: I0218 09:08:10.811850 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Feb 18 09:08:10 crc kubenswrapper[4556]: I0218 09:08:10.922192 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Feb 18 09:08:10 crc kubenswrapper[4556]: I0218 09:08:10.933197 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Feb 18 09:08:10 crc kubenswrapper[4556]: I0218 09:08:10.948788 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Feb 18 09:08:11 crc kubenswrapper[4556]: I0218 09:08:11.115593 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Feb 18 09:08:11 crc kubenswrapper[4556]: I0218 09:08:11.208384 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Feb 18 09:08:11 crc kubenswrapper[4556]: I0218 09:08:11.212257 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Feb 18 09:08:11 crc kubenswrapper[4556]: I0218 09:08:11.401577 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Feb 18 09:08:11 crc kubenswrapper[4556]: I0218 09:08:11.588944 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 18 09:08:11 crc kubenswrapper[4556]: I0218 09:08:11.633070 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Feb 18 09:08:11 crc kubenswrapper[4556]: I0218 09:08:11.719858 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Feb 18 09:08:11 crc kubenswrapper[4556]: I0218 09:08:11.863267 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Feb 18 09:08:11 crc kubenswrapper[4556]: I0218 09:08:11.891401 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Feb 18 09:08:12 crc kubenswrapper[4556]: I0218 09:08:12.007218 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Feb 18 09:08:12 crc kubenswrapper[4556]: I0218 09:08:12.088009 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 18 09:08:12 crc kubenswrapper[4556]: I0218 09:08:12.180255 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Feb 18 09:08:12 crc kubenswrapper[4556]: I0218 09:08:12.341785 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Feb 18 09:08:12 crc kubenswrapper[4556]: I0218 09:08:12.357775 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Feb 18 09:08:12 crc kubenswrapper[4556]: I0218 09:08:12.377696 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Feb 18 09:08:12 crc kubenswrapper[4556]: I0218 09:08:12.378811 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Feb 18 09:08:12 crc kubenswrapper[4556]: I0218 09:08:12.486187 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Feb 18 09:08:12 crc kubenswrapper[4556]: I0218 09:08:12.539651 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Feb 18 09:08:12 crc kubenswrapper[4556]: I0218 09:08:12.550944 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Feb 18 09:08:12 crc kubenswrapper[4556]: I0218 09:08:12.583829 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Feb 18 09:08:12 crc kubenswrapper[4556]: I0218 09:08:12.760253 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Feb 18 09:08:12 crc kubenswrapper[4556]: I0218 09:08:12.771986 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Feb 18 09:08:12 crc kubenswrapper[4556]: I0218 09:08:12.779403 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Feb 18 09:08:12 crc kubenswrapper[4556]: I0218 09:08:12.829885 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Feb 18 09:08:12 crc kubenswrapper[4556]: I0218 09:08:12.857904 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Feb 18 09:08:12 crc kubenswrapper[4556]: I0218 09:08:12.880114 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Feb 18 09:08:12 crc kubenswrapper[4556]: I0218 09:08:12.907220 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Feb 18 09:08:13 crc kubenswrapper[4556]: I0218 09:08:13.024692 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Feb 18 09:08:13 crc kubenswrapper[4556]: I0218 09:08:13.129121 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Feb 18 09:08:13 crc kubenswrapper[4556]: I0218 09:08:13.160876 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Feb 18 09:08:13 crc kubenswrapper[4556]: I0218 09:08:13.172345 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Feb 18 09:08:13 crc kubenswrapper[4556]: I0218 09:08:13.184356 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Feb 18 09:08:13 crc kubenswrapper[4556]: I0218 09:08:13.201299 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Feb 18 09:08:13 crc kubenswrapper[4556]: I0218 09:08:13.260907 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 18 09:08:13 crc kubenswrapper[4556]: I0218 09:08:13.301394 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Feb 18 09:08:13 crc kubenswrapper[4556]: I0218 09:08:13.337562 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 18 09:08:13 crc kubenswrapper[4556]: I0218 09:08:13.381009 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Feb 18 09:08:13 crc kubenswrapper[4556]: I0218 09:08:13.410330 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Feb 18 09:08:13 crc kubenswrapper[4556]: I0218 09:08:13.479425 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Feb 18 09:08:13 crc kubenswrapper[4556]: I0218 09:08:13.536365 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Feb 18 09:08:13 crc kubenswrapper[4556]: I0218 09:08:13.561549 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Feb 18 09:08:13 crc kubenswrapper[4556]: I0218 09:08:13.670318 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Feb 18 09:08:13 crc kubenswrapper[4556]: I0218 09:08:13.677547 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Feb 18 09:08:13 crc kubenswrapper[4556]: I0218 09:08:13.709866 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Feb 18 09:08:13 crc kubenswrapper[4556]: I0218 09:08:13.742804 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 18 09:08:13 crc kubenswrapper[4556]: I0218 09:08:13.776796 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Feb 18 09:08:13 crc kubenswrapper[4556]: I0218 09:08:13.836725 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 18 09:08:13 crc kubenswrapper[4556]: I0218 09:08:13.869648 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Feb 18 09:08:13 crc kubenswrapper[4556]: I0218 09:08:13.877344 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Feb 18 09:08:13 crc kubenswrapper[4556]: I0218 09:08:13.879748 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Feb 18 09:08:13 crc kubenswrapper[4556]: I0218 09:08:13.954980 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Feb 18 09:08:14 crc kubenswrapper[4556]: I0218 09:08:14.013839 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 18 09:08:14 crc kubenswrapper[4556]: I0218 09:08:14.058614 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Feb 18 09:08:14 crc kubenswrapper[4556]: I0218 09:08:14.066385 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Feb 18 09:08:14 crc kubenswrapper[4556]: I0218 09:08:14.278310 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Feb 18 09:08:14 crc kubenswrapper[4556]: I0218 09:08:14.336980 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Feb 18 09:08:14 crc kubenswrapper[4556]: I0218 09:08:14.344859 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Feb 18 09:08:14 crc kubenswrapper[4556]: I0218 09:08:14.394861 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Feb 18 09:08:14 crc kubenswrapper[4556]: I0218 09:08:14.445487 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Feb 18 09:08:14 crc kubenswrapper[4556]: I0218 09:08:14.499254 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Feb 18 09:08:14 crc kubenswrapper[4556]: I0218 09:08:14.502479 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Feb 18 09:08:14 crc kubenswrapper[4556]: I0218 09:08:14.525372 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Feb 18 09:08:14 crc kubenswrapper[4556]: I0218 09:08:14.619385 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 18 09:08:14 crc kubenswrapper[4556]: I0218 09:08:14.639176 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Feb 18 09:08:14 crc kubenswrapper[4556]: I0218 09:08:14.667333 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Feb 18 09:08:14 crc kubenswrapper[4556]: I0218 09:08:14.732607 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Feb 18 09:08:14 crc kubenswrapper[4556]: I0218 09:08:14.764065 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Feb 18 09:08:14 crc kubenswrapper[4556]: I0218 09:08:14.780173 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Feb 18 09:08:14 crc kubenswrapper[4556]: I0218 09:08:14.947118 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Feb 18 09:08:15 crc kubenswrapper[4556]: I0218 09:08:15.005808 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Feb 18 09:08:15 crc kubenswrapper[4556]: I0218 09:08:15.059084 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Feb 18 09:08:15 crc kubenswrapper[4556]: I0218 09:08:15.105905 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Feb 18 09:08:15 crc kubenswrapper[4556]: I0218 09:08:15.119347 4556 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Feb 18 09:08:15 crc kubenswrapper[4556]: I0218 09:08:15.125026 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 18 09:08:15 crc kubenswrapper[4556]: I0218 09:08:15.125072 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 18 09:08:15 crc kubenswrapper[4556]: I0218 09:08:15.128299 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 09:08:15 crc kubenswrapper[4556]: I0218 09:08:15.139425 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=21.139361376 podStartE2EDuration="21.139361376s" podCreationTimestamp="2026-02-18 09:07:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:08:15.136847434 +0000 UTC m=+252.153808414" watchObservedRunningTime="2026-02-18 09:08:15.139361376 +0000 UTC m=+252.156322356" Feb 18 09:08:15 crc kubenswrapper[4556]: I0218 09:08:15.217820 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Feb 18 09:08:15 crc kubenswrapper[4556]: I0218 09:08:15.389607 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Feb 18 09:08:15 crc kubenswrapper[4556]: I0218 09:08:15.499206 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 18 09:08:15 crc kubenswrapper[4556]: I0218 09:08:15.560318 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Feb 18 09:08:15 crc kubenswrapper[4556]: I0218 09:08:15.830687 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Feb 18 09:08:15 crc kubenswrapper[4556]: I0218 09:08:15.845747 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Feb 18 09:08:15 crc kubenswrapper[4556]: I0218 09:08:15.853434 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Feb 18 09:08:15 crc kubenswrapper[4556]: I0218 09:08:15.922355 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Feb 18 09:08:15 crc kubenswrapper[4556]: I0218 09:08:15.995534 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Feb 18 09:08:15 crc kubenswrapper[4556]: I0218 09:08:15.998085 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Feb 18 09:08:16 crc kubenswrapper[4556]: I0218 09:08:16.035920 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Feb 18 09:08:16 crc kubenswrapper[4556]: I0218 09:08:16.077709 4556 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 18 09:08:16 crc kubenswrapper[4556]: I0218 09:08:16.077927 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://00d23fad26f6025c9d80e4438c939e4c205d81d1edb047fc31f1c19450a6bf26" gracePeriod=5 Feb 18 09:08:16 crc kubenswrapper[4556]: I0218 09:08:16.200559 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Feb 18 09:08:16 crc kubenswrapper[4556]: I0218 09:08:16.211653 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Feb 18 09:08:16 crc kubenswrapper[4556]: I0218 09:08:16.267784 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Feb 18 09:08:16 crc kubenswrapper[4556]: I0218 09:08:16.277535 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Feb 18 09:08:16 crc kubenswrapper[4556]: I0218 09:08:16.482190 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Feb 18 09:08:16 crc kubenswrapper[4556]: I0218 09:08:16.514076 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Feb 18 09:08:16 crc kubenswrapper[4556]: I0218 09:08:16.604935 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Feb 18 09:08:16 crc kubenswrapper[4556]: I0218 09:08:16.671442 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Feb 18 09:08:16 crc kubenswrapper[4556]: I0218 09:08:16.708590 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Feb 18 09:08:16 crc kubenswrapper[4556]: I0218 09:08:16.746246 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Feb 18 09:08:16 crc kubenswrapper[4556]: I0218 09:08:16.924467 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Feb 18 09:08:17 crc kubenswrapper[4556]: I0218 09:08:17.191046 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 18 09:08:17 crc kubenswrapper[4556]: I0218 09:08:17.240836 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Feb 18 09:08:17 crc kubenswrapper[4556]: I0218 09:08:17.249862 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 18 09:08:17 crc kubenswrapper[4556]: I0218 09:08:17.301797 4556 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Feb 18 09:08:17 crc kubenswrapper[4556]: I0218 09:08:17.380857 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Feb 18 09:08:17 crc kubenswrapper[4556]: I0218 09:08:17.438393 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Feb 18 09:08:17 crc kubenswrapper[4556]: I0218 09:08:17.499789 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Feb 18 09:08:17 crc kubenswrapper[4556]: I0218 09:08:17.577502 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Feb 18 09:08:17 crc kubenswrapper[4556]: I0218 09:08:17.580061 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Feb 18 09:08:17 crc kubenswrapper[4556]: I0218 09:08:17.653376 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Feb 18 09:08:17 crc kubenswrapper[4556]: I0218 09:08:17.731455 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Feb 18 09:08:17 crc kubenswrapper[4556]: I0218 09:08:17.828822 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Feb 18 09:08:17 crc kubenswrapper[4556]: I0218 09:08:17.848845 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Feb 18 09:08:17 crc kubenswrapper[4556]: I0218 09:08:17.848998 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Feb 18 09:08:18 crc kubenswrapper[4556]: I0218 09:08:18.154403 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Feb 18 09:08:18 crc kubenswrapper[4556]: I0218 09:08:18.155979 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Feb 18 09:08:18 crc kubenswrapper[4556]: I0218 09:08:18.167664 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Feb 18 09:08:18 crc kubenswrapper[4556]: I0218 09:08:18.196775 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Feb 18 09:08:18 crc kubenswrapper[4556]: I0218 09:08:18.221046 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Feb 18 09:08:18 crc kubenswrapper[4556]: I0218 09:08:18.253828 4556 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Feb 18 09:08:18 crc kubenswrapper[4556]: I0218 09:08:18.300709 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Feb 18 09:08:18 crc kubenswrapper[4556]: I0218 09:08:18.388408 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Feb 18 09:08:18 crc kubenswrapper[4556]: I0218 09:08:18.547086 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Feb 18 09:08:18 crc kubenswrapper[4556]: I0218 09:08:18.547943 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Feb 18 09:08:18 crc kubenswrapper[4556]: I0218 09:08:18.552951 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Feb 18 09:08:18 crc kubenswrapper[4556]: I0218 09:08:18.734079 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Feb 18 09:08:18 crc kubenswrapper[4556]: I0218 09:08:18.774546 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Feb 18 09:08:18 crc kubenswrapper[4556]: I0218 09:08:18.778087 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Feb 18 09:08:18 crc kubenswrapper[4556]: I0218 09:08:18.818722 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Feb 18 09:08:18 crc kubenswrapper[4556]: I0218 09:08:18.842540 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Feb 18 09:08:18 crc kubenswrapper[4556]: I0218 09:08:18.950032 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Feb 18 09:08:18 crc kubenswrapper[4556]: I0218 09:08:18.972980 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Feb 18 09:08:19 crc kubenswrapper[4556]: I0218 09:08:19.022578 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Feb 18 09:08:19 crc kubenswrapper[4556]: I0218 09:08:19.164719 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Feb 18 09:08:19 crc kubenswrapper[4556]: I0218 09:08:19.176268 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Feb 18 09:08:19 crc kubenswrapper[4556]: I0218 09:08:19.198032 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Feb 18 09:08:19 crc kubenswrapper[4556]: I0218 09:08:19.314381 4556 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Feb 18 09:08:19 crc kubenswrapper[4556]: I0218 09:08:19.358792 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Feb 18 09:08:19 crc kubenswrapper[4556]: I0218 09:08:19.412493 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Feb 18 09:08:19 crc kubenswrapper[4556]: I0218 09:08:19.519069 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Feb 18 09:08:19 crc kubenswrapper[4556]: I0218 09:08:19.547386 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Feb 18 09:08:19 crc kubenswrapper[4556]: I0218 09:08:19.654983 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 18 09:08:19 crc kubenswrapper[4556]: I0218 09:08:19.695456 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Feb 18 09:08:19 crc kubenswrapper[4556]: I0218 09:08:19.790708 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Feb 18 09:08:19 crc kubenswrapper[4556]: I0218 09:08:19.800257 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Feb 18 09:08:19 crc kubenswrapper[4556]: I0218 09:08:19.840752 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Feb 18 09:08:19 crc kubenswrapper[4556]: I0218 09:08:19.879277 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Feb 18 09:08:19 crc kubenswrapper[4556]: I0218 09:08:19.957514 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Feb 18 09:08:20 crc kubenswrapper[4556]: I0218 09:08:20.010812 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Feb 18 09:08:20 crc kubenswrapper[4556]: I0218 09:08:20.047618 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 18 09:08:20 crc kubenswrapper[4556]: I0218 09:08:20.088248 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Feb 18 09:08:20 crc kubenswrapper[4556]: I0218 09:08:20.227225 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Feb 18 09:08:20 crc kubenswrapper[4556]: I0218 09:08:20.253633 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Feb 18 09:08:20 crc kubenswrapper[4556]: I0218 09:08:20.302570 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Feb 18 09:08:20 crc kubenswrapper[4556]: I0218 09:08:20.307949 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Feb 18 09:08:20 crc kubenswrapper[4556]: I0218 09:08:20.319803 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Feb 18 09:08:20 crc kubenswrapper[4556]: I0218 09:08:20.472531 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Feb 18 09:08:20 crc kubenswrapper[4556]: I0218 09:08:20.514587 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Feb 18 09:08:20 crc kubenswrapper[4556]: I0218 09:08:20.546883 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Feb 18 09:08:20 crc kubenswrapper[4556]: I0218 09:08:20.548518 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Feb 18 09:08:20 crc kubenswrapper[4556]: I0218 09:08:20.700876 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Feb 18 09:08:20 crc kubenswrapper[4556]: I0218 09:08:20.730830 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Feb 18 09:08:20 crc kubenswrapper[4556]: I0218 09:08:20.904011 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Feb 18 09:08:21 crc kubenswrapper[4556]: I0218 09:08:21.017783 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Feb 18 09:08:21 crc kubenswrapper[4556]: I0218 09:08:21.073817 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Feb 18 09:08:21 crc kubenswrapper[4556]: I0218 09:08:21.080274 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Feb 18 09:08:21 crc kubenswrapper[4556]: I0218 09:08:21.102675 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Feb 18 09:08:21 crc kubenswrapper[4556]: I0218 09:08:21.144701 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Feb 18 09:08:21 crc kubenswrapper[4556]: I0218 09:08:21.150861 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Feb 18 09:08:21 crc kubenswrapper[4556]: I0218 09:08:21.152216 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Feb 18 09:08:21 crc kubenswrapper[4556]: I0218 09:08:21.208005 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Feb 18 09:08:21 crc kubenswrapper[4556]: I0218 09:08:21.290201 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Feb 18 09:08:21 crc kubenswrapper[4556]: I0218 09:08:21.480085 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Feb 18 09:08:21 crc kubenswrapper[4556]: I0218 09:08:21.484128 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Feb 18 09:08:21 crc kubenswrapper[4556]: I0218 09:08:21.484185 4556 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="00d23fad26f6025c9d80e4438c939e4c205d81d1edb047fc31f1c19450a6bf26" exitCode=137 Feb 18 09:08:21 crc kubenswrapper[4556]: I0218 09:08:21.554876 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Feb 18 09:08:21 crc kubenswrapper[4556]: I0218 09:08:21.581166 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Feb 18 09:08:21 crc kubenswrapper[4556]: I0218 09:08:21.593645 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Feb 18 09:08:21 crc kubenswrapper[4556]: I0218 09:08:21.627180 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 18 09:08:21 crc kubenswrapper[4556]: I0218 09:08:21.643466 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Feb 18 09:08:21 crc kubenswrapper[4556]: I0218 09:08:21.643530 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 09:08:21 crc kubenswrapper[4556]: I0218 09:08:21.737542 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Feb 18 09:08:21 crc kubenswrapper[4556]: I0218 09:08:21.774871 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 18 09:08:21 crc kubenswrapper[4556]: I0218 09:08:21.774937 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 18 09:08:21 crc kubenswrapper[4556]: I0218 09:08:21.775000 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 18 09:08:21 crc kubenswrapper[4556]: I0218 09:08:21.775015 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 18 09:08:21 crc kubenswrapper[4556]: I0218 09:08:21.775099 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 18 09:08:21 crc kubenswrapper[4556]: I0218 09:08:21.775214 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 09:08:21 crc kubenswrapper[4556]: I0218 09:08:21.775265 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 09:08:21 crc kubenswrapper[4556]: I0218 09:08:21.775292 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 09:08:21 crc kubenswrapper[4556]: I0218 09:08:21.775312 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 09:08:21 crc kubenswrapper[4556]: I0218 09:08:21.775388 4556 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Feb 18 09:08:21 crc kubenswrapper[4556]: I0218 09:08:21.775406 4556 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Feb 18 09:08:21 crc kubenswrapper[4556]: I0218 09:08:21.775415 4556 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 18 09:08:21 crc kubenswrapper[4556]: I0218 09:08:21.775423 4556 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Feb 18 09:08:21 crc kubenswrapper[4556]: I0218 09:08:21.781327 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 09:08:21 crc kubenswrapper[4556]: I0218 09:08:21.809925 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Feb 18 09:08:21 crc kubenswrapper[4556]: I0218 09:08:21.876411 4556 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 18 09:08:21 crc kubenswrapper[4556]: I0218 09:08:21.892370 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Feb 18 09:08:21 crc kubenswrapper[4556]: I0218 09:08:21.954051 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Feb 18 09:08:21 crc kubenswrapper[4556]: I0218 09:08:21.955134 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Feb 18 09:08:22 crc kubenswrapper[4556]: I0218 09:08:22.055868 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Feb 18 09:08:22 crc kubenswrapper[4556]: I0218 09:08:22.076644 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Feb 18 09:08:22 crc kubenswrapper[4556]: I0218 09:08:22.076788 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Feb 18 09:08:22 crc kubenswrapper[4556]: I0218 09:08:22.298645 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Feb 18 09:08:22 crc kubenswrapper[4556]: I0218 09:08:22.490661 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Feb 18 09:08:22 crc kubenswrapper[4556]: I0218 09:08:22.490944 4556 scope.go:117] "RemoveContainer" containerID="00d23fad26f6025c9d80e4438c939e4c205d81d1edb047fc31f1c19450a6bf26" Feb 18 09:08:22 crc kubenswrapper[4556]: I0218 09:08:22.491020 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 09:08:22 crc kubenswrapper[4556]: I0218 09:08:22.627537 4556 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Feb 18 09:08:22 crc kubenswrapper[4556]: I0218 09:08:22.673705 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Feb 18 09:08:22 crc kubenswrapper[4556]: I0218 09:08:22.744751 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Feb 18 09:08:22 crc kubenswrapper[4556]: I0218 09:08:22.828896 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Feb 18 09:08:22 crc kubenswrapper[4556]: I0218 09:08:22.866459 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Feb 18 09:08:23 crc kubenswrapper[4556]: I0218 09:08:23.077448 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Feb 18 09:08:23 crc kubenswrapper[4556]: I0218 09:08:23.172557 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Feb 18 09:08:23 crc kubenswrapper[4556]: I0218 09:08:23.286921 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Feb 18 09:08:23 crc kubenswrapper[4556]: I0218 09:08:23.352258 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Feb 18 09:08:23 crc kubenswrapper[4556]: I0218 09:08:23.611242 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Feb 18 09:08:23 crc kubenswrapper[4556]: I0218 09:08:23.806040 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Feb 18 09:08:25 crc kubenswrapper[4556]: I0218 09:08:25.117062 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Feb 18 09:08:25 crc kubenswrapper[4556]: I0218 09:08:25.497994 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Feb 18 09:08:44 crc kubenswrapper[4556]: I0218 09:08:44.476053 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-vd7f7"] Feb 18 09:08:44 crc kubenswrapper[4556]: I0218 09:08:44.476624 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-vd7f7" podUID="9023a93b-33ee-447e-8278-7b4d752d7a53" containerName="controller-manager" containerID="cri-o://786f093e004a61c1e23c5430334210f564207c4356aeb3a685a705740c5d746e" gracePeriod=30 Feb 18 09:08:44 crc kubenswrapper[4556]: I0218 09:08:44.479143 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-2sdct"] Feb 18 09:08:44 crc kubenswrapper[4556]: I0218 09:08:44.479297 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2sdct" podUID="6f9fe7cf-4384-4a95-abbb-20b487e7bc6c" containerName="route-controller-manager" containerID="cri-o://8373698cd99541c1c4ee31d8fe5af9ed9c9634f75e7ed7da6b5a85dd9a493f96" gracePeriod=30 Feb 18 09:08:44 crc kubenswrapper[4556]: I0218 09:08:44.730783 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2sdct" Feb 18 09:08:44 crc kubenswrapper[4556]: I0218 09:08:44.772168 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-vd7f7" Feb 18 09:08:44 crc kubenswrapper[4556]: I0218 09:08:44.906989 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9023a93b-33ee-447e-8278-7b4d752d7a53-proxy-ca-bundles\") pod \"9023a93b-33ee-447e-8278-7b4d752d7a53\" (UID: \"9023a93b-33ee-447e-8278-7b4d752d7a53\") " Feb 18 09:08:44 crc kubenswrapper[4556]: I0218 09:08:44.907035 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6f9fe7cf-4384-4a95-abbb-20b487e7bc6c-client-ca\") pod \"6f9fe7cf-4384-4a95-abbb-20b487e7bc6c\" (UID: \"6f9fe7cf-4384-4a95-abbb-20b487e7bc6c\") " Feb 18 09:08:44 crc kubenswrapper[4556]: I0218 09:08:44.907063 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ktxmv\" (UniqueName: \"kubernetes.io/projected/6f9fe7cf-4384-4a95-abbb-20b487e7bc6c-kube-api-access-ktxmv\") pod \"6f9fe7cf-4384-4a95-abbb-20b487e7bc6c\" (UID: \"6f9fe7cf-4384-4a95-abbb-20b487e7bc6c\") " Feb 18 09:08:44 crc kubenswrapper[4556]: I0218 09:08:44.907101 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f9fe7cf-4384-4a95-abbb-20b487e7bc6c-config\") pod \"6f9fe7cf-4384-4a95-abbb-20b487e7bc6c\" (UID: \"6f9fe7cf-4384-4a95-abbb-20b487e7bc6c\") " Feb 18 09:08:44 crc kubenswrapper[4556]: I0218 09:08:44.907133 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jlvvv\" (UniqueName: \"kubernetes.io/projected/9023a93b-33ee-447e-8278-7b4d752d7a53-kube-api-access-jlvvv\") pod \"9023a93b-33ee-447e-8278-7b4d752d7a53\" (UID: \"9023a93b-33ee-447e-8278-7b4d752d7a53\") " Feb 18 09:08:44 crc kubenswrapper[4556]: I0218 09:08:44.907193 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9023a93b-33ee-447e-8278-7b4d752d7a53-client-ca\") pod \"9023a93b-33ee-447e-8278-7b4d752d7a53\" (UID: \"9023a93b-33ee-447e-8278-7b4d752d7a53\") " Feb 18 09:08:44 crc kubenswrapper[4556]: I0218 09:08:44.907215 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6f9fe7cf-4384-4a95-abbb-20b487e7bc6c-serving-cert\") pod \"6f9fe7cf-4384-4a95-abbb-20b487e7bc6c\" (UID: \"6f9fe7cf-4384-4a95-abbb-20b487e7bc6c\") " Feb 18 09:08:44 crc kubenswrapper[4556]: I0218 09:08:44.907238 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9023a93b-33ee-447e-8278-7b4d752d7a53-config\") pod \"9023a93b-33ee-447e-8278-7b4d752d7a53\" (UID: \"9023a93b-33ee-447e-8278-7b4d752d7a53\") " Feb 18 09:08:44 crc kubenswrapper[4556]: I0218 09:08:44.907261 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9023a93b-33ee-447e-8278-7b4d752d7a53-serving-cert\") pod \"9023a93b-33ee-447e-8278-7b4d752d7a53\" (UID: \"9023a93b-33ee-447e-8278-7b4d752d7a53\") " Feb 18 09:08:44 crc kubenswrapper[4556]: I0218 09:08:44.907789 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9023a93b-33ee-447e-8278-7b4d752d7a53-client-ca" (OuterVolumeSpecName: "client-ca") pod "9023a93b-33ee-447e-8278-7b4d752d7a53" (UID: "9023a93b-33ee-447e-8278-7b4d752d7a53"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:08:44 crc kubenswrapper[4556]: I0218 09:08:44.907838 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f9fe7cf-4384-4a95-abbb-20b487e7bc6c-client-ca" (OuterVolumeSpecName: "client-ca") pod "6f9fe7cf-4384-4a95-abbb-20b487e7bc6c" (UID: "6f9fe7cf-4384-4a95-abbb-20b487e7bc6c"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:08:44 crc kubenswrapper[4556]: I0218 09:08:44.907875 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9023a93b-33ee-447e-8278-7b4d752d7a53-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "9023a93b-33ee-447e-8278-7b4d752d7a53" (UID: "9023a93b-33ee-447e-8278-7b4d752d7a53"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:08:44 crc kubenswrapper[4556]: I0218 09:08:44.907943 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f9fe7cf-4384-4a95-abbb-20b487e7bc6c-config" (OuterVolumeSpecName: "config") pod "6f9fe7cf-4384-4a95-abbb-20b487e7bc6c" (UID: "6f9fe7cf-4384-4a95-abbb-20b487e7bc6c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:08:44 crc kubenswrapper[4556]: I0218 09:08:44.907974 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9023a93b-33ee-447e-8278-7b4d752d7a53-config" (OuterVolumeSpecName: "config") pod "9023a93b-33ee-447e-8278-7b4d752d7a53" (UID: "9023a93b-33ee-447e-8278-7b4d752d7a53"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:08:44 crc kubenswrapper[4556]: I0218 09:08:44.911641 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f9fe7cf-4384-4a95-abbb-20b487e7bc6c-kube-api-access-ktxmv" (OuterVolumeSpecName: "kube-api-access-ktxmv") pod "6f9fe7cf-4384-4a95-abbb-20b487e7bc6c" (UID: "6f9fe7cf-4384-4a95-abbb-20b487e7bc6c"). InnerVolumeSpecName "kube-api-access-ktxmv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:08:44 crc kubenswrapper[4556]: I0218 09:08:44.911643 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f9fe7cf-4384-4a95-abbb-20b487e7bc6c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6f9fe7cf-4384-4a95-abbb-20b487e7bc6c" (UID: "6f9fe7cf-4384-4a95-abbb-20b487e7bc6c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:08:44 crc kubenswrapper[4556]: I0218 09:08:44.911726 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9023a93b-33ee-447e-8278-7b4d752d7a53-kube-api-access-jlvvv" (OuterVolumeSpecName: "kube-api-access-jlvvv") pod "9023a93b-33ee-447e-8278-7b4d752d7a53" (UID: "9023a93b-33ee-447e-8278-7b4d752d7a53"). InnerVolumeSpecName "kube-api-access-jlvvv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:08:44 crc kubenswrapper[4556]: I0218 09:08:44.911835 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9023a93b-33ee-447e-8278-7b4d752d7a53-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9023a93b-33ee-447e-8278-7b4d752d7a53" (UID: "9023a93b-33ee-447e-8278-7b4d752d7a53"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:08:45 crc kubenswrapper[4556]: I0218 09:08:45.008574 4556 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9023a93b-33ee-447e-8278-7b4d752d7a53-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 09:08:45 crc kubenswrapper[4556]: I0218 09:08:45.008599 4556 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9023a93b-33ee-447e-8278-7b4d752d7a53-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 18 09:08:45 crc kubenswrapper[4556]: I0218 09:08:45.008610 4556 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6f9fe7cf-4384-4a95-abbb-20b487e7bc6c-client-ca\") on node \"crc\" DevicePath \"\"" Feb 18 09:08:45 crc kubenswrapper[4556]: I0218 09:08:45.008620 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ktxmv\" (UniqueName: \"kubernetes.io/projected/6f9fe7cf-4384-4a95-abbb-20b487e7bc6c-kube-api-access-ktxmv\") on node \"crc\" DevicePath \"\"" Feb 18 09:08:45 crc kubenswrapper[4556]: I0218 09:08:45.008628 4556 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f9fe7cf-4384-4a95-abbb-20b487e7bc6c-config\") on node \"crc\" DevicePath \"\"" Feb 18 09:08:45 crc kubenswrapper[4556]: I0218 09:08:45.008638 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jlvvv\" (UniqueName: \"kubernetes.io/projected/9023a93b-33ee-447e-8278-7b4d752d7a53-kube-api-access-jlvvv\") on node \"crc\" DevicePath \"\"" Feb 18 09:08:45 crc kubenswrapper[4556]: I0218 09:08:45.008646 4556 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9023a93b-33ee-447e-8278-7b4d752d7a53-client-ca\") on node \"crc\" DevicePath \"\"" Feb 18 09:08:45 crc kubenswrapper[4556]: I0218 09:08:45.008652 4556 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6f9fe7cf-4384-4a95-abbb-20b487e7bc6c-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 09:08:45 crc kubenswrapper[4556]: I0218 09:08:45.008660 4556 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9023a93b-33ee-447e-8278-7b4d752d7a53-config\") on node \"crc\" DevicePath \"\"" Feb 18 09:08:45 crc kubenswrapper[4556]: I0218 09:08:45.572301 4556 generic.go:334] "Generic (PLEG): container finished" podID="9023a93b-33ee-447e-8278-7b4d752d7a53" containerID="786f093e004a61c1e23c5430334210f564207c4356aeb3a685a705740c5d746e" exitCode=0 Feb 18 09:08:45 crc kubenswrapper[4556]: I0218 09:08:45.572365 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-vd7f7" event={"ID":"9023a93b-33ee-447e-8278-7b4d752d7a53","Type":"ContainerDied","Data":"786f093e004a61c1e23c5430334210f564207c4356aeb3a685a705740c5d746e"} Feb 18 09:08:45 crc kubenswrapper[4556]: I0218 09:08:45.572407 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-vd7f7" Feb 18 09:08:45 crc kubenswrapper[4556]: I0218 09:08:45.572425 4556 scope.go:117] "RemoveContainer" containerID="786f093e004a61c1e23c5430334210f564207c4356aeb3a685a705740c5d746e" Feb 18 09:08:45 crc kubenswrapper[4556]: I0218 09:08:45.572414 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-vd7f7" event={"ID":"9023a93b-33ee-447e-8278-7b4d752d7a53","Type":"ContainerDied","Data":"271e02cd08d44d1fd7f6f684bfc79ff7aea467fdded5ca54f6586a450d7f6e0b"} Feb 18 09:08:45 crc kubenswrapper[4556]: I0218 09:08:45.573626 4556 generic.go:334] "Generic (PLEG): container finished" podID="6f9fe7cf-4384-4a95-abbb-20b487e7bc6c" containerID="8373698cd99541c1c4ee31d8fe5af9ed9c9634f75e7ed7da6b5a85dd9a493f96" exitCode=0 Feb 18 09:08:45 crc kubenswrapper[4556]: I0218 09:08:45.573647 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2sdct" event={"ID":"6f9fe7cf-4384-4a95-abbb-20b487e7bc6c","Type":"ContainerDied","Data":"8373698cd99541c1c4ee31d8fe5af9ed9c9634f75e7ed7da6b5a85dd9a493f96"} Feb 18 09:08:45 crc kubenswrapper[4556]: I0218 09:08:45.573652 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2sdct" Feb 18 09:08:45 crc kubenswrapper[4556]: I0218 09:08:45.573661 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2sdct" event={"ID":"6f9fe7cf-4384-4a95-abbb-20b487e7bc6c","Type":"ContainerDied","Data":"92dfcf5c158636ba7d85bb08f3c643754a755c541be4e5c735fbeb96e693a266"} Feb 18 09:08:45 crc kubenswrapper[4556]: I0218 09:08:45.584085 4556 scope.go:117] "RemoveContainer" containerID="786f093e004a61c1e23c5430334210f564207c4356aeb3a685a705740c5d746e" Feb 18 09:08:45 crc kubenswrapper[4556]: E0218 09:08:45.584376 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"786f093e004a61c1e23c5430334210f564207c4356aeb3a685a705740c5d746e\": container with ID starting with 786f093e004a61c1e23c5430334210f564207c4356aeb3a685a705740c5d746e not found: ID does not exist" containerID="786f093e004a61c1e23c5430334210f564207c4356aeb3a685a705740c5d746e" Feb 18 09:08:45 crc kubenswrapper[4556]: I0218 09:08:45.584414 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"786f093e004a61c1e23c5430334210f564207c4356aeb3a685a705740c5d746e"} err="failed to get container status \"786f093e004a61c1e23c5430334210f564207c4356aeb3a685a705740c5d746e\": rpc error: code = NotFound desc = could not find container \"786f093e004a61c1e23c5430334210f564207c4356aeb3a685a705740c5d746e\": container with ID starting with 786f093e004a61c1e23c5430334210f564207c4356aeb3a685a705740c5d746e not found: ID does not exist" Feb 18 09:08:45 crc kubenswrapper[4556]: I0218 09:08:45.584442 4556 scope.go:117] "RemoveContainer" containerID="8373698cd99541c1c4ee31d8fe5af9ed9c9634f75e7ed7da6b5a85dd9a493f96" Feb 18 09:08:45 crc kubenswrapper[4556]: I0218 09:08:45.584775 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-vd7f7"] Feb 18 09:08:45 crc kubenswrapper[4556]: I0218 09:08:45.587204 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-vd7f7"] Feb 18 09:08:45 crc kubenswrapper[4556]: I0218 09:08:45.592479 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-2sdct"] Feb 18 09:08:45 crc kubenswrapper[4556]: I0218 09:08:45.594256 4556 scope.go:117] "RemoveContainer" containerID="8373698cd99541c1c4ee31d8fe5af9ed9c9634f75e7ed7da6b5a85dd9a493f96" Feb 18 09:08:45 crc kubenswrapper[4556]: E0218 09:08:45.594584 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8373698cd99541c1c4ee31d8fe5af9ed9c9634f75e7ed7da6b5a85dd9a493f96\": container with ID starting with 8373698cd99541c1c4ee31d8fe5af9ed9c9634f75e7ed7da6b5a85dd9a493f96 not found: ID does not exist" containerID="8373698cd99541c1c4ee31d8fe5af9ed9c9634f75e7ed7da6b5a85dd9a493f96" Feb 18 09:08:45 crc kubenswrapper[4556]: I0218 09:08:45.594603 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8373698cd99541c1c4ee31d8fe5af9ed9c9634f75e7ed7da6b5a85dd9a493f96"} err="failed to get container status \"8373698cd99541c1c4ee31d8fe5af9ed9c9634f75e7ed7da6b5a85dd9a493f96\": rpc error: code = NotFound desc = could not find container \"8373698cd99541c1c4ee31d8fe5af9ed9c9634f75e7ed7da6b5a85dd9a493f96\": container with ID starting with 8373698cd99541c1c4ee31d8fe5af9ed9c9634f75e7ed7da6b5a85dd9a493f96 not found: ID does not exist" Feb 18 09:08:45 crc kubenswrapper[4556]: I0218 09:08:45.597735 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-2sdct"] Feb 18 09:08:45 crc kubenswrapper[4556]: I0218 09:08:45.972182 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7d6d847d76-dktfz"] Feb 18 09:08:45 crc kubenswrapper[4556]: E0218 09:08:45.972490 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 18 09:08:45 crc kubenswrapper[4556]: I0218 09:08:45.972505 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 18 09:08:45 crc kubenswrapper[4556]: E0218 09:08:45.972515 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f9fe7cf-4384-4a95-abbb-20b487e7bc6c" containerName="route-controller-manager" Feb 18 09:08:45 crc kubenswrapper[4556]: I0218 09:08:45.972521 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f9fe7cf-4384-4a95-abbb-20b487e7bc6c" containerName="route-controller-manager" Feb 18 09:08:45 crc kubenswrapper[4556]: E0218 09:08:45.972529 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12967c01-f63e-466e-b28f-4828af636833" containerName="installer" Feb 18 09:08:45 crc kubenswrapper[4556]: I0218 09:08:45.972535 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="12967c01-f63e-466e-b28f-4828af636833" containerName="installer" Feb 18 09:08:45 crc kubenswrapper[4556]: E0218 09:08:45.972542 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9023a93b-33ee-447e-8278-7b4d752d7a53" containerName="controller-manager" Feb 18 09:08:45 crc kubenswrapper[4556]: I0218 09:08:45.972563 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="9023a93b-33ee-447e-8278-7b4d752d7a53" containerName="controller-manager" Feb 18 09:08:45 crc kubenswrapper[4556]: I0218 09:08:45.972650 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="12967c01-f63e-466e-b28f-4828af636833" containerName="installer" Feb 18 09:08:45 crc kubenswrapper[4556]: I0218 09:08:45.972662 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 18 09:08:45 crc kubenswrapper[4556]: I0218 09:08:45.972669 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="9023a93b-33ee-447e-8278-7b4d752d7a53" containerName="controller-manager" Feb 18 09:08:45 crc kubenswrapper[4556]: I0218 09:08:45.972675 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f9fe7cf-4384-4a95-abbb-20b487e7bc6c" containerName="route-controller-manager" Feb 18 09:08:45 crc kubenswrapper[4556]: I0218 09:08:45.973066 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7d6d847d76-dktfz" Feb 18 09:08:45 crc kubenswrapper[4556]: I0218 09:08:45.974200 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 18 09:08:45 crc kubenswrapper[4556]: I0218 09:08:45.974812 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 18 09:08:45 crc kubenswrapper[4556]: I0218 09:08:45.975295 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 18 09:08:45 crc kubenswrapper[4556]: I0218 09:08:45.975529 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 18 09:08:45 crc kubenswrapper[4556]: I0218 09:08:45.975708 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 18 09:08:45 crc kubenswrapper[4556]: I0218 09:08:45.976863 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6fcf8b4c5d-bs84s"] Feb 18 09:08:45 crc kubenswrapper[4556]: I0218 09:08:45.977507 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6fcf8b4c5d-bs84s" Feb 18 09:08:45 crc kubenswrapper[4556]: I0218 09:08:45.978139 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 18 09:08:45 crc kubenswrapper[4556]: I0218 09:08:45.978720 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 18 09:08:45 crc kubenswrapper[4556]: I0218 09:08:45.978762 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 18 09:08:45 crc kubenswrapper[4556]: I0218 09:08:45.978840 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 18 09:08:45 crc kubenswrapper[4556]: I0218 09:08:45.978907 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 18 09:08:45 crc kubenswrapper[4556]: I0218 09:08:45.979147 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 18 09:08:45 crc kubenswrapper[4556]: I0218 09:08:45.979491 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 18 09:08:45 crc kubenswrapper[4556]: I0218 09:08:45.986587 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7d6d847d76-dktfz"] Feb 18 09:08:45 crc kubenswrapper[4556]: I0218 09:08:45.991481 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 18 09:08:45 crc kubenswrapper[4556]: I0218 09:08:45.991611 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6fcf8b4c5d-bs84s"] Feb 18 09:08:46 crc kubenswrapper[4556]: I0218 09:08:46.018984 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lh6v\" (UniqueName: \"kubernetes.io/projected/97ee9f2e-ab2e-4436-acdb-12c892f54209-kube-api-access-2lh6v\") pod \"controller-manager-7d6d847d76-dktfz\" (UID: \"97ee9f2e-ab2e-4436-acdb-12c892f54209\") " pod="openshift-controller-manager/controller-manager-7d6d847d76-dktfz" Feb 18 09:08:46 crc kubenswrapper[4556]: I0218 09:08:46.019076 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a9801216-0000-4dac-b960-afd2a02f9228-serving-cert\") pod \"route-controller-manager-6fcf8b4c5d-bs84s\" (UID: \"a9801216-0000-4dac-b960-afd2a02f9228\") " pod="openshift-route-controller-manager/route-controller-manager-6fcf8b4c5d-bs84s" Feb 18 09:08:46 crc kubenswrapper[4556]: I0218 09:08:46.019118 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjmvg\" (UniqueName: \"kubernetes.io/projected/a9801216-0000-4dac-b960-afd2a02f9228-kube-api-access-bjmvg\") pod \"route-controller-manager-6fcf8b4c5d-bs84s\" (UID: \"a9801216-0000-4dac-b960-afd2a02f9228\") " pod="openshift-route-controller-manager/route-controller-manager-6fcf8b4c5d-bs84s" Feb 18 09:08:46 crc kubenswrapper[4556]: I0218 09:08:46.019146 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a9801216-0000-4dac-b960-afd2a02f9228-client-ca\") pod \"route-controller-manager-6fcf8b4c5d-bs84s\" (UID: \"a9801216-0000-4dac-b960-afd2a02f9228\") " pod="openshift-route-controller-manager/route-controller-manager-6fcf8b4c5d-bs84s" Feb 18 09:08:46 crc kubenswrapper[4556]: I0218 09:08:46.019198 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9801216-0000-4dac-b960-afd2a02f9228-config\") pod \"route-controller-manager-6fcf8b4c5d-bs84s\" (UID: \"a9801216-0000-4dac-b960-afd2a02f9228\") " pod="openshift-route-controller-manager/route-controller-manager-6fcf8b4c5d-bs84s" Feb 18 09:08:46 crc kubenswrapper[4556]: I0218 09:08:46.019250 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/97ee9f2e-ab2e-4436-acdb-12c892f54209-client-ca\") pod \"controller-manager-7d6d847d76-dktfz\" (UID: \"97ee9f2e-ab2e-4436-acdb-12c892f54209\") " pod="openshift-controller-manager/controller-manager-7d6d847d76-dktfz" Feb 18 09:08:46 crc kubenswrapper[4556]: I0218 09:08:46.019270 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/97ee9f2e-ab2e-4436-acdb-12c892f54209-serving-cert\") pod \"controller-manager-7d6d847d76-dktfz\" (UID: \"97ee9f2e-ab2e-4436-acdb-12c892f54209\") " pod="openshift-controller-manager/controller-manager-7d6d847d76-dktfz" Feb 18 09:08:46 crc kubenswrapper[4556]: I0218 09:08:46.019322 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97ee9f2e-ab2e-4436-acdb-12c892f54209-config\") pod \"controller-manager-7d6d847d76-dktfz\" (UID: \"97ee9f2e-ab2e-4436-acdb-12c892f54209\") " pod="openshift-controller-manager/controller-manager-7d6d847d76-dktfz" Feb 18 09:08:46 crc kubenswrapper[4556]: I0218 09:08:46.019387 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/97ee9f2e-ab2e-4436-acdb-12c892f54209-proxy-ca-bundles\") pod \"controller-manager-7d6d847d76-dktfz\" (UID: \"97ee9f2e-ab2e-4436-acdb-12c892f54209\") " pod="openshift-controller-manager/controller-manager-7d6d847d76-dktfz" Feb 18 09:08:46 crc kubenswrapper[4556]: I0218 09:08:46.120021 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/97ee9f2e-ab2e-4436-acdb-12c892f54209-proxy-ca-bundles\") pod \"controller-manager-7d6d847d76-dktfz\" (UID: \"97ee9f2e-ab2e-4436-acdb-12c892f54209\") " pod="openshift-controller-manager/controller-manager-7d6d847d76-dktfz" Feb 18 09:08:46 crc kubenswrapper[4556]: I0218 09:08:46.120069 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lh6v\" (UniqueName: \"kubernetes.io/projected/97ee9f2e-ab2e-4436-acdb-12c892f54209-kube-api-access-2lh6v\") pod \"controller-manager-7d6d847d76-dktfz\" (UID: \"97ee9f2e-ab2e-4436-acdb-12c892f54209\") " pod="openshift-controller-manager/controller-manager-7d6d847d76-dktfz" Feb 18 09:08:46 crc kubenswrapper[4556]: I0218 09:08:46.120091 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a9801216-0000-4dac-b960-afd2a02f9228-serving-cert\") pod \"route-controller-manager-6fcf8b4c5d-bs84s\" (UID: \"a9801216-0000-4dac-b960-afd2a02f9228\") " pod="openshift-route-controller-manager/route-controller-manager-6fcf8b4c5d-bs84s" Feb 18 09:08:46 crc kubenswrapper[4556]: I0218 09:08:46.120110 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjmvg\" (UniqueName: \"kubernetes.io/projected/a9801216-0000-4dac-b960-afd2a02f9228-kube-api-access-bjmvg\") pod \"route-controller-manager-6fcf8b4c5d-bs84s\" (UID: \"a9801216-0000-4dac-b960-afd2a02f9228\") " pod="openshift-route-controller-manager/route-controller-manager-6fcf8b4c5d-bs84s" Feb 18 09:08:46 crc kubenswrapper[4556]: I0218 09:08:46.120131 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a9801216-0000-4dac-b960-afd2a02f9228-client-ca\") pod \"route-controller-manager-6fcf8b4c5d-bs84s\" (UID: \"a9801216-0000-4dac-b960-afd2a02f9228\") " pod="openshift-route-controller-manager/route-controller-manager-6fcf8b4c5d-bs84s" Feb 18 09:08:46 crc kubenswrapper[4556]: I0218 09:08:46.120341 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9801216-0000-4dac-b960-afd2a02f9228-config\") pod \"route-controller-manager-6fcf8b4c5d-bs84s\" (UID: \"a9801216-0000-4dac-b960-afd2a02f9228\") " pod="openshift-route-controller-manager/route-controller-manager-6fcf8b4c5d-bs84s" Feb 18 09:08:46 crc kubenswrapper[4556]: I0218 09:08:46.120383 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/97ee9f2e-ab2e-4436-acdb-12c892f54209-client-ca\") pod \"controller-manager-7d6d847d76-dktfz\" (UID: \"97ee9f2e-ab2e-4436-acdb-12c892f54209\") " pod="openshift-controller-manager/controller-manager-7d6d847d76-dktfz" Feb 18 09:08:46 crc kubenswrapper[4556]: I0218 09:08:46.120400 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/97ee9f2e-ab2e-4436-acdb-12c892f54209-serving-cert\") pod \"controller-manager-7d6d847d76-dktfz\" (UID: \"97ee9f2e-ab2e-4436-acdb-12c892f54209\") " pod="openshift-controller-manager/controller-manager-7d6d847d76-dktfz" Feb 18 09:08:46 crc kubenswrapper[4556]: I0218 09:08:46.120415 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97ee9f2e-ab2e-4436-acdb-12c892f54209-config\") pod \"controller-manager-7d6d847d76-dktfz\" (UID: \"97ee9f2e-ab2e-4436-acdb-12c892f54209\") " pod="openshift-controller-manager/controller-manager-7d6d847d76-dktfz" Feb 18 09:08:46 crc kubenswrapper[4556]: I0218 09:08:46.121313 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/97ee9f2e-ab2e-4436-acdb-12c892f54209-proxy-ca-bundles\") pod \"controller-manager-7d6d847d76-dktfz\" (UID: \"97ee9f2e-ab2e-4436-acdb-12c892f54209\") " pod="openshift-controller-manager/controller-manager-7d6d847d76-dktfz" Feb 18 09:08:46 crc kubenswrapper[4556]: I0218 09:08:46.121357 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a9801216-0000-4dac-b960-afd2a02f9228-client-ca\") pod \"route-controller-manager-6fcf8b4c5d-bs84s\" (UID: \"a9801216-0000-4dac-b960-afd2a02f9228\") " pod="openshift-route-controller-manager/route-controller-manager-6fcf8b4c5d-bs84s" Feb 18 09:08:46 crc kubenswrapper[4556]: I0218 09:08:46.121384 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/97ee9f2e-ab2e-4436-acdb-12c892f54209-client-ca\") pod \"controller-manager-7d6d847d76-dktfz\" (UID: \"97ee9f2e-ab2e-4436-acdb-12c892f54209\") " pod="openshift-controller-manager/controller-manager-7d6d847d76-dktfz" Feb 18 09:08:46 crc kubenswrapper[4556]: I0218 09:08:46.121621 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97ee9f2e-ab2e-4436-acdb-12c892f54209-config\") pod \"controller-manager-7d6d847d76-dktfz\" (UID: \"97ee9f2e-ab2e-4436-acdb-12c892f54209\") " pod="openshift-controller-manager/controller-manager-7d6d847d76-dktfz" Feb 18 09:08:46 crc kubenswrapper[4556]: I0218 09:08:46.122944 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9801216-0000-4dac-b960-afd2a02f9228-config\") pod \"route-controller-manager-6fcf8b4c5d-bs84s\" (UID: \"a9801216-0000-4dac-b960-afd2a02f9228\") " pod="openshift-route-controller-manager/route-controller-manager-6fcf8b4c5d-bs84s" Feb 18 09:08:46 crc kubenswrapper[4556]: I0218 09:08:46.123917 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a9801216-0000-4dac-b960-afd2a02f9228-serving-cert\") pod \"route-controller-manager-6fcf8b4c5d-bs84s\" (UID: \"a9801216-0000-4dac-b960-afd2a02f9228\") " pod="openshift-route-controller-manager/route-controller-manager-6fcf8b4c5d-bs84s" Feb 18 09:08:46 crc kubenswrapper[4556]: I0218 09:08:46.124016 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/97ee9f2e-ab2e-4436-acdb-12c892f54209-serving-cert\") pod \"controller-manager-7d6d847d76-dktfz\" (UID: \"97ee9f2e-ab2e-4436-acdb-12c892f54209\") " pod="openshift-controller-manager/controller-manager-7d6d847d76-dktfz" Feb 18 09:08:46 crc kubenswrapper[4556]: I0218 09:08:46.132906 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjmvg\" (UniqueName: \"kubernetes.io/projected/a9801216-0000-4dac-b960-afd2a02f9228-kube-api-access-bjmvg\") pod \"route-controller-manager-6fcf8b4c5d-bs84s\" (UID: \"a9801216-0000-4dac-b960-afd2a02f9228\") " pod="openshift-route-controller-manager/route-controller-manager-6fcf8b4c5d-bs84s" Feb 18 09:08:46 crc kubenswrapper[4556]: I0218 09:08:46.133800 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lh6v\" (UniqueName: \"kubernetes.io/projected/97ee9f2e-ab2e-4436-acdb-12c892f54209-kube-api-access-2lh6v\") pod \"controller-manager-7d6d847d76-dktfz\" (UID: \"97ee9f2e-ab2e-4436-acdb-12c892f54209\") " pod="openshift-controller-manager/controller-manager-7d6d847d76-dktfz" Feb 18 09:08:46 crc kubenswrapper[4556]: I0218 09:08:46.285916 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7d6d847d76-dktfz" Feb 18 09:08:46 crc kubenswrapper[4556]: I0218 09:08:46.294953 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6fcf8b4c5d-bs84s" Feb 18 09:08:46 crc kubenswrapper[4556]: I0218 09:08:46.444666 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6fcf8b4c5d-bs84s"] Feb 18 09:08:46 crc kubenswrapper[4556]: I0218 09:08:46.579262 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6fcf8b4c5d-bs84s" event={"ID":"a9801216-0000-4dac-b960-afd2a02f9228","Type":"ContainerStarted","Data":"5c9ccd41034a49596e7881346348a229ccb590261cd58854a0f8d2391849f45e"} Feb 18 09:08:46 crc kubenswrapper[4556]: I0218 09:08:46.579456 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6fcf8b4c5d-bs84s" event={"ID":"a9801216-0000-4dac-b960-afd2a02f9228","Type":"ContainerStarted","Data":"0ac8dbf9ffaee203a21707591f6c43941935c90f32f00edf337005abc512b086"} Feb 18 09:08:46 crc kubenswrapper[4556]: I0218 09:08:46.579474 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6fcf8b4c5d-bs84s" Feb 18 09:08:46 crc kubenswrapper[4556]: I0218 09:08:46.588966 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6fcf8b4c5d-bs84s" podStartSLOduration=2.588951977 podStartE2EDuration="2.588951977s" podCreationTimestamp="2026-02-18 09:08:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:08:46.588549305 +0000 UTC m=+283.605510285" watchObservedRunningTime="2026-02-18 09:08:46.588951977 +0000 UTC m=+283.605912957" Feb 18 09:08:46 crc kubenswrapper[4556]: I0218 09:08:46.612925 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7d6d847d76-dktfz"] Feb 18 09:08:46 crc kubenswrapper[4556]: W0218 09:08:46.615669 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod97ee9f2e_ab2e_4436_acdb_12c892f54209.slice/crio-19d242311525621a371fa73293c502b333603b760142c8c3d4064f3561f99335 WatchSource:0}: Error finding container 19d242311525621a371fa73293c502b333603b760142c8c3d4064f3561f99335: Status 404 returned error can't find the container with id 19d242311525621a371fa73293c502b333603b760142c8c3d4064f3561f99335 Feb 18 09:08:46 crc kubenswrapper[4556]: I0218 09:08:46.860198 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6fcf8b4c5d-bs84s" Feb 18 09:08:47 crc kubenswrapper[4556]: I0218 09:08:47.286751 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f9fe7cf-4384-4a95-abbb-20b487e7bc6c" path="/var/lib/kubelet/pods/6f9fe7cf-4384-4a95-abbb-20b487e7bc6c/volumes" Feb 18 09:08:47 crc kubenswrapper[4556]: I0218 09:08:47.287417 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9023a93b-33ee-447e-8278-7b4d752d7a53" path="/var/lib/kubelet/pods/9023a93b-33ee-447e-8278-7b4d752d7a53/volumes" Feb 18 09:08:47 crc kubenswrapper[4556]: I0218 09:08:47.585925 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7d6d847d76-dktfz" event={"ID":"97ee9f2e-ab2e-4436-acdb-12c892f54209","Type":"ContainerStarted","Data":"abfca06cddcc30451a0ea1c6ff122cfe5056858630c11acb1f76d28a3f713e25"} Feb 18 09:08:47 crc kubenswrapper[4556]: I0218 09:08:47.585969 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7d6d847d76-dktfz" event={"ID":"97ee9f2e-ab2e-4436-acdb-12c892f54209","Type":"ContainerStarted","Data":"19d242311525621a371fa73293c502b333603b760142c8c3d4064f3561f99335"} Feb 18 09:08:47 crc kubenswrapper[4556]: I0218 09:08:47.598606 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7d6d847d76-dktfz" podStartSLOduration=3.598592016 podStartE2EDuration="3.598592016s" podCreationTimestamp="2026-02-18 09:08:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:08:47.596474565 +0000 UTC m=+284.613435546" watchObservedRunningTime="2026-02-18 09:08:47.598592016 +0000 UTC m=+284.615552997" Feb 18 09:08:48 crc kubenswrapper[4556]: I0218 09:08:48.589649 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7d6d847d76-dktfz" Feb 18 09:08:48 crc kubenswrapper[4556]: I0218 09:08:48.592929 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7d6d847d76-dktfz" Feb 18 09:08:54 crc kubenswrapper[4556]: I0218 09:08:54.431929 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7d6d847d76-dktfz"] Feb 18 09:08:54 crc kubenswrapper[4556]: I0218 09:08:54.432452 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-7d6d847d76-dktfz" podUID="97ee9f2e-ab2e-4436-acdb-12c892f54209" containerName="controller-manager" containerID="cri-o://abfca06cddcc30451a0ea1c6ff122cfe5056858630c11acb1f76d28a3f713e25" gracePeriod=30 Feb 18 09:08:54 crc kubenswrapper[4556]: I0218 09:08:54.450188 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6fcf8b4c5d-bs84s"] Feb 18 09:08:54 crc kubenswrapper[4556]: I0218 09:08:54.450454 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6fcf8b4c5d-bs84s" podUID="a9801216-0000-4dac-b960-afd2a02f9228" containerName="route-controller-manager" containerID="cri-o://5c9ccd41034a49596e7881346348a229ccb590261cd58854a0f8d2391849f45e" gracePeriod=30 Feb 18 09:08:54 crc kubenswrapper[4556]: I0218 09:08:54.618971 4556 generic.go:334] "Generic (PLEG): container finished" podID="a9801216-0000-4dac-b960-afd2a02f9228" containerID="5c9ccd41034a49596e7881346348a229ccb590261cd58854a0f8d2391849f45e" exitCode=0 Feb 18 09:08:54 crc kubenswrapper[4556]: I0218 09:08:54.619035 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6fcf8b4c5d-bs84s" event={"ID":"a9801216-0000-4dac-b960-afd2a02f9228","Type":"ContainerDied","Data":"5c9ccd41034a49596e7881346348a229ccb590261cd58854a0f8d2391849f45e"} Feb 18 09:08:54 crc kubenswrapper[4556]: I0218 09:08:54.620386 4556 generic.go:334] "Generic (PLEG): container finished" podID="97ee9f2e-ab2e-4436-acdb-12c892f54209" containerID="abfca06cddcc30451a0ea1c6ff122cfe5056858630c11acb1f76d28a3f713e25" exitCode=0 Feb 18 09:08:54 crc kubenswrapper[4556]: I0218 09:08:54.620426 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7d6d847d76-dktfz" event={"ID":"97ee9f2e-ab2e-4436-acdb-12c892f54209","Type":"ContainerDied","Data":"abfca06cddcc30451a0ea1c6ff122cfe5056858630c11acb1f76d28a3f713e25"} Feb 18 09:08:54 crc kubenswrapper[4556]: I0218 09:08:54.828419 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6fcf8b4c5d-bs84s" Feb 18 09:08:54 crc kubenswrapper[4556]: I0218 09:08:54.857842 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7d6d847d76-dktfz" Feb 18 09:08:54 crc kubenswrapper[4556]: I0218 09:08:54.858393 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7c4d96c475-dgppt"] Feb 18 09:08:54 crc kubenswrapper[4556]: E0218 09:08:54.858585 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97ee9f2e-ab2e-4436-acdb-12c892f54209" containerName="controller-manager" Feb 18 09:08:54 crc kubenswrapper[4556]: I0218 09:08:54.858600 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="97ee9f2e-ab2e-4436-acdb-12c892f54209" containerName="controller-manager" Feb 18 09:08:54 crc kubenswrapper[4556]: E0218 09:08:54.858612 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9801216-0000-4dac-b960-afd2a02f9228" containerName="route-controller-manager" Feb 18 09:08:54 crc kubenswrapper[4556]: I0218 09:08:54.858618 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9801216-0000-4dac-b960-afd2a02f9228" containerName="route-controller-manager" Feb 18 09:08:54 crc kubenswrapper[4556]: I0218 09:08:54.858702 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9801216-0000-4dac-b960-afd2a02f9228" containerName="route-controller-manager" Feb 18 09:08:54 crc kubenswrapper[4556]: I0218 09:08:54.858713 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="97ee9f2e-ab2e-4436-acdb-12c892f54209" containerName="controller-manager" Feb 18 09:08:54 crc kubenswrapper[4556]: I0218 09:08:54.859041 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7c4d96c475-dgppt" Feb 18 09:08:54 crc kubenswrapper[4556]: I0218 09:08:54.865439 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7c4d96c475-dgppt"] Feb 18 09:08:54 crc kubenswrapper[4556]: I0218 09:08:54.910858 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0cba6771-0129-4ae1-b8e0-c13b427939b1-config\") pod \"route-controller-manager-7c4d96c475-dgppt\" (UID: \"0cba6771-0129-4ae1-b8e0-c13b427939b1\") " pod="openshift-route-controller-manager/route-controller-manager-7c4d96c475-dgppt" Feb 18 09:08:54 crc kubenswrapper[4556]: I0218 09:08:54.910954 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82grg\" (UniqueName: \"kubernetes.io/projected/0cba6771-0129-4ae1-b8e0-c13b427939b1-kube-api-access-82grg\") pod \"route-controller-manager-7c4d96c475-dgppt\" (UID: \"0cba6771-0129-4ae1-b8e0-c13b427939b1\") " pod="openshift-route-controller-manager/route-controller-manager-7c4d96c475-dgppt" Feb 18 09:08:54 crc kubenswrapper[4556]: I0218 09:08:54.910988 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0cba6771-0129-4ae1-b8e0-c13b427939b1-client-ca\") pod \"route-controller-manager-7c4d96c475-dgppt\" (UID: \"0cba6771-0129-4ae1-b8e0-c13b427939b1\") " pod="openshift-route-controller-manager/route-controller-manager-7c4d96c475-dgppt" Feb 18 09:08:54 crc kubenswrapper[4556]: I0218 09:08:54.911023 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0cba6771-0129-4ae1-b8e0-c13b427939b1-serving-cert\") pod \"route-controller-manager-7c4d96c475-dgppt\" (UID: \"0cba6771-0129-4ae1-b8e0-c13b427939b1\") " pod="openshift-route-controller-manager/route-controller-manager-7c4d96c475-dgppt" Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.012265 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/97ee9f2e-ab2e-4436-acdb-12c892f54209-serving-cert\") pod \"97ee9f2e-ab2e-4436-acdb-12c892f54209\" (UID: \"97ee9f2e-ab2e-4436-acdb-12c892f54209\") " Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.012319 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bjmvg\" (UniqueName: \"kubernetes.io/projected/a9801216-0000-4dac-b960-afd2a02f9228-kube-api-access-bjmvg\") pod \"a9801216-0000-4dac-b960-afd2a02f9228\" (UID: \"a9801216-0000-4dac-b960-afd2a02f9228\") " Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.012341 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a9801216-0000-4dac-b960-afd2a02f9228-client-ca\") pod \"a9801216-0000-4dac-b960-afd2a02f9228\" (UID: \"a9801216-0000-4dac-b960-afd2a02f9228\") " Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.012360 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2lh6v\" (UniqueName: \"kubernetes.io/projected/97ee9f2e-ab2e-4436-acdb-12c892f54209-kube-api-access-2lh6v\") pod \"97ee9f2e-ab2e-4436-acdb-12c892f54209\" (UID: \"97ee9f2e-ab2e-4436-acdb-12c892f54209\") " Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.012388 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a9801216-0000-4dac-b960-afd2a02f9228-serving-cert\") pod \"a9801216-0000-4dac-b960-afd2a02f9228\" (UID: \"a9801216-0000-4dac-b960-afd2a02f9228\") " Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.012458 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9801216-0000-4dac-b960-afd2a02f9228-config\") pod \"a9801216-0000-4dac-b960-afd2a02f9228\" (UID: \"a9801216-0000-4dac-b960-afd2a02f9228\") " Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.012483 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97ee9f2e-ab2e-4436-acdb-12c892f54209-config\") pod \"97ee9f2e-ab2e-4436-acdb-12c892f54209\" (UID: \"97ee9f2e-ab2e-4436-acdb-12c892f54209\") " Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.012605 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/97ee9f2e-ab2e-4436-acdb-12c892f54209-proxy-ca-bundles\") pod \"97ee9f2e-ab2e-4436-acdb-12c892f54209\" (UID: \"97ee9f2e-ab2e-4436-acdb-12c892f54209\") " Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.013261 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97ee9f2e-ab2e-4436-acdb-12c892f54209-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "97ee9f2e-ab2e-4436-acdb-12c892f54209" (UID: "97ee9f2e-ab2e-4436-acdb-12c892f54209"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.013287 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97ee9f2e-ab2e-4436-acdb-12c892f54209-config" (OuterVolumeSpecName: "config") pod "97ee9f2e-ab2e-4436-acdb-12c892f54209" (UID: "97ee9f2e-ab2e-4436-acdb-12c892f54209"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.013323 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9801216-0000-4dac-b960-afd2a02f9228-config" (OuterVolumeSpecName: "config") pod "a9801216-0000-4dac-b960-afd2a02f9228" (UID: "a9801216-0000-4dac-b960-afd2a02f9228"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.013315 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/97ee9f2e-ab2e-4436-acdb-12c892f54209-client-ca\") pod \"97ee9f2e-ab2e-4436-acdb-12c892f54209\" (UID: \"97ee9f2e-ab2e-4436-acdb-12c892f54209\") " Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.013355 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9801216-0000-4dac-b960-afd2a02f9228-client-ca" (OuterVolumeSpecName: "client-ca") pod "a9801216-0000-4dac-b960-afd2a02f9228" (UID: "a9801216-0000-4dac-b960-afd2a02f9228"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.013472 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82grg\" (UniqueName: \"kubernetes.io/projected/0cba6771-0129-4ae1-b8e0-c13b427939b1-kube-api-access-82grg\") pod \"route-controller-manager-7c4d96c475-dgppt\" (UID: \"0cba6771-0129-4ae1-b8e0-c13b427939b1\") " pod="openshift-route-controller-manager/route-controller-manager-7c4d96c475-dgppt" Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.013505 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0cba6771-0129-4ae1-b8e0-c13b427939b1-client-ca\") pod \"route-controller-manager-7c4d96c475-dgppt\" (UID: \"0cba6771-0129-4ae1-b8e0-c13b427939b1\") " pod="openshift-route-controller-manager/route-controller-manager-7c4d96c475-dgppt" Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.013534 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0cba6771-0129-4ae1-b8e0-c13b427939b1-serving-cert\") pod \"route-controller-manager-7c4d96c475-dgppt\" (UID: \"0cba6771-0129-4ae1-b8e0-c13b427939b1\") " pod="openshift-route-controller-manager/route-controller-manager-7c4d96c475-dgppt" Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.013580 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0cba6771-0129-4ae1-b8e0-c13b427939b1-config\") pod \"route-controller-manager-7c4d96c475-dgppt\" (UID: \"0cba6771-0129-4ae1-b8e0-c13b427939b1\") " pod="openshift-route-controller-manager/route-controller-manager-7c4d96c475-dgppt" Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.013643 4556 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a9801216-0000-4dac-b960-afd2a02f9228-client-ca\") on node \"crc\" DevicePath \"\"" Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.013650 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97ee9f2e-ab2e-4436-acdb-12c892f54209-client-ca" (OuterVolumeSpecName: "client-ca") pod "97ee9f2e-ab2e-4436-acdb-12c892f54209" (UID: "97ee9f2e-ab2e-4436-acdb-12c892f54209"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.013658 4556 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9801216-0000-4dac-b960-afd2a02f9228-config\") on node \"crc\" DevicePath \"\"" Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.013694 4556 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97ee9f2e-ab2e-4436-acdb-12c892f54209-config\") on node \"crc\" DevicePath \"\"" Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.013704 4556 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/97ee9f2e-ab2e-4436-acdb-12c892f54209-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.014665 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0cba6771-0129-4ae1-b8e0-c13b427939b1-client-ca\") pod \"route-controller-manager-7c4d96c475-dgppt\" (UID: \"0cba6771-0129-4ae1-b8e0-c13b427939b1\") " pod="openshift-route-controller-manager/route-controller-manager-7c4d96c475-dgppt" Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.014755 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0cba6771-0129-4ae1-b8e0-c13b427939b1-config\") pod \"route-controller-manager-7c4d96c475-dgppt\" (UID: \"0cba6771-0129-4ae1-b8e0-c13b427939b1\") " pod="openshift-route-controller-manager/route-controller-manager-7c4d96c475-dgppt" Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.017607 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97ee9f2e-ab2e-4436-acdb-12c892f54209-kube-api-access-2lh6v" (OuterVolumeSpecName: "kube-api-access-2lh6v") pod "97ee9f2e-ab2e-4436-acdb-12c892f54209" (UID: "97ee9f2e-ab2e-4436-acdb-12c892f54209"). InnerVolumeSpecName "kube-api-access-2lh6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.017612 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9801216-0000-4dac-b960-afd2a02f9228-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "a9801216-0000-4dac-b960-afd2a02f9228" (UID: "a9801216-0000-4dac-b960-afd2a02f9228"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.017630 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97ee9f2e-ab2e-4436-acdb-12c892f54209-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "97ee9f2e-ab2e-4436-acdb-12c892f54209" (UID: "97ee9f2e-ab2e-4436-acdb-12c892f54209"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.017698 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9801216-0000-4dac-b960-afd2a02f9228-kube-api-access-bjmvg" (OuterVolumeSpecName: "kube-api-access-bjmvg") pod "a9801216-0000-4dac-b960-afd2a02f9228" (UID: "a9801216-0000-4dac-b960-afd2a02f9228"). InnerVolumeSpecName "kube-api-access-bjmvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.018069 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0cba6771-0129-4ae1-b8e0-c13b427939b1-serving-cert\") pod \"route-controller-manager-7c4d96c475-dgppt\" (UID: \"0cba6771-0129-4ae1-b8e0-c13b427939b1\") " pod="openshift-route-controller-manager/route-controller-manager-7c4d96c475-dgppt" Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.026344 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82grg\" (UniqueName: \"kubernetes.io/projected/0cba6771-0129-4ae1-b8e0-c13b427939b1-kube-api-access-82grg\") pod \"route-controller-manager-7c4d96c475-dgppt\" (UID: \"0cba6771-0129-4ae1-b8e0-c13b427939b1\") " pod="openshift-route-controller-manager/route-controller-manager-7c4d96c475-dgppt" Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.114763 4556 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/97ee9f2e-ab2e-4436-acdb-12c892f54209-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.114790 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bjmvg\" (UniqueName: \"kubernetes.io/projected/a9801216-0000-4dac-b960-afd2a02f9228-kube-api-access-bjmvg\") on node \"crc\" DevicePath \"\"" Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.114801 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2lh6v\" (UniqueName: \"kubernetes.io/projected/97ee9f2e-ab2e-4436-acdb-12c892f54209-kube-api-access-2lh6v\") on node \"crc\" DevicePath \"\"" Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.114810 4556 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a9801216-0000-4dac-b960-afd2a02f9228-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.114819 4556 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/97ee9f2e-ab2e-4436-acdb-12c892f54209-client-ca\") on node \"crc\" DevicePath \"\"" Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.173386 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7c4d96c475-dgppt" Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.297922 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7c4d96c475-dgppt"] Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.625634 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6fcf8b4c5d-bs84s" event={"ID":"a9801216-0000-4dac-b960-afd2a02f9228","Type":"ContainerDied","Data":"0ac8dbf9ffaee203a21707591f6c43941935c90f32f00edf337005abc512b086"} Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.625865 4556 scope.go:117] "RemoveContainer" containerID="5c9ccd41034a49596e7881346348a229ccb590261cd58854a0f8d2391849f45e" Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.625648 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6fcf8b4c5d-bs84s" Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.627130 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7c4d96c475-dgppt" event={"ID":"0cba6771-0129-4ae1-b8e0-c13b427939b1","Type":"ContainerStarted","Data":"b20fb33b326fe5a0f4dc3819bdaa9e6ecd4abdfb9921328997860afe284d9088"} Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.627175 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7c4d96c475-dgppt" event={"ID":"0cba6771-0129-4ae1-b8e0-c13b427939b1","Type":"ContainerStarted","Data":"e6fe62ea209df75edd3b3a20e869dbe7a73235fc374b7c5e7b2e02ea59eedb7a"} Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.627372 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7c4d96c475-dgppt" Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.628704 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7d6d847d76-dktfz" event={"ID":"97ee9f2e-ab2e-4436-acdb-12c892f54209","Type":"ContainerDied","Data":"19d242311525621a371fa73293c502b333603b760142c8c3d4064f3561f99335"} Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.628728 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7d6d847d76-dktfz" Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.633531 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7c4d96c475-dgppt" Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.640196 4556 scope.go:117] "RemoveContainer" containerID="abfca06cddcc30451a0ea1c6ff122cfe5056858630c11acb1f76d28a3f713e25" Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.640854 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7c4d96c475-dgppt" podStartSLOduration=1.640844577 podStartE2EDuration="1.640844577s" podCreationTimestamp="2026-02-18 09:08:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:08:55.639268707 +0000 UTC m=+292.656229688" watchObservedRunningTime="2026-02-18 09:08:55.640844577 +0000 UTC m=+292.657805556" Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.653111 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6fcf8b4c5d-bs84s"] Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.656694 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6fcf8b4c5d-bs84s"] Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.668563 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7d6d847d76-dktfz"] Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.673228 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-7d6d847d76-dktfz"] Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.977698 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-754b797845-2t4bs"] Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.978375 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-754b797845-2t4bs" Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.979846 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.979946 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.980052 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.980199 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.980821 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.981937 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.985084 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 18 09:08:55 crc kubenswrapper[4556]: I0218 09:08:55.988116 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-754b797845-2t4bs"] Feb 18 09:08:56 crc kubenswrapper[4556]: I0218 09:08:56.124971 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7cf2e5ac-c19f-4694-af7f-70bc6bcbc720-serving-cert\") pod \"controller-manager-754b797845-2t4bs\" (UID: \"7cf2e5ac-c19f-4694-af7f-70bc6bcbc720\") " pod="openshift-controller-manager/controller-manager-754b797845-2t4bs" Feb 18 09:08:56 crc kubenswrapper[4556]: I0218 09:08:56.125006 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7cf2e5ac-c19f-4694-af7f-70bc6bcbc720-proxy-ca-bundles\") pod \"controller-manager-754b797845-2t4bs\" (UID: \"7cf2e5ac-c19f-4694-af7f-70bc6bcbc720\") " pod="openshift-controller-manager/controller-manager-754b797845-2t4bs" Feb 18 09:08:56 crc kubenswrapper[4556]: I0218 09:08:56.125040 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cf2e5ac-c19f-4694-af7f-70bc6bcbc720-config\") pod \"controller-manager-754b797845-2t4bs\" (UID: \"7cf2e5ac-c19f-4694-af7f-70bc6bcbc720\") " pod="openshift-controller-manager/controller-manager-754b797845-2t4bs" Feb 18 09:08:56 crc kubenswrapper[4556]: I0218 09:08:56.125256 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4ft6\" (UniqueName: \"kubernetes.io/projected/7cf2e5ac-c19f-4694-af7f-70bc6bcbc720-kube-api-access-s4ft6\") pod \"controller-manager-754b797845-2t4bs\" (UID: \"7cf2e5ac-c19f-4694-af7f-70bc6bcbc720\") " pod="openshift-controller-manager/controller-manager-754b797845-2t4bs" Feb 18 09:08:56 crc kubenswrapper[4556]: I0218 09:08:56.125307 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7cf2e5ac-c19f-4694-af7f-70bc6bcbc720-client-ca\") pod \"controller-manager-754b797845-2t4bs\" (UID: \"7cf2e5ac-c19f-4694-af7f-70bc6bcbc720\") " pod="openshift-controller-manager/controller-manager-754b797845-2t4bs" Feb 18 09:08:56 crc kubenswrapper[4556]: I0218 09:08:56.225924 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4ft6\" (UniqueName: \"kubernetes.io/projected/7cf2e5ac-c19f-4694-af7f-70bc6bcbc720-kube-api-access-s4ft6\") pod \"controller-manager-754b797845-2t4bs\" (UID: \"7cf2e5ac-c19f-4694-af7f-70bc6bcbc720\") " pod="openshift-controller-manager/controller-manager-754b797845-2t4bs" Feb 18 09:08:56 crc kubenswrapper[4556]: I0218 09:08:56.225971 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7cf2e5ac-c19f-4694-af7f-70bc6bcbc720-client-ca\") pod \"controller-manager-754b797845-2t4bs\" (UID: \"7cf2e5ac-c19f-4694-af7f-70bc6bcbc720\") " pod="openshift-controller-manager/controller-manager-754b797845-2t4bs" Feb 18 09:08:56 crc kubenswrapper[4556]: I0218 09:08:56.226052 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7cf2e5ac-c19f-4694-af7f-70bc6bcbc720-serving-cert\") pod \"controller-manager-754b797845-2t4bs\" (UID: \"7cf2e5ac-c19f-4694-af7f-70bc6bcbc720\") " pod="openshift-controller-manager/controller-manager-754b797845-2t4bs" Feb 18 09:08:56 crc kubenswrapper[4556]: I0218 09:08:56.226067 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7cf2e5ac-c19f-4694-af7f-70bc6bcbc720-proxy-ca-bundles\") pod \"controller-manager-754b797845-2t4bs\" (UID: \"7cf2e5ac-c19f-4694-af7f-70bc6bcbc720\") " pod="openshift-controller-manager/controller-manager-754b797845-2t4bs" Feb 18 09:08:56 crc kubenswrapper[4556]: I0218 09:08:56.226109 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cf2e5ac-c19f-4694-af7f-70bc6bcbc720-config\") pod \"controller-manager-754b797845-2t4bs\" (UID: \"7cf2e5ac-c19f-4694-af7f-70bc6bcbc720\") " pod="openshift-controller-manager/controller-manager-754b797845-2t4bs" Feb 18 09:08:56 crc kubenswrapper[4556]: I0218 09:08:56.226897 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7cf2e5ac-c19f-4694-af7f-70bc6bcbc720-client-ca\") pod \"controller-manager-754b797845-2t4bs\" (UID: \"7cf2e5ac-c19f-4694-af7f-70bc6bcbc720\") " pod="openshift-controller-manager/controller-manager-754b797845-2t4bs" Feb 18 09:08:56 crc kubenswrapper[4556]: I0218 09:08:56.226981 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7cf2e5ac-c19f-4694-af7f-70bc6bcbc720-proxy-ca-bundles\") pod \"controller-manager-754b797845-2t4bs\" (UID: \"7cf2e5ac-c19f-4694-af7f-70bc6bcbc720\") " pod="openshift-controller-manager/controller-manager-754b797845-2t4bs" Feb 18 09:08:56 crc kubenswrapper[4556]: I0218 09:08:56.227337 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cf2e5ac-c19f-4694-af7f-70bc6bcbc720-config\") pod \"controller-manager-754b797845-2t4bs\" (UID: \"7cf2e5ac-c19f-4694-af7f-70bc6bcbc720\") " pod="openshift-controller-manager/controller-manager-754b797845-2t4bs" Feb 18 09:08:56 crc kubenswrapper[4556]: I0218 09:08:56.229880 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7cf2e5ac-c19f-4694-af7f-70bc6bcbc720-serving-cert\") pod \"controller-manager-754b797845-2t4bs\" (UID: \"7cf2e5ac-c19f-4694-af7f-70bc6bcbc720\") " pod="openshift-controller-manager/controller-manager-754b797845-2t4bs" Feb 18 09:08:56 crc kubenswrapper[4556]: I0218 09:08:56.238025 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4ft6\" (UniqueName: \"kubernetes.io/projected/7cf2e5ac-c19f-4694-af7f-70bc6bcbc720-kube-api-access-s4ft6\") pod \"controller-manager-754b797845-2t4bs\" (UID: \"7cf2e5ac-c19f-4694-af7f-70bc6bcbc720\") " pod="openshift-controller-manager/controller-manager-754b797845-2t4bs" Feb 18 09:08:56 crc kubenswrapper[4556]: I0218 09:08:56.289751 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-754b797845-2t4bs" Feb 18 09:08:56 crc kubenswrapper[4556]: I0218 09:08:56.613461 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-754b797845-2t4bs"] Feb 18 09:08:56 crc kubenswrapper[4556]: W0218 09:08:56.616801 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7cf2e5ac_c19f_4694_af7f_70bc6bcbc720.slice/crio-7f3370e4ea8bbef9353573382e866188b7e6602a23f1c15c97d8633b6ddd6281 WatchSource:0}: Error finding container 7f3370e4ea8bbef9353573382e866188b7e6602a23f1c15c97d8633b6ddd6281: Status 404 returned error can't find the container with id 7f3370e4ea8bbef9353573382e866188b7e6602a23f1c15c97d8633b6ddd6281 Feb 18 09:08:56 crc kubenswrapper[4556]: I0218 09:08:56.633387 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-754b797845-2t4bs" event={"ID":"7cf2e5ac-c19f-4694-af7f-70bc6bcbc720","Type":"ContainerStarted","Data":"7f3370e4ea8bbef9353573382e866188b7e6602a23f1c15c97d8633b6ddd6281"} Feb 18 09:08:57 crc kubenswrapper[4556]: I0218 09:08:57.286930 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97ee9f2e-ab2e-4436-acdb-12c892f54209" path="/var/lib/kubelet/pods/97ee9f2e-ab2e-4436-acdb-12c892f54209/volumes" Feb 18 09:08:57 crc kubenswrapper[4556]: I0218 09:08:57.287727 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9801216-0000-4dac-b960-afd2a02f9228" path="/var/lib/kubelet/pods/a9801216-0000-4dac-b960-afd2a02f9228/volumes" Feb 18 09:08:57 crc kubenswrapper[4556]: I0218 09:08:57.641300 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-754b797845-2t4bs" event={"ID":"7cf2e5ac-c19f-4694-af7f-70bc6bcbc720","Type":"ContainerStarted","Data":"377564960a52e642e12544eef0a23dbe3d4907b01520dbbeec8f5e8fe13ef34b"} Feb 18 09:08:57 crc kubenswrapper[4556]: I0218 09:08:57.653317 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-754b797845-2t4bs" podStartSLOduration=3.653301993 podStartE2EDuration="3.653301993s" podCreationTimestamp="2026-02-18 09:08:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:08:57.651794635 +0000 UTC m=+294.668755615" watchObservedRunningTime="2026-02-18 09:08:57.653301993 +0000 UTC m=+294.670262974" Feb 18 09:08:58 crc kubenswrapper[4556]: I0218 09:08:58.644809 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-754b797845-2t4bs" Feb 18 09:08:58 crc kubenswrapper[4556]: I0218 09:08:58.649767 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-754b797845-2t4bs" Feb 18 09:09:03 crc kubenswrapper[4556]: I0218 09:09:03.183822 4556 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Feb 18 09:09:18 crc kubenswrapper[4556]: I0218 09:09:18.735466 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-6ldbx"] Feb 18 09:09:18 crc kubenswrapper[4556]: I0218 09:09:18.736985 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-6ldbx" Feb 18 09:09:18 crc kubenswrapper[4556]: I0218 09:09:18.745358 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-6ldbx"] Feb 18 09:09:18 crc kubenswrapper[4556]: I0218 09:09:18.904766 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/200ae95d-c17e-4b50-ba27-5f815b88e509-registry-tls\") pod \"image-registry-66df7c8f76-6ldbx\" (UID: \"200ae95d-c17e-4b50-ba27-5f815b88e509\") " pod="openshift-image-registry/image-registry-66df7c8f76-6ldbx" Feb 18 09:09:18 crc kubenswrapper[4556]: I0218 09:09:18.904830 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-6ldbx\" (UID: \"200ae95d-c17e-4b50-ba27-5f815b88e509\") " pod="openshift-image-registry/image-registry-66df7c8f76-6ldbx" Feb 18 09:09:18 crc kubenswrapper[4556]: I0218 09:09:18.904901 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/200ae95d-c17e-4b50-ba27-5f815b88e509-trusted-ca\") pod \"image-registry-66df7c8f76-6ldbx\" (UID: \"200ae95d-c17e-4b50-ba27-5f815b88e509\") " pod="openshift-image-registry/image-registry-66df7c8f76-6ldbx" Feb 18 09:09:18 crc kubenswrapper[4556]: I0218 09:09:18.904926 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/200ae95d-c17e-4b50-ba27-5f815b88e509-installation-pull-secrets\") pod \"image-registry-66df7c8f76-6ldbx\" (UID: \"200ae95d-c17e-4b50-ba27-5f815b88e509\") " pod="openshift-image-registry/image-registry-66df7c8f76-6ldbx" Feb 18 09:09:18 crc kubenswrapper[4556]: I0218 09:09:18.904945 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/200ae95d-c17e-4b50-ba27-5f815b88e509-registry-certificates\") pod \"image-registry-66df7c8f76-6ldbx\" (UID: \"200ae95d-c17e-4b50-ba27-5f815b88e509\") " pod="openshift-image-registry/image-registry-66df7c8f76-6ldbx" Feb 18 09:09:18 crc kubenswrapper[4556]: I0218 09:09:18.904967 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhh64\" (UniqueName: \"kubernetes.io/projected/200ae95d-c17e-4b50-ba27-5f815b88e509-kube-api-access-rhh64\") pod \"image-registry-66df7c8f76-6ldbx\" (UID: \"200ae95d-c17e-4b50-ba27-5f815b88e509\") " pod="openshift-image-registry/image-registry-66df7c8f76-6ldbx" Feb 18 09:09:18 crc kubenswrapper[4556]: I0218 09:09:18.905223 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/200ae95d-c17e-4b50-ba27-5f815b88e509-bound-sa-token\") pod \"image-registry-66df7c8f76-6ldbx\" (UID: \"200ae95d-c17e-4b50-ba27-5f815b88e509\") " pod="openshift-image-registry/image-registry-66df7c8f76-6ldbx" Feb 18 09:09:18 crc kubenswrapper[4556]: I0218 09:09:18.905441 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/200ae95d-c17e-4b50-ba27-5f815b88e509-ca-trust-extracted\") pod \"image-registry-66df7c8f76-6ldbx\" (UID: \"200ae95d-c17e-4b50-ba27-5f815b88e509\") " pod="openshift-image-registry/image-registry-66df7c8f76-6ldbx" Feb 18 09:09:18 crc kubenswrapper[4556]: I0218 09:09:18.920775 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-6ldbx\" (UID: \"200ae95d-c17e-4b50-ba27-5f815b88e509\") " pod="openshift-image-registry/image-registry-66df7c8f76-6ldbx" Feb 18 09:09:19 crc kubenswrapper[4556]: I0218 09:09:19.007079 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/200ae95d-c17e-4b50-ba27-5f815b88e509-ca-trust-extracted\") pod \"image-registry-66df7c8f76-6ldbx\" (UID: \"200ae95d-c17e-4b50-ba27-5f815b88e509\") " pod="openshift-image-registry/image-registry-66df7c8f76-6ldbx" Feb 18 09:09:19 crc kubenswrapper[4556]: I0218 09:09:19.007124 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/200ae95d-c17e-4b50-ba27-5f815b88e509-registry-tls\") pod \"image-registry-66df7c8f76-6ldbx\" (UID: \"200ae95d-c17e-4b50-ba27-5f815b88e509\") " pod="openshift-image-registry/image-registry-66df7c8f76-6ldbx" Feb 18 09:09:19 crc kubenswrapper[4556]: I0218 09:09:19.007206 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/200ae95d-c17e-4b50-ba27-5f815b88e509-trusted-ca\") pod \"image-registry-66df7c8f76-6ldbx\" (UID: \"200ae95d-c17e-4b50-ba27-5f815b88e509\") " pod="openshift-image-registry/image-registry-66df7c8f76-6ldbx" Feb 18 09:09:19 crc kubenswrapper[4556]: I0218 09:09:19.007245 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/200ae95d-c17e-4b50-ba27-5f815b88e509-installation-pull-secrets\") pod \"image-registry-66df7c8f76-6ldbx\" (UID: \"200ae95d-c17e-4b50-ba27-5f815b88e509\") " pod="openshift-image-registry/image-registry-66df7c8f76-6ldbx" Feb 18 09:09:19 crc kubenswrapper[4556]: I0218 09:09:19.007268 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/200ae95d-c17e-4b50-ba27-5f815b88e509-registry-certificates\") pod \"image-registry-66df7c8f76-6ldbx\" (UID: \"200ae95d-c17e-4b50-ba27-5f815b88e509\") " pod="openshift-image-registry/image-registry-66df7c8f76-6ldbx" Feb 18 09:09:19 crc kubenswrapper[4556]: I0218 09:09:19.007291 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhh64\" (UniqueName: \"kubernetes.io/projected/200ae95d-c17e-4b50-ba27-5f815b88e509-kube-api-access-rhh64\") pod \"image-registry-66df7c8f76-6ldbx\" (UID: \"200ae95d-c17e-4b50-ba27-5f815b88e509\") " pod="openshift-image-registry/image-registry-66df7c8f76-6ldbx" Feb 18 09:09:19 crc kubenswrapper[4556]: I0218 09:09:19.007326 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/200ae95d-c17e-4b50-ba27-5f815b88e509-bound-sa-token\") pod \"image-registry-66df7c8f76-6ldbx\" (UID: \"200ae95d-c17e-4b50-ba27-5f815b88e509\") " pod="openshift-image-registry/image-registry-66df7c8f76-6ldbx" Feb 18 09:09:19 crc kubenswrapper[4556]: I0218 09:09:19.007572 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/200ae95d-c17e-4b50-ba27-5f815b88e509-ca-trust-extracted\") pod \"image-registry-66df7c8f76-6ldbx\" (UID: \"200ae95d-c17e-4b50-ba27-5f815b88e509\") " pod="openshift-image-registry/image-registry-66df7c8f76-6ldbx" Feb 18 09:09:19 crc kubenswrapper[4556]: I0218 09:09:19.008391 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/200ae95d-c17e-4b50-ba27-5f815b88e509-trusted-ca\") pod \"image-registry-66df7c8f76-6ldbx\" (UID: \"200ae95d-c17e-4b50-ba27-5f815b88e509\") " pod="openshift-image-registry/image-registry-66df7c8f76-6ldbx" Feb 18 09:09:19 crc kubenswrapper[4556]: I0218 09:09:19.008422 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/200ae95d-c17e-4b50-ba27-5f815b88e509-registry-certificates\") pod \"image-registry-66df7c8f76-6ldbx\" (UID: \"200ae95d-c17e-4b50-ba27-5f815b88e509\") " pod="openshift-image-registry/image-registry-66df7c8f76-6ldbx" Feb 18 09:09:19 crc kubenswrapper[4556]: I0218 09:09:19.013004 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/200ae95d-c17e-4b50-ba27-5f815b88e509-registry-tls\") pod \"image-registry-66df7c8f76-6ldbx\" (UID: \"200ae95d-c17e-4b50-ba27-5f815b88e509\") " pod="openshift-image-registry/image-registry-66df7c8f76-6ldbx" Feb 18 09:09:19 crc kubenswrapper[4556]: I0218 09:09:19.013166 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/200ae95d-c17e-4b50-ba27-5f815b88e509-installation-pull-secrets\") pod \"image-registry-66df7c8f76-6ldbx\" (UID: \"200ae95d-c17e-4b50-ba27-5f815b88e509\") " pod="openshift-image-registry/image-registry-66df7c8f76-6ldbx" Feb 18 09:09:19 crc kubenswrapper[4556]: I0218 09:09:19.020976 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/200ae95d-c17e-4b50-ba27-5f815b88e509-bound-sa-token\") pod \"image-registry-66df7c8f76-6ldbx\" (UID: \"200ae95d-c17e-4b50-ba27-5f815b88e509\") " pod="openshift-image-registry/image-registry-66df7c8f76-6ldbx" Feb 18 09:09:19 crc kubenswrapper[4556]: I0218 09:09:19.021106 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhh64\" (UniqueName: \"kubernetes.io/projected/200ae95d-c17e-4b50-ba27-5f815b88e509-kube-api-access-rhh64\") pod \"image-registry-66df7c8f76-6ldbx\" (UID: \"200ae95d-c17e-4b50-ba27-5f815b88e509\") " pod="openshift-image-registry/image-registry-66df7c8f76-6ldbx" Feb 18 09:09:19 crc kubenswrapper[4556]: I0218 09:09:19.053032 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-6ldbx" Feb 18 09:09:19 crc kubenswrapper[4556]: I0218 09:09:19.402278 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-6ldbx"] Feb 18 09:09:19 crc kubenswrapper[4556]: W0218 09:09:19.407472 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod200ae95d_c17e_4b50_ba27_5f815b88e509.slice/crio-d9ce20665d0194f61b749a284d6b6098bd352752574063308223c3ad164c9413 WatchSource:0}: Error finding container d9ce20665d0194f61b749a284d6b6098bd352752574063308223c3ad164c9413: Status 404 returned error can't find the container with id d9ce20665d0194f61b749a284d6b6098bd352752574063308223c3ad164c9413 Feb 18 09:09:19 crc kubenswrapper[4556]: I0218 09:09:19.734584 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-6ldbx" event={"ID":"200ae95d-c17e-4b50-ba27-5f815b88e509","Type":"ContainerStarted","Data":"ee477995ce90d811638cfea45183a4e3d8fde53140ac84778ec9fa9aa63749c3"} Feb 18 09:09:19 crc kubenswrapper[4556]: I0218 09:09:19.734632 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-6ldbx" event={"ID":"200ae95d-c17e-4b50-ba27-5f815b88e509","Type":"ContainerStarted","Data":"d9ce20665d0194f61b749a284d6b6098bd352752574063308223c3ad164c9413"} Feb 18 09:09:19 crc kubenswrapper[4556]: I0218 09:09:19.734732 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-6ldbx" Feb 18 09:09:19 crc kubenswrapper[4556]: I0218 09:09:19.750591 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-6ldbx" podStartSLOduration=1.750575867 podStartE2EDuration="1.750575867s" podCreationTimestamp="2026-02-18 09:09:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:09:19.747934555 +0000 UTC m=+316.764895536" watchObservedRunningTime="2026-02-18 09:09:19.750575867 +0000 UTC m=+316.767536847" Feb 18 09:09:24 crc kubenswrapper[4556]: I0218 09:09:24.461968 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-754b797845-2t4bs"] Feb 18 09:09:24 crc kubenswrapper[4556]: I0218 09:09:24.462476 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-754b797845-2t4bs" podUID="7cf2e5ac-c19f-4694-af7f-70bc6bcbc720" containerName="controller-manager" containerID="cri-o://377564960a52e642e12544eef0a23dbe3d4907b01520dbbeec8f5e8fe13ef34b" gracePeriod=30 Feb 18 09:09:24 crc kubenswrapper[4556]: I0218 09:09:24.471135 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7c4d96c475-dgppt"] Feb 18 09:09:24 crc kubenswrapper[4556]: I0218 09:09:24.471320 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-7c4d96c475-dgppt" podUID="0cba6771-0129-4ae1-b8e0-c13b427939b1" containerName="route-controller-manager" containerID="cri-o://b20fb33b326fe5a0f4dc3819bdaa9e6ecd4abdfb9921328997860afe284d9088" gracePeriod=30 Feb 18 09:09:24 crc kubenswrapper[4556]: I0218 09:09:24.769405 4556 generic.go:334] "Generic (PLEG): container finished" podID="0cba6771-0129-4ae1-b8e0-c13b427939b1" containerID="b20fb33b326fe5a0f4dc3819bdaa9e6ecd4abdfb9921328997860afe284d9088" exitCode=0 Feb 18 09:09:24 crc kubenswrapper[4556]: I0218 09:09:24.769567 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7c4d96c475-dgppt" event={"ID":"0cba6771-0129-4ae1-b8e0-c13b427939b1","Type":"ContainerDied","Data":"b20fb33b326fe5a0f4dc3819bdaa9e6ecd4abdfb9921328997860afe284d9088"} Feb 18 09:09:24 crc kubenswrapper[4556]: I0218 09:09:24.771481 4556 generic.go:334] "Generic (PLEG): container finished" podID="7cf2e5ac-c19f-4694-af7f-70bc6bcbc720" containerID="377564960a52e642e12544eef0a23dbe3d4907b01520dbbeec8f5e8fe13ef34b" exitCode=0 Feb 18 09:09:24 crc kubenswrapper[4556]: I0218 09:09:24.771541 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-754b797845-2t4bs" event={"ID":"7cf2e5ac-c19f-4694-af7f-70bc6bcbc720","Type":"ContainerDied","Data":"377564960a52e642e12544eef0a23dbe3d4907b01520dbbeec8f5e8fe13ef34b"} Feb 18 09:09:24 crc kubenswrapper[4556]: I0218 09:09:24.893186 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7c4d96c475-dgppt" Feb 18 09:09:24 crc kubenswrapper[4556]: I0218 09:09:24.967659 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-754b797845-2t4bs" Feb 18 09:09:25 crc kubenswrapper[4556]: I0218 09:09:25.084185 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0cba6771-0129-4ae1-b8e0-c13b427939b1-config\") pod \"0cba6771-0129-4ae1-b8e0-c13b427939b1\" (UID: \"0cba6771-0129-4ae1-b8e0-c13b427939b1\") " Feb 18 09:09:25 crc kubenswrapper[4556]: I0218 09:09:25.084240 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-82grg\" (UniqueName: \"kubernetes.io/projected/0cba6771-0129-4ae1-b8e0-c13b427939b1-kube-api-access-82grg\") pod \"0cba6771-0129-4ae1-b8e0-c13b427939b1\" (UID: \"0cba6771-0129-4ae1-b8e0-c13b427939b1\") " Feb 18 09:09:25 crc kubenswrapper[4556]: I0218 09:09:25.084264 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7cf2e5ac-c19f-4694-af7f-70bc6bcbc720-client-ca\") pod \"7cf2e5ac-c19f-4694-af7f-70bc6bcbc720\" (UID: \"7cf2e5ac-c19f-4694-af7f-70bc6bcbc720\") " Feb 18 09:09:25 crc kubenswrapper[4556]: I0218 09:09:25.084292 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7cf2e5ac-c19f-4694-af7f-70bc6bcbc720-serving-cert\") pod \"7cf2e5ac-c19f-4694-af7f-70bc6bcbc720\" (UID: \"7cf2e5ac-c19f-4694-af7f-70bc6bcbc720\") " Feb 18 09:09:25 crc kubenswrapper[4556]: I0218 09:09:25.084320 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cf2e5ac-c19f-4694-af7f-70bc6bcbc720-config\") pod \"7cf2e5ac-c19f-4694-af7f-70bc6bcbc720\" (UID: \"7cf2e5ac-c19f-4694-af7f-70bc6bcbc720\") " Feb 18 09:09:25 crc kubenswrapper[4556]: I0218 09:09:25.084990 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7cf2e5ac-c19f-4694-af7f-70bc6bcbc720-client-ca" (OuterVolumeSpecName: "client-ca") pod "7cf2e5ac-c19f-4694-af7f-70bc6bcbc720" (UID: "7cf2e5ac-c19f-4694-af7f-70bc6bcbc720"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:09:25 crc kubenswrapper[4556]: I0218 09:09:25.085008 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4ft6\" (UniqueName: \"kubernetes.io/projected/7cf2e5ac-c19f-4694-af7f-70bc6bcbc720-kube-api-access-s4ft6\") pod \"7cf2e5ac-c19f-4694-af7f-70bc6bcbc720\" (UID: \"7cf2e5ac-c19f-4694-af7f-70bc6bcbc720\") " Feb 18 09:09:25 crc kubenswrapper[4556]: I0218 09:09:25.085323 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0cba6771-0129-4ae1-b8e0-c13b427939b1-serving-cert\") pod \"0cba6771-0129-4ae1-b8e0-c13b427939b1\" (UID: \"0cba6771-0129-4ae1-b8e0-c13b427939b1\") " Feb 18 09:09:25 crc kubenswrapper[4556]: I0218 09:09:25.085051 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0cba6771-0129-4ae1-b8e0-c13b427939b1-config" (OuterVolumeSpecName: "config") pod "0cba6771-0129-4ae1-b8e0-c13b427939b1" (UID: "0cba6771-0129-4ae1-b8e0-c13b427939b1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:09:25 crc kubenswrapper[4556]: I0218 09:09:25.085092 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7cf2e5ac-c19f-4694-af7f-70bc6bcbc720-config" (OuterVolumeSpecName: "config") pod "7cf2e5ac-c19f-4694-af7f-70bc6bcbc720" (UID: "7cf2e5ac-c19f-4694-af7f-70bc6bcbc720"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:09:25 crc kubenswrapper[4556]: I0218 09:09:25.085420 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7cf2e5ac-c19f-4694-af7f-70bc6bcbc720-proxy-ca-bundles\") pod \"7cf2e5ac-c19f-4694-af7f-70bc6bcbc720\" (UID: \"7cf2e5ac-c19f-4694-af7f-70bc6bcbc720\") " Feb 18 09:09:25 crc kubenswrapper[4556]: I0218 09:09:25.085526 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0cba6771-0129-4ae1-b8e0-c13b427939b1-client-ca\") pod \"0cba6771-0129-4ae1-b8e0-c13b427939b1\" (UID: \"0cba6771-0129-4ae1-b8e0-c13b427939b1\") " Feb 18 09:09:25 crc kubenswrapper[4556]: I0218 09:09:25.085735 4556 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0cba6771-0129-4ae1-b8e0-c13b427939b1-config\") on node \"crc\" DevicePath \"\"" Feb 18 09:09:25 crc kubenswrapper[4556]: I0218 09:09:25.085750 4556 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7cf2e5ac-c19f-4694-af7f-70bc6bcbc720-client-ca\") on node \"crc\" DevicePath \"\"" Feb 18 09:09:25 crc kubenswrapper[4556]: I0218 09:09:25.085759 4556 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cf2e5ac-c19f-4694-af7f-70bc6bcbc720-config\") on node \"crc\" DevicePath \"\"" Feb 18 09:09:25 crc kubenswrapper[4556]: I0218 09:09:25.085909 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7cf2e5ac-c19f-4694-af7f-70bc6bcbc720-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7cf2e5ac-c19f-4694-af7f-70bc6bcbc720" (UID: "7cf2e5ac-c19f-4694-af7f-70bc6bcbc720"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:09:25 crc kubenswrapper[4556]: I0218 09:09:25.085960 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0cba6771-0129-4ae1-b8e0-c13b427939b1-client-ca" (OuterVolumeSpecName: "client-ca") pod "0cba6771-0129-4ae1-b8e0-c13b427939b1" (UID: "0cba6771-0129-4ae1-b8e0-c13b427939b1"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:09:25 crc kubenswrapper[4556]: I0218 09:09:25.089557 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0cba6771-0129-4ae1-b8e0-c13b427939b1-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0cba6771-0129-4ae1-b8e0-c13b427939b1" (UID: "0cba6771-0129-4ae1-b8e0-c13b427939b1"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:09:25 crc kubenswrapper[4556]: I0218 09:09:25.089596 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7cf2e5ac-c19f-4694-af7f-70bc6bcbc720-kube-api-access-s4ft6" (OuterVolumeSpecName: "kube-api-access-s4ft6") pod "7cf2e5ac-c19f-4694-af7f-70bc6bcbc720" (UID: "7cf2e5ac-c19f-4694-af7f-70bc6bcbc720"). InnerVolumeSpecName "kube-api-access-s4ft6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:09:25 crc kubenswrapper[4556]: I0218 09:09:25.089653 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0cba6771-0129-4ae1-b8e0-c13b427939b1-kube-api-access-82grg" (OuterVolumeSpecName: "kube-api-access-82grg") pod "0cba6771-0129-4ae1-b8e0-c13b427939b1" (UID: "0cba6771-0129-4ae1-b8e0-c13b427939b1"). InnerVolumeSpecName "kube-api-access-82grg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:09:25 crc kubenswrapper[4556]: I0218 09:09:25.089717 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7cf2e5ac-c19f-4694-af7f-70bc6bcbc720-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7cf2e5ac-c19f-4694-af7f-70bc6bcbc720" (UID: "7cf2e5ac-c19f-4694-af7f-70bc6bcbc720"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:09:25 crc kubenswrapper[4556]: I0218 09:09:25.187415 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4ft6\" (UniqueName: \"kubernetes.io/projected/7cf2e5ac-c19f-4694-af7f-70bc6bcbc720-kube-api-access-s4ft6\") on node \"crc\" DevicePath \"\"" Feb 18 09:09:25 crc kubenswrapper[4556]: I0218 09:09:25.187506 4556 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0cba6771-0129-4ae1-b8e0-c13b427939b1-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 09:09:25 crc kubenswrapper[4556]: I0218 09:09:25.187560 4556 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7cf2e5ac-c19f-4694-af7f-70bc6bcbc720-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 18 09:09:25 crc kubenswrapper[4556]: I0218 09:09:25.187611 4556 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0cba6771-0129-4ae1-b8e0-c13b427939b1-client-ca\") on node \"crc\" DevicePath \"\"" Feb 18 09:09:25 crc kubenswrapper[4556]: I0218 09:09:25.187657 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-82grg\" (UniqueName: \"kubernetes.io/projected/0cba6771-0129-4ae1-b8e0-c13b427939b1-kube-api-access-82grg\") on node \"crc\" DevicePath \"\"" Feb 18 09:09:25 crc kubenswrapper[4556]: I0218 09:09:25.187700 4556 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7cf2e5ac-c19f-4694-af7f-70bc6bcbc720-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 09:09:25 crc kubenswrapper[4556]: I0218 09:09:25.777367 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7c4d96c475-dgppt" event={"ID":"0cba6771-0129-4ae1-b8e0-c13b427939b1","Type":"ContainerDied","Data":"e6fe62ea209df75edd3b3a20e869dbe7a73235fc374b7c5e7b2e02ea59eedb7a"} Feb 18 09:09:25 crc kubenswrapper[4556]: I0218 09:09:25.777383 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7c4d96c475-dgppt" Feb 18 09:09:25 crc kubenswrapper[4556]: I0218 09:09:25.778083 4556 scope.go:117] "RemoveContainer" containerID="b20fb33b326fe5a0f4dc3819bdaa9e6ecd4abdfb9921328997860afe284d9088" Feb 18 09:09:25 crc kubenswrapper[4556]: I0218 09:09:25.779254 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-754b797845-2t4bs" event={"ID":"7cf2e5ac-c19f-4694-af7f-70bc6bcbc720","Type":"ContainerDied","Data":"7f3370e4ea8bbef9353573382e866188b7e6602a23f1c15c97d8633b6ddd6281"} Feb 18 09:09:25 crc kubenswrapper[4556]: I0218 09:09:25.779314 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-754b797845-2t4bs" Feb 18 09:09:25 crc kubenswrapper[4556]: I0218 09:09:25.794409 4556 scope.go:117] "RemoveContainer" containerID="377564960a52e642e12544eef0a23dbe3d4907b01520dbbeec8f5e8fe13ef34b" Feb 18 09:09:25 crc kubenswrapper[4556]: I0218 09:09:25.796557 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-754b797845-2t4bs"] Feb 18 09:09:25 crc kubenswrapper[4556]: I0218 09:09:25.799956 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-754b797845-2t4bs"] Feb 18 09:09:25 crc kubenswrapper[4556]: I0218 09:09:25.806188 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7c4d96c475-dgppt"] Feb 18 09:09:25 crc kubenswrapper[4556]: I0218 09:09:25.807884 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7c4d96c475-dgppt"] Feb 18 09:09:25 crc kubenswrapper[4556]: I0218 09:09:25.999709 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7bd4bf85d5-67z5x"] Feb 18 09:09:25 crc kubenswrapper[4556]: E0218 09:09:25.999971 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cf2e5ac-c19f-4694-af7f-70bc6bcbc720" containerName="controller-manager" Feb 18 09:09:25 crc kubenswrapper[4556]: I0218 09:09:25.999991 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cf2e5ac-c19f-4694-af7f-70bc6bcbc720" containerName="controller-manager" Feb 18 09:09:26 crc kubenswrapper[4556]: E0218 09:09:26.000008 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cba6771-0129-4ae1-b8e0-c13b427939b1" containerName="route-controller-manager" Feb 18 09:09:26 crc kubenswrapper[4556]: I0218 09:09:26.000014 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cba6771-0129-4ae1-b8e0-c13b427939b1" containerName="route-controller-manager" Feb 18 09:09:26 crc kubenswrapper[4556]: I0218 09:09:26.000127 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cba6771-0129-4ae1-b8e0-c13b427939b1" containerName="route-controller-manager" Feb 18 09:09:26 crc kubenswrapper[4556]: I0218 09:09:26.000146 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="7cf2e5ac-c19f-4694-af7f-70bc6bcbc720" containerName="controller-manager" Feb 18 09:09:26 crc kubenswrapper[4556]: I0218 09:09:26.001266 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7bd4bf85d5-67z5x" Feb 18 09:09:26 crc kubenswrapper[4556]: I0218 09:09:26.003077 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-569ff4fbdf-hzfrc"] Feb 18 09:09:26 crc kubenswrapper[4556]: I0218 09:09:26.003569 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 18 09:09:26 crc kubenswrapper[4556]: I0218 09:09:26.003774 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 18 09:09:26 crc kubenswrapper[4556]: I0218 09:09:26.003731 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 18 09:09:26 crc kubenswrapper[4556]: I0218 09:09:26.003853 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-569ff4fbdf-hzfrc" Feb 18 09:09:26 crc kubenswrapper[4556]: I0218 09:09:26.003867 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 18 09:09:26 crc kubenswrapper[4556]: I0218 09:09:26.005471 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 18 09:09:26 crc kubenswrapper[4556]: I0218 09:09:26.005665 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 18 09:09:26 crc kubenswrapper[4556]: I0218 09:09:26.007476 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 18 09:09:26 crc kubenswrapper[4556]: I0218 09:09:26.007625 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 18 09:09:26 crc kubenswrapper[4556]: I0218 09:09:26.007829 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 18 09:09:26 crc kubenswrapper[4556]: I0218 09:09:26.009416 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 18 09:09:26 crc kubenswrapper[4556]: I0218 09:09:26.009481 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 18 09:09:26 crc kubenswrapper[4556]: I0218 09:09:26.011370 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-569ff4fbdf-hzfrc"] Feb 18 09:09:26 crc kubenswrapper[4556]: I0218 09:09:26.014077 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 18 09:09:26 crc kubenswrapper[4556]: I0218 09:09:26.015147 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7bd4bf85d5-67z5x"] Feb 18 09:09:26 crc kubenswrapper[4556]: I0218 09:09:26.015679 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 18 09:09:26 crc kubenswrapper[4556]: I0218 09:09:26.100332 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f59d57b-9765-4379-8216-6c0d57041a94-serving-cert\") pod \"route-controller-manager-7bd4bf85d5-67z5x\" (UID: \"8f59d57b-9765-4379-8216-6c0d57041a94\") " pod="openshift-route-controller-manager/route-controller-manager-7bd4bf85d5-67z5x" Feb 18 09:09:26 crc kubenswrapper[4556]: I0218 09:09:26.100404 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvcmv\" (UniqueName: \"kubernetes.io/projected/8f59d57b-9765-4379-8216-6c0d57041a94-kube-api-access-jvcmv\") pod \"route-controller-manager-7bd4bf85d5-67z5x\" (UID: \"8f59d57b-9765-4379-8216-6c0d57041a94\") " pod="openshift-route-controller-manager/route-controller-manager-7bd4bf85d5-67z5x" Feb 18 09:09:26 crc kubenswrapper[4556]: I0218 09:09:26.100624 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8f59d57b-9765-4379-8216-6c0d57041a94-client-ca\") pod \"route-controller-manager-7bd4bf85d5-67z5x\" (UID: \"8f59d57b-9765-4379-8216-6c0d57041a94\") " pod="openshift-route-controller-manager/route-controller-manager-7bd4bf85d5-67z5x" Feb 18 09:09:26 crc kubenswrapper[4556]: I0218 09:09:26.100762 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f59d57b-9765-4379-8216-6c0d57041a94-config\") pod \"route-controller-manager-7bd4bf85d5-67z5x\" (UID: \"8f59d57b-9765-4379-8216-6c0d57041a94\") " pod="openshift-route-controller-manager/route-controller-manager-7bd4bf85d5-67z5x" Feb 18 09:09:26 crc kubenswrapper[4556]: I0218 09:09:26.201888 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvcmv\" (UniqueName: \"kubernetes.io/projected/8f59d57b-9765-4379-8216-6c0d57041a94-kube-api-access-jvcmv\") pod \"route-controller-manager-7bd4bf85d5-67z5x\" (UID: \"8f59d57b-9765-4379-8216-6c0d57041a94\") " pod="openshift-route-controller-manager/route-controller-manager-7bd4bf85d5-67z5x" Feb 18 09:09:26 crc kubenswrapper[4556]: I0218 09:09:26.201947 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4347df98-d929-4a76-9be5-2da99a4696c4-serving-cert\") pod \"controller-manager-569ff4fbdf-hzfrc\" (UID: \"4347df98-d929-4a76-9be5-2da99a4696c4\") " pod="openshift-controller-manager/controller-manager-569ff4fbdf-hzfrc" Feb 18 09:09:26 crc kubenswrapper[4556]: I0218 09:09:26.201991 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4347df98-d929-4a76-9be5-2da99a4696c4-config\") pod \"controller-manager-569ff4fbdf-hzfrc\" (UID: \"4347df98-d929-4a76-9be5-2da99a4696c4\") " pod="openshift-controller-manager/controller-manager-569ff4fbdf-hzfrc" Feb 18 09:09:26 crc kubenswrapper[4556]: I0218 09:09:26.202024 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4347df98-d929-4a76-9be5-2da99a4696c4-proxy-ca-bundles\") pod \"controller-manager-569ff4fbdf-hzfrc\" (UID: \"4347df98-d929-4a76-9be5-2da99a4696c4\") " pod="openshift-controller-manager/controller-manager-569ff4fbdf-hzfrc" Feb 18 09:09:26 crc kubenswrapper[4556]: I0218 09:09:26.202060 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8f59d57b-9765-4379-8216-6c0d57041a94-client-ca\") pod \"route-controller-manager-7bd4bf85d5-67z5x\" (UID: \"8f59d57b-9765-4379-8216-6c0d57041a94\") " pod="openshift-route-controller-manager/route-controller-manager-7bd4bf85d5-67z5x" Feb 18 09:09:26 crc kubenswrapper[4556]: I0218 09:09:26.202085 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8lt8\" (UniqueName: \"kubernetes.io/projected/4347df98-d929-4a76-9be5-2da99a4696c4-kube-api-access-c8lt8\") pod \"controller-manager-569ff4fbdf-hzfrc\" (UID: \"4347df98-d929-4a76-9be5-2da99a4696c4\") " pod="openshift-controller-manager/controller-manager-569ff4fbdf-hzfrc" Feb 18 09:09:26 crc kubenswrapper[4556]: I0218 09:09:26.202109 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4347df98-d929-4a76-9be5-2da99a4696c4-client-ca\") pod \"controller-manager-569ff4fbdf-hzfrc\" (UID: \"4347df98-d929-4a76-9be5-2da99a4696c4\") " pod="openshift-controller-manager/controller-manager-569ff4fbdf-hzfrc" Feb 18 09:09:26 crc kubenswrapper[4556]: I0218 09:09:26.202493 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f59d57b-9765-4379-8216-6c0d57041a94-config\") pod \"route-controller-manager-7bd4bf85d5-67z5x\" (UID: \"8f59d57b-9765-4379-8216-6c0d57041a94\") " pod="openshift-route-controller-manager/route-controller-manager-7bd4bf85d5-67z5x" Feb 18 09:09:26 crc kubenswrapper[4556]: I0218 09:09:26.202605 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f59d57b-9765-4379-8216-6c0d57041a94-serving-cert\") pod \"route-controller-manager-7bd4bf85d5-67z5x\" (UID: \"8f59d57b-9765-4379-8216-6c0d57041a94\") " pod="openshift-route-controller-manager/route-controller-manager-7bd4bf85d5-67z5x" Feb 18 09:09:26 crc kubenswrapper[4556]: I0218 09:09:26.203148 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8f59d57b-9765-4379-8216-6c0d57041a94-client-ca\") pod \"route-controller-manager-7bd4bf85d5-67z5x\" (UID: \"8f59d57b-9765-4379-8216-6c0d57041a94\") " pod="openshift-route-controller-manager/route-controller-manager-7bd4bf85d5-67z5x" Feb 18 09:09:26 crc kubenswrapper[4556]: I0218 09:09:26.203645 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f59d57b-9765-4379-8216-6c0d57041a94-config\") pod \"route-controller-manager-7bd4bf85d5-67z5x\" (UID: \"8f59d57b-9765-4379-8216-6c0d57041a94\") " pod="openshift-route-controller-manager/route-controller-manager-7bd4bf85d5-67z5x" Feb 18 09:09:26 crc kubenswrapper[4556]: I0218 09:09:26.207194 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f59d57b-9765-4379-8216-6c0d57041a94-serving-cert\") pod \"route-controller-manager-7bd4bf85d5-67z5x\" (UID: \"8f59d57b-9765-4379-8216-6c0d57041a94\") " pod="openshift-route-controller-manager/route-controller-manager-7bd4bf85d5-67z5x" Feb 18 09:09:26 crc kubenswrapper[4556]: I0218 09:09:26.215969 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvcmv\" (UniqueName: \"kubernetes.io/projected/8f59d57b-9765-4379-8216-6c0d57041a94-kube-api-access-jvcmv\") pod \"route-controller-manager-7bd4bf85d5-67z5x\" (UID: \"8f59d57b-9765-4379-8216-6c0d57041a94\") " pod="openshift-route-controller-manager/route-controller-manager-7bd4bf85d5-67z5x" Feb 18 09:09:26 crc kubenswrapper[4556]: I0218 09:09:26.303869 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4347df98-d929-4a76-9be5-2da99a4696c4-client-ca\") pod \"controller-manager-569ff4fbdf-hzfrc\" (UID: \"4347df98-d929-4a76-9be5-2da99a4696c4\") " pod="openshift-controller-manager/controller-manager-569ff4fbdf-hzfrc" Feb 18 09:09:26 crc kubenswrapper[4556]: I0218 09:09:26.303923 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8lt8\" (UniqueName: \"kubernetes.io/projected/4347df98-d929-4a76-9be5-2da99a4696c4-kube-api-access-c8lt8\") pod \"controller-manager-569ff4fbdf-hzfrc\" (UID: \"4347df98-d929-4a76-9be5-2da99a4696c4\") " pod="openshift-controller-manager/controller-manager-569ff4fbdf-hzfrc" Feb 18 09:09:26 crc kubenswrapper[4556]: I0218 09:09:26.304020 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4347df98-d929-4a76-9be5-2da99a4696c4-serving-cert\") pod \"controller-manager-569ff4fbdf-hzfrc\" (UID: \"4347df98-d929-4a76-9be5-2da99a4696c4\") " pod="openshift-controller-manager/controller-manager-569ff4fbdf-hzfrc" Feb 18 09:09:26 crc kubenswrapper[4556]: I0218 09:09:26.304059 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4347df98-d929-4a76-9be5-2da99a4696c4-config\") pod \"controller-manager-569ff4fbdf-hzfrc\" (UID: \"4347df98-d929-4a76-9be5-2da99a4696c4\") " pod="openshift-controller-manager/controller-manager-569ff4fbdf-hzfrc" Feb 18 09:09:26 crc kubenswrapper[4556]: I0218 09:09:26.304090 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4347df98-d929-4a76-9be5-2da99a4696c4-proxy-ca-bundles\") pod \"controller-manager-569ff4fbdf-hzfrc\" (UID: \"4347df98-d929-4a76-9be5-2da99a4696c4\") " pod="openshift-controller-manager/controller-manager-569ff4fbdf-hzfrc" Feb 18 09:09:26 crc kubenswrapper[4556]: I0218 09:09:26.304923 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4347df98-d929-4a76-9be5-2da99a4696c4-client-ca\") pod \"controller-manager-569ff4fbdf-hzfrc\" (UID: \"4347df98-d929-4a76-9be5-2da99a4696c4\") " pod="openshift-controller-manager/controller-manager-569ff4fbdf-hzfrc" Feb 18 09:09:26 crc kubenswrapper[4556]: I0218 09:09:26.305390 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4347df98-d929-4a76-9be5-2da99a4696c4-config\") pod \"controller-manager-569ff4fbdf-hzfrc\" (UID: \"4347df98-d929-4a76-9be5-2da99a4696c4\") " pod="openshift-controller-manager/controller-manager-569ff4fbdf-hzfrc" Feb 18 09:09:26 crc kubenswrapper[4556]: I0218 09:09:26.305477 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4347df98-d929-4a76-9be5-2da99a4696c4-proxy-ca-bundles\") pod \"controller-manager-569ff4fbdf-hzfrc\" (UID: \"4347df98-d929-4a76-9be5-2da99a4696c4\") " pod="openshift-controller-manager/controller-manager-569ff4fbdf-hzfrc" Feb 18 09:09:26 crc kubenswrapper[4556]: I0218 09:09:26.307301 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4347df98-d929-4a76-9be5-2da99a4696c4-serving-cert\") pod \"controller-manager-569ff4fbdf-hzfrc\" (UID: \"4347df98-d929-4a76-9be5-2da99a4696c4\") " pod="openshift-controller-manager/controller-manager-569ff4fbdf-hzfrc" Feb 18 09:09:26 crc kubenswrapper[4556]: I0218 09:09:26.318677 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8lt8\" (UniqueName: \"kubernetes.io/projected/4347df98-d929-4a76-9be5-2da99a4696c4-kube-api-access-c8lt8\") pod \"controller-manager-569ff4fbdf-hzfrc\" (UID: \"4347df98-d929-4a76-9be5-2da99a4696c4\") " pod="openshift-controller-manager/controller-manager-569ff4fbdf-hzfrc" Feb 18 09:09:26 crc kubenswrapper[4556]: I0218 09:09:26.319143 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7bd4bf85d5-67z5x" Feb 18 09:09:26 crc kubenswrapper[4556]: I0218 09:09:26.324503 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-569ff4fbdf-hzfrc" Feb 18 09:09:26 crc kubenswrapper[4556]: I0218 09:09:26.688581 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7bd4bf85d5-67z5x"] Feb 18 09:09:26 crc kubenswrapper[4556]: I0218 09:09:26.720779 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-569ff4fbdf-hzfrc"] Feb 18 09:09:26 crc kubenswrapper[4556]: W0218 09:09:26.746863 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4347df98_d929_4a76_9be5_2da99a4696c4.slice/crio-f0b6a677ed1710ebd92e312f59445350c3ca4e48a15a561e43329920d316d700 WatchSource:0}: Error finding container f0b6a677ed1710ebd92e312f59445350c3ca4e48a15a561e43329920d316d700: Status 404 returned error can't find the container with id f0b6a677ed1710ebd92e312f59445350c3ca4e48a15a561e43329920d316d700 Feb 18 09:09:26 crc kubenswrapper[4556]: I0218 09:09:26.785505 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-569ff4fbdf-hzfrc" event={"ID":"4347df98-d929-4a76-9be5-2da99a4696c4","Type":"ContainerStarted","Data":"f0b6a677ed1710ebd92e312f59445350c3ca4e48a15a561e43329920d316d700"} Feb 18 09:09:26 crc kubenswrapper[4556]: I0218 09:09:26.786624 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7bd4bf85d5-67z5x" event={"ID":"8f59d57b-9765-4379-8216-6c0d57041a94","Type":"ContainerStarted","Data":"5dcbc2ca2d7f85d85aa2d80d69f35f09a2c1dc5b941b7c8dda5aab18bdd7f5c0"} Feb 18 09:09:27 crc kubenswrapper[4556]: I0218 09:09:27.288522 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0cba6771-0129-4ae1-b8e0-c13b427939b1" path="/var/lib/kubelet/pods/0cba6771-0129-4ae1-b8e0-c13b427939b1/volumes" Feb 18 09:09:27 crc kubenswrapper[4556]: I0218 09:09:27.289319 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7cf2e5ac-c19f-4694-af7f-70bc6bcbc720" path="/var/lib/kubelet/pods/7cf2e5ac-c19f-4694-af7f-70bc6bcbc720/volumes" Feb 18 09:09:27 crc kubenswrapper[4556]: I0218 09:09:27.792784 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-569ff4fbdf-hzfrc" event={"ID":"4347df98-d929-4a76-9be5-2da99a4696c4","Type":"ContainerStarted","Data":"641eece3d8adf7ac5658540fe525b55115f7bc06532159cee1474fdcd3e0bc9f"} Feb 18 09:09:27 crc kubenswrapper[4556]: I0218 09:09:27.793573 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-569ff4fbdf-hzfrc" Feb 18 09:09:27 crc kubenswrapper[4556]: I0218 09:09:27.795000 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7bd4bf85d5-67z5x" event={"ID":"8f59d57b-9765-4379-8216-6c0d57041a94","Type":"ContainerStarted","Data":"41c1f35a3e5f1d4c7d829a48914ead541b3ceb626eb034a4546cbdfdf9b64e13"} Feb 18 09:09:27 crc kubenswrapper[4556]: I0218 09:09:27.795563 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7bd4bf85d5-67z5x" Feb 18 09:09:27 crc kubenswrapper[4556]: I0218 09:09:27.798693 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-569ff4fbdf-hzfrc" Feb 18 09:09:27 crc kubenswrapper[4556]: I0218 09:09:27.799652 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7bd4bf85d5-67z5x" Feb 18 09:09:27 crc kubenswrapper[4556]: I0218 09:09:27.809881 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-569ff4fbdf-hzfrc" podStartSLOduration=3.809870182 podStartE2EDuration="3.809870182s" podCreationTimestamp="2026-02-18 09:09:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:09:27.807426324 +0000 UTC m=+324.824387304" watchObservedRunningTime="2026-02-18 09:09:27.809870182 +0000 UTC m=+324.826831162" Feb 18 09:09:27 crc kubenswrapper[4556]: I0218 09:09:27.840495 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7bd4bf85d5-67z5x" podStartSLOduration=3.840475372 podStartE2EDuration="3.840475372s" podCreationTimestamp="2026-02-18 09:09:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:09:27.839785628 +0000 UTC m=+324.856746609" watchObservedRunningTime="2026-02-18 09:09:27.840475372 +0000 UTC m=+324.857436352" Feb 18 09:09:31 crc kubenswrapper[4556]: I0218 09:09:31.728042 4556 patch_prober.go:28] interesting pod/machine-config-daemon-f76hs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 09:09:31 crc kubenswrapper[4556]: I0218 09:09:31.729050 4556 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 09:09:39 crc kubenswrapper[4556]: I0218 09:09:39.056680 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-6ldbx" Feb 18 09:09:39 crc kubenswrapper[4556]: I0218 09:09:39.098046 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-wz622"] Feb 18 09:10:01 crc kubenswrapper[4556]: I0218 09:10:01.727873 4556 patch_prober.go:28] interesting pod/machine-config-daemon-f76hs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 09:10:01 crc kubenswrapper[4556]: I0218 09:10:01.728304 4556 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 09:10:04 crc kubenswrapper[4556]: I0218 09:10:04.125881 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-wz622" podUID="709f253c-93c3-4aaf-b3db-630775991aeb" containerName="registry" containerID="cri-o://910086a9ffd35fe9f573814c9d702a1ea8340e0bc824e9adcdc009f52d38ffda" gracePeriod=30 Feb 18 09:10:04 crc kubenswrapper[4556]: I0218 09:10:04.453740 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:10:04 crc kubenswrapper[4556]: I0218 09:10:04.582818 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/709f253c-93c3-4aaf-b3db-630775991aeb-registry-tls\") pod \"709f253c-93c3-4aaf-b3db-630775991aeb\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " Feb 18 09:10:04 crc kubenswrapper[4556]: I0218 09:10:04.582884 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/709f253c-93c3-4aaf-b3db-630775991aeb-ca-trust-extracted\") pod \"709f253c-93c3-4aaf-b3db-630775991aeb\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " Feb 18 09:10:04 crc kubenswrapper[4556]: I0218 09:10:04.582917 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/709f253c-93c3-4aaf-b3db-630775991aeb-installation-pull-secrets\") pod \"709f253c-93c3-4aaf-b3db-630775991aeb\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " Feb 18 09:10:04 crc kubenswrapper[4556]: I0218 09:10:04.583027 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"709f253c-93c3-4aaf-b3db-630775991aeb\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " Feb 18 09:10:04 crc kubenswrapper[4556]: I0218 09:10:04.583054 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/709f253c-93c3-4aaf-b3db-630775991aeb-bound-sa-token\") pod \"709f253c-93c3-4aaf-b3db-630775991aeb\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " Feb 18 09:10:04 crc kubenswrapper[4556]: I0218 09:10:04.583096 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/709f253c-93c3-4aaf-b3db-630775991aeb-trusted-ca\") pod \"709f253c-93c3-4aaf-b3db-630775991aeb\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " Feb 18 09:10:04 crc kubenswrapper[4556]: I0218 09:10:04.583130 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/709f253c-93c3-4aaf-b3db-630775991aeb-registry-certificates\") pod \"709f253c-93c3-4aaf-b3db-630775991aeb\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " Feb 18 09:10:04 crc kubenswrapper[4556]: I0218 09:10:04.583172 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2vrqj\" (UniqueName: \"kubernetes.io/projected/709f253c-93c3-4aaf-b3db-630775991aeb-kube-api-access-2vrqj\") pod \"709f253c-93c3-4aaf-b3db-630775991aeb\" (UID: \"709f253c-93c3-4aaf-b3db-630775991aeb\") " Feb 18 09:10:04 crc kubenswrapper[4556]: I0218 09:10:04.583733 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/709f253c-93c3-4aaf-b3db-630775991aeb-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "709f253c-93c3-4aaf-b3db-630775991aeb" (UID: "709f253c-93c3-4aaf-b3db-630775991aeb"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:10:04 crc kubenswrapper[4556]: I0218 09:10:04.583817 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/709f253c-93c3-4aaf-b3db-630775991aeb-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "709f253c-93c3-4aaf-b3db-630775991aeb" (UID: "709f253c-93c3-4aaf-b3db-630775991aeb"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:10:04 crc kubenswrapper[4556]: I0218 09:10:04.587750 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/709f253c-93c3-4aaf-b3db-630775991aeb-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "709f253c-93c3-4aaf-b3db-630775991aeb" (UID: "709f253c-93c3-4aaf-b3db-630775991aeb"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:10:04 crc kubenswrapper[4556]: I0218 09:10:04.587745 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/709f253c-93c3-4aaf-b3db-630775991aeb-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "709f253c-93c3-4aaf-b3db-630775991aeb" (UID: "709f253c-93c3-4aaf-b3db-630775991aeb"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:10:04 crc kubenswrapper[4556]: I0218 09:10:04.588116 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/709f253c-93c3-4aaf-b3db-630775991aeb-kube-api-access-2vrqj" (OuterVolumeSpecName: "kube-api-access-2vrqj") pod "709f253c-93c3-4aaf-b3db-630775991aeb" (UID: "709f253c-93c3-4aaf-b3db-630775991aeb"). InnerVolumeSpecName "kube-api-access-2vrqj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:10:04 crc kubenswrapper[4556]: I0218 09:10:04.588210 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/709f253c-93c3-4aaf-b3db-630775991aeb-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "709f253c-93c3-4aaf-b3db-630775991aeb" (UID: "709f253c-93c3-4aaf-b3db-630775991aeb"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:10:04 crc kubenswrapper[4556]: I0218 09:10:04.590825 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "709f253c-93c3-4aaf-b3db-630775991aeb" (UID: "709f253c-93c3-4aaf-b3db-630775991aeb"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 18 09:10:04 crc kubenswrapper[4556]: I0218 09:10:04.597734 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/709f253c-93c3-4aaf-b3db-630775991aeb-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "709f253c-93c3-4aaf-b3db-630775991aeb" (UID: "709f253c-93c3-4aaf-b3db-630775991aeb"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:10:04 crc kubenswrapper[4556]: I0218 09:10:04.683844 4556 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/709f253c-93c3-4aaf-b3db-630775991aeb-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 18 09:10:04 crc kubenswrapper[4556]: I0218 09:10:04.683867 4556 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/709f253c-93c3-4aaf-b3db-630775991aeb-registry-certificates\") on node \"crc\" DevicePath \"\"" Feb 18 09:10:04 crc kubenswrapper[4556]: I0218 09:10:04.683878 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2vrqj\" (UniqueName: \"kubernetes.io/projected/709f253c-93c3-4aaf-b3db-630775991aeb-kube-api-access-2vrqj\") on node \"crc\" DevicePath \"\"" Feb 18 09:10:04 crc kubenswrapper[4556]: I0218 09:10:04.683886 4556 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/709f253c-93c3-4aaf-b3db-630775991aeb-registry-tls\") on node \"crc\" DevicePath \"\"" Feb 18 09:10:04 crc kubenswrapper[4556]: I0218 09:10:04.683897 4556 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/709f253c-93c3-4aaf-b3db-630775991aeb-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Feb 18 09:10:04 crc kubenswrapper[4556]: I0218 09:10:04.683905 4556 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/709f253c-93c3-4aaf-b3db-630775991aeb-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Feb 18 09:10:04 crc kubenswrapper[4556]: I0218 09:10:04.683913 4556 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/709f253c-93c3-4aaf-b3db-630775991aeb-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 18 09:10:04 crc kubenswrapper[4556]: I0218 09:10:04.957109 4556 generic.go:334] "Generic (PLEG): container finished" podID="709f253c-93c3-4aaf-b3db-630775991aeb" containerID="910086a9ffd35fe9f573814c9d702a1ea8340e0bc824e9adcdc009f52d38ffda" exitCode=0 Feb 18 09:10:04 crc kubenswrapper[4556]: I0218 09:10:04.957141 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-wz622" event={"ID":"709f253c-93c3-4aaf-b3db-630775991aeb","Type":"ContainerDied","Data":"910086a9ffd35fe9f573814c9d702a1ea8340e0bc824e9adcdc009f52d38ffda"} Feb 18 09:10:04 crc kubenswrapper[4556]: I0218 09:10:04.957193 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-wz622" event={"ID":"709f253c-93c3-4aaf-b3db-630775991aeb","Type":"ContainerDied","Data":"a9de573ef0d283c2f4720857e5f6a03636743796f09c3e066c47154a783e9a04"} Feb 18 09:10:04 crc kubenswrapper[4556]: I0218 09:10:04.957211 4556 scope.go:117] "RemoveContainer" containerID="910086a9ffd35fe9f573814c9d702a1ea8340e0bc824e9adcdc009f52d38ffda" Feb 18 09:10:04 crc kubenswrapper[4556]: I0218 09:10:04.957482 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-wz622" Feb 18 09:10:04 crc kubenswrapper[4556]: I0218 09:10:04.969420 4556 scope.go:117] "RemoveContainer" containerID="910086a9ffd35fe9f573814c9d702a1ea8340e0bc824e9adcdc009f52d38ffda" Feb 18 09:10:04 crc kubenswrapper[4556]: E0218 09:10:04.969971 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"910086a9ffd35fe9f573814c9d702a1ea8340e0bc824e9adcdc009f52d38ffda\": container with ID starting with 910086a9ffd35fe9f573814c9d702a1ea8340e0bc824e9adcdc009f52d38ffda not found: ID does not exist" containerID="910086a9ffd35fe9f573814c9d702a1ea8340e0bc824e9adcdc009f52d38ffda" Feb 18 09:10:04 crc kubenswrapper[4556]: I0218 09:10:04.970068 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"910086a9ffd35fe9f573814c9d702a1ea8340e0bc824e9adcdc009f52d38ffda"} err="failed to get container status \"910086a9ffd35fe9f573814c9d702a1ea8340e0bc824e9adcdc009f52d38ffda\": rpc error: code = NotFound desc = could not find container \"910086a9ffd35fe9f573814c9d702a1ea8340e0bc824e9adcdc009f52d38ffda\": container with ID starting with 910086a9ffd35fe9f573814c9d702a1ea8340e0bc824e9adcdc009f52d38ffda not found: ID does not exist" Feb 18 09:10:04 crc kubenswrapper[4556]: I0218 09:10:04.977643 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-wz622"] Feb 18 09:10:04 crc kubenswrapper[4556]: I0218 09:10:04.980384 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-wz622"] Feb 18 09:10:05 crc kubenswrapper[4556]: I0218 09:10:05.292310 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="709f253c-93c3-4aaf-b3db-630775991aeb" path="/var/lib/kubelet/pods/709f253c-93c3-4aaf-b3db-630775991aeb/volumes" Feb 18 09:10:31 crc kubenswrapper[4556]: I0218 09:10:31.727990 4556 patch_prober.go:28] interesting pod/machine-config-daemon-f76hs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 09:10:31 crc kubenswrapper[4556]: I0218 09:10:31.728395 4556 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 09:10:31 crc kubenswrapper[4556]: I0218 09:10:31.728442 4556 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" Feb 18 09:10:31 crc kubenswrapper[4556]: I0218 09:10:31.728871 4556 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8b54607784a82541a6af0430808ac511ecebb51bf04484e5e4e70f755a71305e"} pod="openshift-machine-config-operator/machine-config-daemon-f76hs" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 18 09:10:31 crc kubenswrapper[4556]: I0218 09:10:31.728920 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerName="machine-config-daemon" containerID="cri-o://8b54607784a82541a6af0430808ac511ecebb51bf04484e5e4e70f755a71305e" gracePeriod=600 Feb 18 09:10:32 crc kubenswrapper[4556]: I0218 09:10:32.061057 4556 generic.go:334] "Generic (PLEG): container finished" podID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerID="8b54607784a82541a6af0430808ac511ecebb51bf04484e5e4e70f755a71305e" exitCode=0 Feb 18 09:10:32 crc kubenswrapper[4556]: I0218 09:10:32.061133 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" event={"ID":"8dac7f27-d3d1-4778-9e54-f273035a1d37","Type":"ContainerDied","Data":"8b54607784a82541a6af0430808ac511ecebb51bf04484e5e4e70f755a71305e"} Feb 18 09:10:32 crc kubenswrapper[4556]: I0218 09:10:32.061477 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" event={"ID":"8dac7f27-d3d1-4778-9e54-f273035a1d37","Type":"ContainerStarted","Data":"dc67a95c5ea909b9cea5fe464915331393bc298bd4bcd198fcb5ebe8f8ba8a4b"} Feb 18 09:10:32 crc kubenswrapper[4556]: I0218 09:10:32.061501 4556 scope.go:117] "RemoveContainer" containerID="6d41b85076e695bbc51a41fce200e0e47c099d88b4d825b4d133d33ed96237e2" Feb 18 09:11:41 crc kubenswrapper[4556]: I0218 09:11:41.077075 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-wfnn4"] Feb 18 09:11:41 crc kubenswrapper[4556]: E0218 09:11:41.077860 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="709f253c-93c3-4aaf-b3db-630775991aeb" containerName="registry" Feb 18 09:11:41 crc kubenswrapper[4556]: I0218 09:11:41.077875 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="709f253c-93c3-4aaf-b3db-630775991aeb" containerName="registry" Feb 18 09:11:41 crc kubenswrapper[4556]: I0218 09:11:41.077984 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="709f253c-93c3-4aaf-b3db-630775991aeb" containerName="registry" Feb 18 09:11:41 crc kubenswrapper[4556]: I0218 09:11:41.078462 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-wfnn4" Feb 18 09:11:41 crc kubenswrapper[4556]: I0218 09:11:41.083024 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-858654f9db-bw685"] Feb 18 09:11:41 crc kubenswrapper[4556]: I0218 09:11:41.083910 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-bw685" Feb 18 09:11:41 crc kubenswrapper[4556]: I0218 09:11:41.086989 4556 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-76z9m" Feb 18 09:11:41 crc kubenswrapper[4556]: I0218 09:11:41.086994 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Feb 18 09:11:41 crc kubenswrapper[4556]: I0218 09:11:41.087016 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Feb 18 09:11:41 crc kubenswrapper[4556]: I0218 09:11:41.087372 4556 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-49xcc" Feb 18 09:11:41 crc kubenswrapper[4556]: I0218 09:11:41.096691 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-wfnn4"] Feb 18 09:11:41 crc kubenswrapper[4556]: I0218 09:11:41.100246 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-bw685"] Feb 18 09:11:41 crc kubenswrapper[4556]: I0218 09:11:41.108316 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-j67wz"] Feb 18 09:11:41 crc kubenswrapper[4556]: I0218 09:11:41.109020 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-j67wz" Feb 18 09:11:41 crc kubenswrapper[4556]: I0218 09:11:41.110824 4556 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-qd6tx" Feb 18 09:11:41 crc kubenswrapper[4556]: I0218 09:11:41.118651 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-j67wz"] Feb 18 09:11:41 crc kubenswrapper[4556]: I0218 09:11:41.126488 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bc6l9\" (UniqueName: \"kubernetes.io/projected/58f7a628-b1a3-4fd5-a730-0a80524f751c-kube-api-access-bc6l9\") pod \"cert-manager-webhook-687f57d79b-j67wz\" (UID: \"58f7a628-b1a3-4fd5-a730-0a80524f751c\") " pod="cert-manager/cert-manager-webhook-687f57d79b-j67wz" Feb 18 09:11:41 crc kubenswrapper[4556]: I0218 09:11:41.126526 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mw6n\" (UniqueName: \"kubernetes.io/projected/428e282a-fe3e-47f5-9d19-deb95236a4c9-kube-api-access-6mw6n\") pod \"cert-manager-cainjector-cf98fcc89-wfnn4\" (UID: \"428e282a-fe3e-47f5-9d19-deb95236a4c9\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-wfnn4" Feb 18 09:11:41 crc kubenswrapper[4556]: I0218 09:11:41.126562 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6lm6s\" (UniqueName: \"kubernetes.io/projected/257b9c79-fc7b-4bc5-88eb-d767e4762f77-kube-api-access-6lm6s\") pod \"cert-manager-858654f9db-bw685\" (UID: \"257b9c79-fc7b-4bc5-88eb-d767e4762f77\") " pod="cert-manager/cert-manager-858654f9db-bw685" Feb 18 09:11:41 crc kubenswrapper[4556]: I0218 09:11:41.227980 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bc6l9\" (UniqueName: \"kubernetes.io/projected/58f7a628-b1a3-4fd5-a730-0a80524f751c-kube-api-access-bc6l9\") pod \"cert-manager-webhook-687f57d79b-j67wz\" (UID: \"58f7a628-b1a3-4fd5-a730-0a80524f751c\") " pod="cert-manager/cert-manager-webhook-687f57d79b-j67wz" Feb 18 09:11:41 crc kubenswrapper[4556]: I0218 09:11:41.228043 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mw6n\" (UniqueName: \"kubernetes.io/projected/428e282a-fe3e-47f5-9d19-deb95236a4c9-kube-api-access-6mw6n\") pod \"cert-manager-cainjector-cf98fcc89-wfnn4\" (UID: \"428e282a-fe3e-47f5-9d19-deb95236a4c9\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-wfnn4" Feb 18 09:11:41 crc kubenswrapper[4556]: I0218 09:11:41.228133 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6lm6s\" (UniqueName: \"kubernetes.io/projected/257b9c79-fc7b-4bc5-88eb-d767e4762f77-kube-api-access-6lm6s\") pod \"cert-manager-858654f9db-bw685\" (UID: \"257b9c79-fc7b-4bc5-88eb-d767e4762f77\") " pod="cert-manager/cert-manager-858654f9db-bw685" Feb 18 09:11:41 crc kubenswrapper[4556]: I0218 09:11:41.247787 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bc6l9\" (UniqueName: \"kubernetes.io/projected/58f7a628-b1a3-4fd5-a730-0a80524f751c-kube-api-access-bc6l9\") pod \"cert-manager-webhook-687f57d79b-j67wz\" (UID: \"58f7a628-b1a3-4fd5-a730-0a80524f751c\") " pod="cert-manager/cert-manager-webhook-687f57d79b-j67wz" Feb 18 09:11:41 crc kubenswrapper[4556]: I0218 09:11:41.248177 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mw6n\" (UniqueName: \"kubernetes.io/projected/428e282a-fe3e-47f5-9d19-deb95236a4c9-kube-api-access-6mw6n\") pod \"cert-manager-cainjector-cf98fcc89-wfnn4\" (UID: \"428e282a-fe3e-47f5-9d19-deb95236a4c9\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-wfnn4" Feb 18 09:11:41 crc kubenswrapper[4556]: I0218 09:11:41.249691 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6lm6s\" (UniqueName: \"kubernetes.io/projected/257b9c79-fc7b-4bc5-88eb-d767e4762f77-kube-api-access-6lm6s\") pod \"cert-manager-858654f9db-bw685\" (UID: \"257b9c79-fc7b-4bc5-88eb-d767e4762f77\") " pod="cert-manager/cert-manager-858654f9db-bw685" Feb 18 09:11:41 crc kubenswrapper[4556]: I0218 09:11:41.397851 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-wfnn4" Feb 18 09:11:41 crc kubenswrapper[4556]: I0218 09:11:41.402376 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-bw685" Feb 18 09:11:41 crc kubenswrapper[4556]: I0218 09:11:41.423841 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-j67wz" Feb 18 09:11:41 crc kubenswrapper[4556]: I0218 09:11:41.628787 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-j67wz"] Feb 18 09:11:41 crc kubenswrapper[4556]: I0218 09:11:41.638058 4556 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 18 09:11:41 crc kubenswrapper[4556]: I0218 09:11:41.773206 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-wfnn4"] Feb 18 09:11:41 crc kubenswrapper[4556]: W0218 09:11:41.777720 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod428e282a_fe3e_47f5_9d19_deb95236a4c9.slice/crio-c32d646f4ea8da7c6b838c00cd8e62e358b7cfc7c7c923cfc361f0950daa43ad WatchSource:0}: Error finding container c32d646f4ea8da7c6b838c00cd8e62e358b7cfc7c7c923cfc361f0950daa43ad: Status 404 returned error can't find the container with id c32d646f4ea8da7c6b838c00cd8e62e358b7cfc7c7c923cfc361f0950daa43ad Feb 18 09:11:41 crc kubenswrapper[4556]: I0218 09:11:41.777844 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-bw685"] Feb 18 09:11:41 crc kubenswrapper[4556]: W0218 09:11:41.779753 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod257b9c79_fc7b_4bc5_88eb_d767e4762f77.slice/crio-7fdd0b399aee77ac9f1002e384547fbed956d3eb8a52daafd733840dc7ac68a1 WatchSource:0}: Error finding container 7fdd0b399aee77ac9f1002e384547fbed956d3eb8a52daafd733840dc7ac68a1: Status 404 returned error can't find the container with id 7fdd0b399aee77ac9f1002e384547fbed956d3eb8a52daafd733840dc7ac68a1 Feb 18 09:11:42 crc kubenswrapper[4556]: I0218 09:11:42.333737 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-wfnn4" event={"ID":"428e282a-fe3e-47f5-9d19-deb95236a4c9","Type":"ContainerStarted","Data":"c32d646f4ea8da7c6b838c00cd8e62e358b7cfc7c7c923cfc361f0950daa43ad"} Feb 18 09:11:42 crc kubenswrapper[4556]: I0218 09:11:42.334681 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-j67wz" event={"ID":"58f7a628-b1a3-4fd5-a730-0a80524f751c","Type":"ContainerStarted","Data":"6bf10a0299bc9071c5802eedadfe02a9a751c2d5ead96675b09b5599b5613f89"} Feb 18 09:11:42 crc kubenswrapper[4556]: I0218 09:11:42.335443 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-bw685" event={"ID":"257b9c79-fc7b-4bc5-88eb-d767e4762f77","Type":"ContainerStarted","Data":"7fdd0b399aee77ac9f1002e384547fbed956d3eb8a52daafd733840dc7ac68a1"} Feb 18 09:11:45 crc kubenswrapper[4556]: I0218 09:11:45.350603 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-bw685" event={"ID":"257b9c79-fc7b-4bc5-88eb-d767e4762f77","Type":"ContainerStarted","Data":"c2adbf5e8272df0213c54b65ee88890020da9098d66c35bcf5ce0e71f0254902"} Feb 18 09:11:45 crc kubenswrapper[4556]: I0218 09:11:45.352262 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-j67wz" event={"ID":"58f7a628-b1a3-4fd5-a730-0a80524f751c","Type":"ContainerStarted","Data":"95d0e50c3e86ecc88990fd7f43cb91ca2bbe4c38912fa1211687c277ae4f9f89"} Feb 18 09:11:45 crc kubenswrapper[4556]: I0218 09:11:45.352404 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-687f57d79b-j67wz" Feb 18 09:11:45 crc kubenswrapper[4556]: I0218 09:11:45.353698 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-wfnn4" event={"ID":"428e282a-fe3e-47f5-9d19-deb95236a4c9","Type":"ContainerStarted","Data":"cbd82e6df96de830c74aea7fc6b007b5b0122c62b236e0c94d44b173542a798d"} Feb 18 09:11:45 crc kubenswrapper[4556]: I0218 09:11:45.367918 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-858654f9db-bw685" podStartSLOduration=1.508941359 podStartE2EDuration="4.367901844s" podCreationTimestamp="2026-02-18 09:11:41 +0000 UTC" firstStartedPulling="2026-02-18 09:11:41.781693398 +0000 UTC m=+458.798654378" lastFinishedPulling="2026-02-18 09:11:44.640653883 +0000 UTC m=+461.657614863" observedRunningTime="2026-02-18 09:11:45.363924254 +0000 UTC m=+462.380885233" watchObservedRunningTime="2026-02-18 09:11:45.367901844 +0000 UTC m=+462.384862824" Feb 18 09:11:45 crc kubenswrapper[4556]: I0218 09:11:45.377998 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-cf98fcc89-wfnn4" podStartSLOduration=1.512631166 podStartE2EDuration="4.377984624s" podCreationTimestamp="2026-02-18 09:11:41 +0000 UTC" firstStartedPulling="2026-02-18 09:11:41.780241035 +0000 UTC m=+458.797202015" lastFinishedPulling="2026-02-18 09:11:44.645594493 +0000 UTC m=+461.662555473" observedRunningTime="2026-02-18 09:11:45.374892274 +0000 UTC m=+462.391853255" watchObservedRunningTime="2026-02-18 09:11:45.377984624 +0000 UTC m=+462.394945604" Feb 18 09:11:51 crc kubenswrapper[4556]: I0218 09:11:51.427100 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-687f57d79b-j67wz" Feb 18 09:11:51 crc kubenswrapper[4556]: I0218 09:11:51.442193 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-687f57d79b-j67wz" podStartSLOduration=7.435308873 podStartE2EDuration="10.442175387s" podCreationTimestamp="2026-02-18 09:11:41 +0000 UTC" firstStartedPulling="2026-02-18 09:11:41.637806008 +0000 UTC m=+458.654766988" lastFinishedPulling="2026-02-18 09:11:44.644672522 +0000 UTC m=+461.661633502" observedRunningTime="2026-02-18 09:11:45.394544587 +0000 UTC m=+462.411505567" watchObservedRunningTime="2026-02-18 09:11:51.442175387 +0000 UTC m=+468.459136367" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.329228 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-b7fsj"] Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.329943 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" podUID="09971522-31bc-4e10-8831-4fb62675e220" containerName="ovn-controller" containerID="cri-o://0c0f15d331af9bc2ca3082c1573dd167c66773eb390593844e681f8d0b455d3d" gracePeriod=30 Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.330015 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" podUID="09971522-31bc-4e10-8831-4fb62675e220" containerName="nbdb" containerID="cri-o://4cd6e758ee396c9000b1bdb6ded2176ec803fa9da6201cdd9b8146940a2c00cb" gracePeriod=30 Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.330071 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" podUID="09971522-31bc-4e10-8831-4fb62675e220" containerName="kube-rbac-proxy-node" containerID="cri-o://9170bdcbaf1065995d03e6448e40bdced0741c716c9e81f37a72b015132f1161" gracePeriod=30 Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.330083 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" podUID="09971522-31bc-4e10-8831-4fb62675e220" containerName="ovn-acl-logging" containerID="cri-o://318cb69e50df4c14df9bf5ef57d5c381421a9bdb72135c9b9620f55c0547b410" gracePeriod=30 Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.330085 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" podUID="09971522-31bc-4e10-8831-4fb62675e220" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://01c526c0613f19e1a1a584b7129bd56001faf08d0b97fd77feeb69f764ed01f8" gracePeriod=30 Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.330110 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" podUID="09971522-31bc-4e10-8831-4fb62675e220" containerName="northd" containerID="cri-o://9b0412f3330999c3348ac726ac5a0c2d541571f4b13786ca24cb3109a44aa653" gracePeriod=30 Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.330728 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" podUID="09971522-31bc-4e10-8831-4fb62675e220" containerName="sbdb" containerID="cri-o://aec6030df0a59ad8902d5c694d5cd30d249ca1e895155c57a458e359fa531b7c" gracePeriod=30 Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.351124 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" podUID="09971522-31bc-4e10-8831-4fb62675e220" containerName="ovnkube-controller" containerID="cri-o://ad37b48f74a075b5562885db93abefd92df9f91f81a3b2ff685ee97a13b5e403" gracePeriod=30 Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.419441 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-qfksk_e226b189-cfa9-47fc-b94d-19b5cbe0859f/kube-multus/1.log" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.419985 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-qfksk_e226b189-cfa9-47fc-b94d-19b5cbe0859f/kube-multus/0.log" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.420037 4556 generic.go:334] "Generic (PLEG): container finished" podID="e226b189-cfa9-47fc-b94d-19b5cbe0859f" containerID="23f488cd416c6d49acf193e27beba9e9f1dcfb36430f9ae540742c2e2bd0f144" exitCode=2 Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.420068 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-qfksk" event={"ID":"e226b189-cfa9-47fc-b94d-19b5cbe0859f","Type":"ContainerDied","Data":"23f488cd416c6d49acf193e27beba9e9f1dcfb36430f9ae540742c2e2bd0f144"} Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.420099 4556 scope.go:117] "RemoveContainer" containerID="fc2cb92a7735d16b8e5918797e52e1e488d9261a0418ed5ebbb94c033ca9328a" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.420555 4556 scope.go:117] "RemoveContainer" containerID="23f488cd416c6d49acf193e27beba9e9f1dcfb36430f9ae540742c2e2bd0f144" Feb 18 09:11:58 crc kubenswrapper[4556]: E0218 09:11:58.420725 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-qfksk_openshift-multus(e226b189-cfa9-47fc-b94d-19b5cbe0859f)\"" pod="openshift-multus/multus-qfksk" podUID="e226b189-cfa9-47fc-b94d-19b5cbe0859f" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.571086 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-b7fsj_09971522-31bc-4e10-8831-4fb62675e220/ovnkube-controller/3.log" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.572947 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-b7fsj_09971522-31bc-4e10-8831-4fb62675e220/ovn-acl-logging/0.log" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.573367 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-b7fsj_09971522-31bc-4e10-8831-4fb62675e220/ovn-controller/0.log" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.573771 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.609819 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-wdg5l"] Feb 18 09:11:58 crc kubenswrapper[4556]: E0218 09:11:58.610025 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09971522-31bc-4e10-8831-4fb62675e220" containerName="ovnkube-controller" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.610036 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="09971522-31bc-4e10-8831-4fb62675e220" containerName="ovnkube-controller" Feb 18 09:11:58 crc kubenswrapper[4556]: E0218 09:11:58.610047 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09971522-31bc-4e10-8831-4fb62675e220" containerName="kube-rbac-proxy-node" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.610052 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="09971522-31bc-4e10-8831-4fb62675e220" containerName="kube-rbac-proxy-node" Feb 18 09:11:58 crc kubenswrapper[4556]: E0218 09:11:58.610063 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09971522-31bc-4e10-8831-4fb62675e220" containerName="ovnkube-controller" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.610069 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="09971522-31bc-4e10-8831-4fb62675e220" containerName="ovnkube-controller" Feb 18 09:11:58 crc kubenswrapper[4556]: E0218 09:11:58.610075 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09971522-31bc-4e10-8831-4fb62675e220" containerName="sbdb" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.610081 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="09971522-31bc-4e10-8831-4fb62675e220" containerName="sbdb" Feb 18 09:11:58 crc kubenswrapper[4556]: E0218 09:11:58.610091 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09971522-31bc-4e10-8831-4fb62675e220" containerName="northd" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.610096 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="09971522-31bc-4e10-8831-4fb62675e220" containerName="northd" Feb 18 09:11:58 crc kubenswrapper[4556]: E0218 09:11:58.610103 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09971522-31bc-4e10-8831-4fb62675e220" containerName="ovnkube-controller" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.610108 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="09971522-31bc-4e10-8831-4fb62675e220" containerName="ovnkube-controller" Feb 18 09:11:58 crc kubenswrapper[4556]: E0218 09:11:58.610115 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09971522-31bc-4e10-8831-4fb62675e220" containerName="nbdb" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.610120 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="09971522-31bc-4e10-8831-4fb62675e220" containerName="nbdb" Feb 18 09:11:58 crc kubenswrapper[4556]: E0218 09:11:58.610127 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09971522-31bc-4e10-8831-4fb62675e220" containerName="ovnkube-controller" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.610132 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="09971522-31bc-4e10-8831-4fb62675e220" containerName="ovnkube-controller" Feb 18 09:11:58 crc kubenswrapper[4556]: E0218 09:11:58.610142 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09971522-31bc-4e10-8831-4fb62675e220" containerName="kubecfg-setup" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.610147 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="09971522-31bc-4e10-8831-4fb62675e220" containerName="kubecfg-setup" Feb 18 09:11:58 crc kubenswrapper[4556]: E0218 09:11:58.610177 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09971522-31bc-4e10-8831-4fb62675e220" containerName="ovn-acl-logging" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.610182 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="09971522-31bc-4e10-8831-4fb62675e220" containerName="ovn-acl-logging" Feb 18 09:11:58 crc kubenswrapper[4556]: E0218 09:11:58.610189 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09971522-31bc-4e10-8831-4fb62675e220" containerName="kube-rbac-proxy-ovn-metrics" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.610194 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="09971522-31bc-4e10-8831-4fb62675e220" containerName="kube-rbac-proxy-ovn-metrics" Feb 18 09:11:58 crc kubenswrapper[4556]: E0218 09:11:58.610201 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09971522-31bc-4e10-8831-4fb62675e220" containerName="ovn-controller" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.610207 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="09971522-31bc-4e10-8831-4fb62675e220" containerName="ovn-controller" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.610294 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="09971522-31bc-4e10-8831-4fb62675e220" containerName="ovnkube-controller" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.610303 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="09971522-31bc-4e10-8831-4fb62675e220" containerName="sbdb" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.610309 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="09971522-31bc-4e10-8831-4fb62675e220" containerName="ovnkube-controller" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.610316 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="09971522-31bc-4e10-8831-4fb62675e220" containerName="ovn-acl-logging" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.610325 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="09971522-31bc-4e10-8831-4fb62675e220" containerName="northd" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.610331 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="09971522-31bc-4e10-8831-4fb62675e220" containerName="nbdb" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.610339 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="09971522-31bc-4e10-8831-4fb62675e220" containerName="ovnkube-controller" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.610345 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="09971522-31bc-4e10-8831-4fb62675e220" containerName="kube-rbac-proxy-node" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.610351 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="09971522-31bc-4e10-8831-4fb62675e220" containerName="ovnkube-controller" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.610358 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="09971522-31bc-4e10-8831-4fb62675e220" containerName="kube-rbac-proxy-ovn-metrics" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.610364 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="09971522-31bc-4e10-8831-4fb62675e220" containerName="ovn-controller" Feb 18 09:11:58 crc kubenswrapper[4556]: E0218 09:11:58.610451 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09971522-31bc-4e10-8831-4fb62675e220" containerName="ovnkube-controller" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.610458 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="09971522-31bc-4e10-8831-4fb62675e220" containerName="ovnkube-controller" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.610526 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="09971522-31bc-4e10-8831-4fb62675e220" containerName="ovnkube-controller" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.611949 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.640361 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-host-run-ovn-kubernetes\") pod \"09971522-31bc-4e10-8831-4fb62675e220\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.640398 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-host-cni-bin\") pod \"09971522-31bc-4e10-8831-4fb62675e220\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.640421 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-run-ovn\") pod \"09971522-31bc-4e10-8831-4fb62675e220\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.640417 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "09971522-31bc-4e10-8831-4fb62675e220" (UID: "09971522-31bc-4e10-8831-4fb62675e220"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.640439 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "09971522-31bc-4e10-8831-4fb62675e220" (UID: "09971522-31bc-4e10-8831-4fb62675e220"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.640464 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "09971522-31bc-4e10-8831-4fb62675e220" (UID: "09971522-31bc-4e10-8831-4fb62675e220"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.640448 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-var-lib-openvswitch\") pod \"09971522-31bc-4e10-8831-4fb62675e220\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.640514 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-node-log\") pod \"09971522-31bc-4e10-8831-4fb62675e220\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.640481 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "09971522-31bc-4e10-8831-4fb62675e220" (UID: "09971522-31bc-4e10-8831-4fb62675e220"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.640544 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/09971522-31bc-4e10-8831-4fb62675e220-ovnkube-config\") pod \"09971522-31bc-4e10-8831-4fb62675e220\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.640574 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-node-log" (OuterVolumeSpecName: "node-log") pod "09971522-31bc-4e10-8831-4fb62675e220" (UID: "09971522-31bc-4e10-8831-4fb62675e220"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.640586 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-host-var-lib-cni-networks-ovn-kubernetes\") pod \"09971522-31bc-4e10-8831-4fb62675e220\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.640645 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-run-systemd\") pod \"09971522-31bc-4e10-8831-4fb62675e220\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.640666 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-log-socket\") pod \"09971522-31bc-4e10-8831-4fb62675e220\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.640683 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-host-cni-netd\") pod \"09971522-31bc-4e10-8831-4fb62675e220\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.640700 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-host-kubelet\") pod \"09971522-31bc-4e10-8831-4fb62675e220\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.640746 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/09971522-31bc-4e10-8831-4fb62675e220-env-overrides\") pod \"09971522-31bc-4e10-8831-4fb62675e220\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.640766 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bcd67\" (UniqueName: \"kubernetes.io/projected/09971522-31bc-4e10-8831-4fb62675e220-kube-api-access-bcd67\") pod \"09971522-31bc-4e10-8831-4fb62675e220\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.640837 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/09971522-31bc-4e10-8831-4fb62675e220-ovn-node-metrics-cert\") pod \"09971522-31bc-4e10-8831-4fb62675e220\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.640855 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-host-slash\") pod \"09971522-31bc-4e10-8831-4fb62675e220\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.640891 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-host-run-netns\") pod \"09971522-31bc-4e10-8831-4fb62675e220\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.640910 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-systemd-units\") pod \"09971522-31bc-4e10-8831-4fb62675e220\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.640929 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-etc-openvswitch\") pod \"09971522-31bc-4e10-8831-4fb62675e220\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.640966 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-run-openvswitch\") pod \"09971522-31bc-4e10-8831-4fb62675e220\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.641006 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/09971522-31bc-4e10-8831-4fb62675e220-ovnkube-script-lib\") pod \"09971522-31bc-4e10-8831-4fb62675e220\" (UID: \"09971522-31bc-4e10-8831-4fb62675e220\") " Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.640689 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "09971522-31bc-4e10-8831-4fb62675e220" (UID: "09971522-31bc-4e10-8831-4fb62675e220"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.640733 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-log-socket" (OuterVolumeSpecName: "log-socket") pod "09971522-31bc-4e10-8831-4fb62675e220" (UID: "09971522-31bc-4e10-8831-4fb62675e220"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.641358 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "09971522-31bc-4e10-8831-4fb62675e220" (UID: "09971522-31bc-4e10-8831-4fb62675e220"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.640752 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "09971522-31bc-4e10-8831-4fb62675e220" (UID: "09971522-31bc-4e10-8831-4fb62675e220"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.640779 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "09971522-31bc-4e10-8831-4fb62675e220" (UID: "09971522-31bc-4e10-8831-4fb62675e220"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.640940 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-host-slash" (OuterVolumeSpecName: "host-slash") pod "09971522-31bc-4e10-8831-4fb62675e220" (UID: "09971522-31bc-4e10-8831-4fb62675e220"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.640966 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "09971522-31bc-4e10-8831-4fb62675e220" (UID: "09971522-31bc-4e10-8831-4fb62675e220"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.641011 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "09971522-31bc-4e10-8831-4fb62675e220" (UID: "09971522-31bc-4e10-8831-4fb62675e220"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.641056 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "09971522-31bc-4e10-8831-4fb62675e220" (UID: "09971522-31bc-4e10-8831-4fb62675e220"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.641484 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09971522-31bc-4e10-8831-4fb62675e220-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "09971522-31bc-4e10-8831-4fb62675e220" (UID: "09971522-31bc-4e10-8831-4fb62675e220"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.641666 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1c39975c-67c8-47f3-bd1e-c2deb655d009-run-openvswitch\") pod \"ovnkube-node-wdg5l\" (UID: \"1c39975c-67c8-47f3-bd1e-c2deb655d009\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.641707 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1c39975c-67c8-47f3-bd1e-c2deb655d009-host-slash\") pod \"ovnkube-node-wdg5l\" (UID: \"1c39975c-67c8-47f3-bd1e-c2deb655d009\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.641749 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1c39975c-67c8-47f3-bd1e-c2deb655d009-ovn-node-metrics-cert\") pod \"ovnkube-node-wdg5l\" (UID: \"1c39975c-67c8-47f3-bd1e-c2deb655d009\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.641744 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09971522-31bc-4e10-8831-4fb62675e220-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "09971522-31bc-4e10-8831-4fb62675e220" (UID: "09971522-31bc-4e10-8831-4fb62675e220"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.641982 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1c39975c-67c8-47f3-bd1e-c2deb655d009-var-lib-openvswitch\") pod \"ovnkube-node-wdg5l\" (UID: \"1c39975c-67c8-47f3-bd1e-c2deb655d009\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.642025 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1c39975c-67c8-47f3-bd1e-c2deb655d009-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-wdg5l\" (UID: \"1c39975c-67c8-47f3-bd1e-c2deb655d009\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.642080 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1c39975c-67c8-47f3-bd1e-c2deb655d009-log-socket\") pod \"ovnkube-node-wdg5l\" (UID: \"1c39975c-67c8-47f3-bd1e-c2deb655d009\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.642107 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1c39975c-67c8-47f3-bd1e-c2deb655d009-run-ovn\") pod \"ovnkube-node-wdg5l\" (UID: \"1c39975c-67c8-47f3-bd1e-c2deb655d009\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.642174 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1c39975c-67c8-47f3-bd1e-c2deb655d009-systemd-units\") pod \"ovnkube-node-wdg5l\" (UID: \"1c39975c-67c8-47f3-bd1e-c2deb655d009\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.642194 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1c39975c-67c8-47f3-bd1e-c2deb655d009-host-kubelet\") pod \"ovnkube-node-wdg5l\" (UID: \"1c39975c-67c8-47f3-bd1e-c2deb655d009\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.642208 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1c39975c-67c8-47f3-bd1e-c2deb655d009-env-overrides\") pod \"ovnkube-node-wdg5l\" (UID: \"1c39975c-67c8-47f3-bd1e-c2deb655d009\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.642265 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1c39975c-67c8-47f3-bd1e-c2deb655d009-run-systemd\") pod \"ovnkube-node-wdg5l\" (UID: \"1c39975c-67c8-47f3-bd1e-c2deb655d009\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.642343 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1c39975c-67c8-47f3-bd1e-c2deb655d009-host-run-netns\") pod \"ovnkube-node-wdg5l\" (UID: \"1c39975c-67c8-47f3-bd1e-c2deb655d009\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.642405 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1c39975c-67c8-47f3-bd1e-c2deb655d009-ovnkube-config\") pod \"ovnkube-node-wdg5l\" (UID: \"1c39975c-67c8-47f3-bd1e-c2deb655d009\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.642447 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1c39975c-67c8-47f3-bd1e-c2deb655d009-node-log\") pod \"ovnkube-node-wdg5l\" (UID: \"1c39975c-67c8-47f3-bd1e-c2deb655d009\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.642474 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1c39975c-67c8-47f3-bd1e-c2deb655d009-etc-openvswitch\") pod \"ovnkube-node-wdg5l\" (UID: \"1c39975c-67c8-47f3-bd1e-c2deb655d009\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.642519 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1c39975c-67c8-47f3-bd1e-c2deb655d009-host-run-ovn-kubernetes\") pod \"ovnkube-node-wdg5l\" (UID: \"1c39975c-67c8-47f3-bd1e-c2deb655d009\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.642551 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1c39975c-67c8-47f3-bd1e-c2deb655d009-host-cni-netd\") pod \"ovnkube-node-wdg5l\" (UID: \"1c39975c-67c8-47f3-bd1e-c2deb655d009\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.642607 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1c39975c-67c8-47f3-bd1e-c2deb655d009-host-cni-bin\") pod \"ovnkube-node-wdg5l\" (UID: \"1c39975c-67c8-47f3-bd1e-c2deb655d009\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.642633 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rbkc\" (UniqueName: \"kubernetes.io/projected/1c39975c-67c8-47f3-bd1e-c2deb655d009-kube-api-access-9rbkc\") pod \"ovnkube-node-wdg5l\" (UID: \"1c39975c-67c8-47f3-bd1e-c2deb655d009\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.642724 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1c39975c-67c8-47f3-bd1e-c2deb655d009-ovnkube-script-lib\") pod \"ovnkube-node-wdg5l\" (UID: \"1c39975c-67c8-47f3-bd1e-c2deb655d009\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.642823 4556 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.642837 4556 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-node-log\") on node \"crc\" DevicePath \"\"" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.642848 4556 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/09971522-31bc-4e10-8831-4fb62675e220-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.642858 4556 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.642873 4556 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-log-socket\") on node \"crc\" DevicePath \"\"" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.642882 4556 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-host-kubelet\") on node \"crc\" DevicePath \"\"" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.642892 4556 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-host-cni-netd\") on node \"crc\" DevicePath \"\"" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.642934 4556 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/09971522-31bc-4e10-8831-4fb62675e220-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.642952 4556 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-host-slash\") on node \"crc\" DevicePath \"\"" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.642960 4556 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-host-run-netns\") on node \"crc\" DevicePath \"\"" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.642968 4556 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.642978 4556 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-systemd-units\") on node \"crc\" DevicePath \"\"" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.642986 4556 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-run-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.643006 4556 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.643015 4556 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-host-cni-bin\") on node \"crc\" DevicePath \"\"" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.643026 4556 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-run-ovn\") on node \"crc\" DevicePath \"\"" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.643684 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09971522-31bc-4e10-8831-4fb62675e220-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "09971522-31bc-4e10-8831-4fb62675e220" (UID: "09971522-31bc-4e10-8831-4fb62675e220"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.649823 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09971522-31bc-4e10-8831-4fb62675e220-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "09971522-31bc-4e10-8831-4fb62675e220" (UID: "09971522-31bc-4e10-8831-4fb62675e220"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.650405 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09971522-31bc-4e10-8831-4fb62675e220-kube-api-access-bcd67" (OuterVolumeSpecName: "kube-api-access-bcd67") pod "09971522-31bc-4e10-8831-4fb62675e220" (UID: "09971522-31bc-4e10-8831-4fb62675e220"). InnerVolumeSpecName "kube-api-access-bcd67". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.653144 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "09971522-31bc-4e10-8831-4fb62675e220" (UID: "09971522-31bc-4e10-8831-4fb62675e220"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.744580 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1c39975c-67c8-47f3-bd1e-c2deb655d009-host-cni-bin\") pod \"ovnkube-node-wdg5l\" (UID: \"1c39975c-67c8-47f3-bd1e-c2deb655d009\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.744622 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rbkc\" (UniqueName: \"kubernetes.io/projected/1c39975c-67c8-47f3-bd1e-c2deb655d009-kube-api-access-9rbkc\") pod \"ovnkube-node-wdg5l\" (UID: \"1c39975c-67c8-47f3-bd1e-c2deb655d009\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.744662 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1c39975c-67c8-47f3-bd1e-c2deb655d009-ovnkube-script-lib\") pod \"ovnkube-node-wdg5l\" (UID: \"1c39975c-67c8-47f3-bd1e-c2deb655d009\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.744680 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1c39975c-67c8-47f3-bd1e-c2deb655d009-run-openvswitch\") pod \"ovnkube-node-wdg5l\" (UID: \"1c39975c-67c8-47f3-bd1e-c2deb655d009\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.744697 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1c39975c-67c8-47f3-bd1e-c2deb655d009-host-slash\") pod \"ovnkube-node-wdg5l\" (UID: \"1c39975c-67c8-47f3-bd1e-c2deb655d009\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.744718 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1c39975c-67c8-47f3-bd1e-c2deb655d009-ovn-node-metrics-cert\") pod \"ovnkube-node-wdg5l\" (UID: \"1c39975c-67c8-47f3-bd1e-c2deb655d009\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.744718 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1c39975c-67c8-47f3-bd1e-c2deb655d009-host-cni-bin\") pod \"ovnkube-node-wdg5l\" (UID: \"1c39975c-67c8-47f3-bd1e-c2deb655d009\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.744737 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1c39975c-67c8-47f3-bd1e-c2deb655d009-var-lib-openvswitch\") pod \"ovnkube-node-wdg5l\" (UID: \"1c39975c-67c8-47f3-bd1e-c2deb655d009\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.744752 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1c39975c-67c8-47f3-bd1e-c2deb655d009-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-wdg5l\" (UID: \"1c39975c-67c8-47f3-bd1e-c2deb655d009\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.744777 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1c39975c-67c8-47f3-bd1e-c2deb655d009-log-socket\") pod \"ovnkube-node-wdg5l\" (UID: \"1c39975c-67c8-47f3-bd1e-c2deb655d009\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.744794 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1c39975c-67c8-47f3-bd1e-c2deb655d009-run-ovn\") pod \"ovnkube-node-wdg5l\" (UID: \"1c39975c-67c8-47f3-bd1e-c2deb655d009\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.744794 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1c39975c-67c8-47f3-bd1e-c2deb655d009-host-slash\") pod \"ovnkube-node-wdg5l\" (UID: \"1c39975c-67c8-47f3-bd1e-c2deb655d009\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.744830 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1c39975c-67c8-47f3-bd1e-c2deb655d009-systemd-units\") pod \"ovnkube-node-wdg5l\" (UID: \"1c39975c-67c8-47f3-bd1e-c2deb655d009\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.744810 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1c39975c-67c8-47f3-bd1e-c2deb655d009-systemd-units\") pod \"ovnkube-node-wdg5l\" (UID: \"1c39975c-67c8-47f3-bd1e-c2deb655d009\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.744858 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1c39975c-67c8-47f3-bd1e-c2deb655d009-var-lib-openvswitch\") pod \"ovnkube-node-wdg5l\" (UID: \"1c39975c-67c8-47f3-bd1e-c2deb655d009\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.744879 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1c39975c-67c8-47f3-bd1e-c2deb655d009-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-wdg5l\" (UID: \"1c39975c-67c8-47f3-bd1e-c2deb655d009\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.744881 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1c39975c-67c8-47f3-bd1e-c2deb655d009-host-kubelet\") pod \"ovnkube-node-wdg5l\" (UID: \"1c39975c-67c8-47f3-bd1e-c2deb655d009\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.744897 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1c39975c-67c8-47f3-bd1e-c2deb655d009-log-socket\") pod \"ovnkube-node-wdg5l\" (UID: \"1c39975c-67c8-47f3-bd1e-c2deb655d009\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.744906 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1c39975c-67c8-47f3-bd1e-c2deb655d009-env-overrides\") pod \"ovnkube-node-wdg5l\" (UID: \"1c39975c-67c8-47f3-bd1e-c2deb655d009\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.744917 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1c39975c-67c8-47f3-bd1e-c2deb655d009-run-ovn\") pod \"ovnkube-node-wdg5l\" (UID: \"1c39975c-67c8-47f3-bd1e-c2deb655d009\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.744928 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1c39975c-67c8-47f3-bd1e-c2deb655d009-run-systemd\") pod \"ovnkube-node-wdg5l\" (UID: \"1c39975c-67c8-47f3-bd1e-c2deb655d009\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.744937 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1c39975c-67c8-47f3-bd1e-c2deb655d009-host-kubelet\") pod \"ovnkube-node-wdg5l\" (UID: \"1c39975c-67c8-47f3-bd1e-c2deb655d009\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.744958 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1c39975c-67c8-47f3-bd1e-c2deb655d009-run-openvswitch\") pod \"ovnkube-node-wdg5l\" (UID: \"1c39975c-67c8-47f3-bd1e-c2deb655d009\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.744958 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1c39975c-67c8-47f3-bd1e-c2deb655d009-host-run-netns\") pod \"ovnkube-node-wdg5l\" (UID: \"1c39975c-67c8-47f3-bd1e-c2deb655d009\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.744980 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1c39975c-67c8-47f3-bd1e-c2deb655d009-host-run-netns\") pod \"ovnkube-node-wdg5l\" (UID: \"1c39975c-67c8-47f3-bd1e-c2deb655d009\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.744990 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1c39975c-67c8-47f3-bd1e-c2deb655d009-ovnkube-config\") pod \"ovnkube-node-wdg5l\" (UID: \"1c39975c-67c8-47f3-bd1e-c2deb655d009\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.745025 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1c39975c-67c8-47f3-bd1e-c2deb655d009-node-log\") pod \"ovnkube-node-wdg5l\" (UID: \"1c39975c-67c8-47f3-bd1e-c2deb655d009\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.745038 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1c39975c-67c8-47f3-bd1e-c2deb655d009-etc-openvswitch\") pod \"ovnkube-node-wdg5l\" (UID: \"1c39975c-67c8-47f3-bd1e-c2deb655d009\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.745043 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1c39975c-67c8-47f3-bd1e-c2deb655d009-run-systemd\") pod \"ovnkube-node-wdg5l\" (UID: \"1c39975c-67c8-47f3-bd1e-c2deb655d009\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.745058 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1c39975c-67c8-47f3-bd1e-c2deb655d009-host-run-ovn-kubernetes\") pod \"ovnkube-node-wdg5l\" (UID: \"1c39975c-67c8-47f3-bd1e-c2deb655d009\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.745079 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1c39975c-67c8-47f3-bd1e-c2deb655d009-host-cni-netd\") pod \"ovnkube-node-wdg5l\" (UID: \"1c39975c-67c8-47f3-bd1e-c2deb655d009\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.745085 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1c39975c-67c8-47f3-bd1e-c2deb655d009-node-log\") pod \"ovnkube-node-wdg5l\" (UID: \"1c39975c-67c8-47f3-bd1e-c2deb655d009\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.745125 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1c39975c-67c8-47f3-bd1e-c2deb655d009-host-run-ovn-kubernetes\") pod \"ovnkube-node-wdg5l\" (UID: \"1c39975c-67c8-47f3-bd1e-c2deb655d009\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.745130 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1c39975c-67c8-47f3-bd1e-c2deb655d009-host-cni-netd\") pod \"ovnkube-node-wdg5l\" (UID: \"1c39975c-67c8-47f3-bd1e-c2deb655d009\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.745249 4556 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/09971522-31bc-4e10-8831-4fb62675e220-run-systemd\") on node \"crc\" DevicePath \"\"" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.745270 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1c39975c-67c8-47f3-bd1e-c2deb655d009-etc-openvswitch\") pod \"ovnkube-node-wdg5l\" (UID: \"1c39975c-67c8-47f3-bd1e-c2deb655d009\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.745285 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bcd67\" (UniqueName: \"kubernetes.io/projected/09971522-31bc-4e10-8831-4fb62675e220-kube-api-access-bcd67\") on node \"crc\" DevicePath \"\"" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.745294 4556 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/09971522-31bc-4e10-8831-4fb62675e220-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.745303 4556 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/09971522-31bc-4e10-8831-4fb62675e220-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.745465 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1c39975c-67c8-47f3-bd1e-c2deb655d009-env-overrides\") pod \"ovnkube-node-wdg5l\" (UID: \"1c39975c-67c8-47f3-bd1e-c2deb655d009\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.745568 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1c39975c-67c8-47f3-bd1e-c2deb655d009-ovnkube-script-lib\") pod \"ovnkube-node-wdg5l\" (UID: \"1c39975c-67c8-47f3-bd1e-c2deb655d009\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.745784 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1c39975c-67c8-47f3-bd1e-c2deb655d009-ovnkube-config\") pod \"ovnkube-node-wdg5l\" (UID: \"1c39975c-67c8-47f3-bd1e-c2deb655d009\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.748058 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1c39975c-67c8-47f3-bd1e-c2deb655d009-ovn-node-metrics-cert\") pod \"ovnkube-node-wdg5l\" (UID: \"1c39975c-67c8-47f3-bd1e-c2deb655d009\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.757487 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rbkc\" (UniqueName: \"kubernetes.io/projected/1c39975c-67c8-47f3-bd1e-c2deb655d009-kube-api-access-9rbkc\") pod \"ovnkube-node-wdg5l\" (UID: \"1c39975c-67c8-47f3-bd1e-c2deb655d009\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: I0218 09:11:58.924126 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:11:58 crc kubenswrapper[4556]: W0218 09:11:58.939507 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1c39975c_67c8_47f3_bd1e_c2deb655d009.slice/crio-dcec54834142487c0aecfadab71b699a6d9f1bc4a48c8950fce0593349883873 WatchSource:0}: Error finding container dcec54834142487c0aecfadab71b699a6d9f1bc4a48c8950fce0593349883873: Status 404 returned error can't find the container with id dcec54834142487c0aecfadab71b699a6d9f1bc4a48c8950fce0593349883873 Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.425337 4556 generic.go:334] "Generic (PLEG): container finished" podID="1c39975c-67c8-47f3-bd1e-c2deb655d009" containerID="f423d7e8f74c6e992035e11fc0a22626b5baa01b5b4782ecccc28ad43077be10" exitCode=0 Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.425415 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" event={"ID":"1c39975c-67c8-47f3-bd1e-c2deb655d009","Type":"ContainerDied","Data":"f423d7e8f74c6e992035e11fc0a22626b5baa01b5b4782ecccc28ad43077be10"} Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.425442 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" event={"ID":"1c39975c-67c8-47f3-bd1e-c2deb655d009","Type":"ContainerStarted","Data":"dcec54834142487c0aecfadab71b699a6d9f1bc4a48c8950fce0593349883873"} Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.428076 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-qfksk_e226b189-cfa9-47fc-b94d-19b5cbe0859f/kube-multus/1.log" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.433933 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-b7fsj_09971522-31bc-4e10-8831-4fb62675e220/ovnkube-controller/3.log" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.436318 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-b7fsj_09971522-31bc-4e10-8831-4fb62675e220/ovn-acl-logging/0.log" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.436803 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-b7fsj_09971522-31bc-4e10-8831-4fb62675e220/ovn-controller/0.log" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.437326 4556 generic.go:334] "Generic (PLEG): container finished" podID="09971522-31bc-4e10-8831-4fb62675e220" containerID="ad37b48f74a075b5562885db93abefd92df9f91f81a3b2ff685ee97a13b5e403" exitCode=0 Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.437350 4556 generic.go:334] "Generic (PLEG): container finished" podID="09971522-31bc-4e10-8831-4fb62675e220" containerID="aec6030df0a59ad8902d5c694d5cd30d249ca1e895155c57a458e359fa531b7c" exitCode=0 Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.437359 4556 generic.go:334] "Generic (PLEG): container finished" podID="09971522-31bc-4e10-8831-4fb62675e220" containerID="4cd6e758ee396c9000b1bdb6ded2176ec803fa9da6201cdd9b8146940a2c00cb" exitCode=0 Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.437367 4556 generic.go:334] "Generic (PLEG): container finished" podID="09971522-31bc-4e10-8831-4fb62675e220" containerID="9b0412f3330999c3348ac726ac5a0c2d541571f4b13786ca24cb3109a44aa653" exitCode=0 Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.437374 4556 generic.go:334] "Generic (PLEG): container finished" podID="09971522-31bc-4e10-8831-4fb62675e220" containerID="01c526c0613f19e1a1a584b7129bd56001faf08d0b97fd77feeb69f764ed01f8" exitCode=0 Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.437381 4556 generic.go:334] "Generic (PLEG): container finished" podID="09971522-31bc-4e10-8831-4fb62675e220" containerID="9170bdcbaf1065995d03e6448e40bdced0741c716c9e81f37a72b015132f1161" exitCode=0 Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.437388 4556 generic.go:334] "Generic (PLEG): container finished" podID="09971522-31bc-4e10-8831-4fb62675e220" containerID="318cb69e50df4c14df9bf5ef57d5c381421a9bdb72135c9b9620f55c0547b410" exitCode=143 Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.437394 4556 generic.go:334] "Generic (PLEG): container finished" podID="09971522-31bc-4e10-8831-4fb62675e220" containerID="0c0f15d331af9bc2ca3082c1573dd167c66773eb390593844e681f8d0b455d3d" exitCode=143 Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.437410 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" event={"ID":"09971522-31bc-4e10-8831-4fb62675e220","Type":"ContainerDied","Data":"ad37b48f74a075b5562885db93abefd92df9f91f81a3b2ff685ee97a13b5e403"} Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.437429 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" event={"ID":"09971522-31bc-4e10-8831-4fb62675e220","Type":"ContainerDied","Data":"aec6030df0a59ad8902d5c694d5cd30d249ca1e895155c57a458e359fa531b7c"} Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.437440 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" event={"ID":"09971522-31bc-4e10-8831-4fb62675e220","Type":"ContainerDied","Data":"4cd6e758ee396c9000b1bdb6ded2176ec803fa9da6201cdd9b8146940a2c00cb"} Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.437449 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" event={"ID":"09971522-31bc-4e10-8831-4fb62675e220","Type":"ContainerDied","Data":"9b0412f3330999c3348ac726ac5a0c2d541571f4b13786ca24cb3109a44aa653"} Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.437457 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" event={"ID":"09971522-31bc-4e10-8831-4fb62675e220","Type":"ContainerDied","Data":"01c526c0613f19e1a1a584b7129bd56001faf08d0b97fd77feeb69f764ed01f8"} Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.437465 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" event={"ID":"09971522-31bc-4e10-8831-4fb62675e220","Type":"ContainerDied","Data":"9170bdcbaf1065995d03e6448e40bdced0741c716c9e81f37a72b015132f1161"} Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.437474 4556 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c2f16fe4bf2d659398ec141ba12d2dda32727810b03b96bfaf491602656027cb"} Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.437485 4556 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aec6030df0a59ad8902d5c694d5cd30d249ca1e895155c57a458e359fa531b7c"} Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.437489 4556 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4cd6e758ee396c9000b1bdb6ded2176ec803fa9da6201cdd9b8146940a2c00cb"} Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.437494 4556 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9b0412f3330999c3348ac726ac5a0c2d541571f4b13786ca24cb3109a44aa653"} Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.437499 4556 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"01c526c0613f19e1a1a584b7129bd56001faf08d0b97fd77feeb69f764ed01f8"} Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.437503 4556 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9170bdcbaf1065995d03e6448e40bdced0741c716c9e81f37a72b015132f1161"} Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.437508 4556 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"318cb69e50df4c14df9bf5ef57d5c381421a9bdb72135c9b9620f55c0547b410"} Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.437512 4556 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0c0f15d331af9bc2ca3082c1573dd167c66773eb390593844e681f8d0b455d3d"} Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.437517 4556 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e"} Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.437523 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" event={"ID":"09971522-31bc-4e10-8831-4fb62675e220","Type":"ContainerDied","Data":"318cb69e50df4c14df9bf5ef57d5c381421a9bdb72135c9b9620f55c0547b410"} Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.437531 4556 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ad37b48f74a075b5562885db93abefd92df9f91f81a3b2ff685ee97a13b5e403"} Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.437536 4556 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c2f16fe4bf2d659398ec141ba12d2dda32727810b03b96bfaf491602656027cb"} Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.437540 4556 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aec6030df0a59ad8902d5c694d5cd30d249ca1e895155c57a458e359fa531b7c"} Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.437545 4556 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4cd6e758ee396c9000b1bdb6ded2176ec803fa9da6201cdd9b8146940a2c00cb"} Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.437551 4556 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9b0412f3330999c3348ac726ac5a0c2d541571f4b13786ca24cb3109a44aa653"} Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.437555 4556 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"01c526c0613f19e1a1a584b7129bd56001faf08d0b97fd77feeb69f764ed01f8"} Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.437560 4556 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9170bdcbaf1065995d03e6448e40bdced0741c716c9e81f37a72b015132f1161"} Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.437565 4556 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"318cb69e50df4c14df9bf5ef57d5c381421a9bdb72135c9b9620f55c0547b410"} Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.437569 4556 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0c0f15d331af9bc2ca3082c1573dd167c66773eb390593844e681f8d0b455d3d"} Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.437574 4556 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e"} Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.437580 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" event={"ID":"09971522-31bc-4e10-8831-4fb62675e220","Type":"ContainerDied","Data":"0c0f15d331af9bc2ca3082c1573dd167c66773eb390593844e681f8d0b455d3d"} Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.437587 4556 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ad37b48f74a075b5562885db93abefd92df9f91f81a3b2ff685ee97a13b5e403"} Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.437599 4556 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c2f16fe4bf2d659398ec141ba12d2dda32727810b03b96bfaf491602656027cb"} Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.437604 4556 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aec6030df0a59ad8902d5c694d5cd30d249ca1e895155c57a458e359fa531b7c"} Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.437609 4556 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4cd6e758ee396c9000b1bdb6ded2176ec803fa9da6201cdd9b8146940a2c00cb"} Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.437615 4556 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9b0412f3330999c3348ac726ac5a0c2d541571f4b13786ca24cb3109a44aa653"} Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.437619 4556 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"01c526c0613f19e1a1a584b7129bd56001faf08d0b97fd77feeb69f764ed01f8"} Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.437624 4556 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9170bdcbaf1065995d03e6448e40bdced0741c716c9e81f37a72b015132f1161"} Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.437629 4556 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"318cb69e50df4c14df9bf5ef57d5c381421a9bdb72135c9b9620f55c0547b410"} Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.437633 4556 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0c0f15d331af9bc2ca3082c1573dd167c66773eb390593844e681f8d0b455d3d"} Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.437639 4556 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e"} Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.437645 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" event={"ID":"09971522-31bc-4e10-8831-4fb62675e220","Type":"ContainerDied","Data":"67f192b1fd186997f2fc73bbdf9672b329d777c219cb80c49cfef1f0e2fbfebc"} Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.437655 4556 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ad37b48f74a075b5562885db93abefd92df9f91f81a3b2ff685ee97a13b5e403"} Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.437660 4556 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c2f16fe4bf2d659398ec141ba12d2dda32727810b03b96bfaf491602656027cb"} Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.437664 4556 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aec6030df0a59ad8902d5c694d5cd30d249ca1e895155c57a458e359fa531b7c"} Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.437669 4556 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4cd6e758ee396c9000b1bdb6ded2176ec803fa9da6201cdd9b8146940a2c00cb"} Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.437673 4556 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9b0412f3330999c3348ac726ac5a0c2d541571f4b13786ca24cb3109a44aa653"} Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.437678 4556 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"01c526c0613f19e1a1a584b7129bd56001faf08d0b97fd77feeb69f764ed01f8"} Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.437682 4556 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9170bdcbaf1065995d03e6448e40bdced0741c716c9e81f37a72b015132f1161"} Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.437687 4556 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"318cb69e50df4c14df9bf5ef57d5c381421a9bdb72135c9b9620f55c0547b410"} Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.437692 4556 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0c0f15d331af9bc2ca3082c1573dd167c66773eb390593844e681f8d0b455d3d"} Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.437696 4556 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e"} Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.437708 4556 scope.go:117] "RemoveContainer" containerID="ad37b48f74a075b5562885db93abefd92df9f91f81a3b2ff685ee97a13b5e403" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.437837 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-b7fsj" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.470346 4556 scope.go:117] "RemoveContainer" containerID="c2f16fe4bf2d659398ec141ba12d2dda32727810b03b96bfaf491602656027cb" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.478282 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-b7fsj"] Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.480686 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-b7fsj"] Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.487186 4556 scope.go:117] "RemoveContainer" containerID="aec6030df0a59ad8902d5c694d5cd30d249ca1e895155c57a458e359fa531b7c" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.500693 4556 scope.go:117] "RemoveContainer" containerID="4cd6e758ee396c9000b1bdb6ded2176ec803fa9da6201cdd9b8146940a2c00cb" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.513210 4556 scope.go:117] "RemoveContainer" containerID="9b0412f3330999c3348ac726ac5a0c2d541571f4b13786ca24cb3109a44aa653" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.523043 4556 scope.go:117] "RemoveContainer" containerID="01c526c0613f19e1a1a584b7129bd56001faf08d0b97fd77feeb69f764ed01f8" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.533597 4556 scope.go:117] "RemoveContainer" containerID="9170bdcbaf1065995d03e6448e40bdced0741c716c9e81f37a72b015132f1161" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.543909 4556 scope.go:117] "RemoveContainer" containerID="318cb69e50df4c14df9bf5ef57d5c381421a9bdb72135c9b9620f55c0547b410" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.554887 4556 scope.go:117] "RemoveContainer" containerID="0c0f15d331af9bc2ca3082c1573dd167c66773eb390593844e681f8d0b455d3d" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.574273 4556 scope.go:117] "RemoveContainer" containerID="323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.594782 4556 scope.go:117] "RemoveContainer" containerID="ad37b48f74a075b5562885db93abefd92df9f91f81a3b2ff685ee97a13b5e403" Feb 18 09:11:59 crc kubenswrapper[4556]: E0218 09:11:59.595116 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad37b48f74a075b5562885db93abefd92df9f91f81a3b2ff685ee97a13b5e403\": container with ID starting with ad37b48f74a075b5562885db93abefd92df9f91f81a3b2ff685ee97a13b5e403 not found: ID does not exist" containerID="ad37b48f74a075b5562885db93abefd92df9f91f81a3b2ff685ee97a13b5e403" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.595178 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad37b48f74a075b5562885db93abefd92df9f91f81a3b2ff685ee97a13b5e403"} err="failed to get container status \"ad37b48f74a075b5562885db93abefd92df9f91f81a3b2ff685ee97a13b5e403\": rpc error: code = NotFound desc = could not find container \"ad37b48f74a075b5562885db93abefd92df9f91f81a3b2ff685ee97a13b5e403\": container with ID starting with ad37b48f74a075b5562885db93abefd92df9f91f81a3b2ff685ee97a13b5e403 not found: ID does not exist" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.595205 4556 scope.go:117] "RemoveContainer" containerID="c2f16fe4bf2d659398ec141ba12d2dda32727810b03b96bfaf491602656027cb" Feb 18 09:11:59 crc kubenswrapper[4556]: E0218 09:11:59.595484 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2f16fe4bf2d659398ec141ba12d2dda32727810b03b96bfaf491602656027cb\": container with ID starting with c2f16fe4bf2d659398ec141ba12d2dda32727810b03b96bfaf491602656027cb not found: ID does not exist" containerID="c2f16fe4bf2d659398ec141ba12d2dda32727810b03b96bfaf491602656027cb" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.595511 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2f16fe4bf2d659398ec141ba12d2dda32727810b03b96bfaf491602656027cb"} err="failed to get container status \"c2f16fe4bf2d659398ec141ba12d2dda32727810b03b96bfaf491602656027cb\": rpc error: code = NotFound desc = could not find container \"c2f16fe4bf2d659398ec141ba12d2dda32727810b03b96bfaf491602656027cb\": container with ID starting with c2f16fe4bf2d659398ec141ba12d2dda32727810b03b96bfaf491602656027cb not found: ID does not exist" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.595531 4556 scope.go:117] "RemoveContainer" containerID="aec6030df0a59ad8902d5c694d5cd30d249ca1e895155c57a458e359fa531b7c" Feb 18 09:11:59 crc kubenswrapper[4556]: E0218 09:11:59.595737 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aec6030df0a59ad8902d5c694d5cd30d249ca1e895155c57a458e359fa531b7c\": container with ID starting with aec6030df0a59ad8902d5c694d5cd30d249ca1e895155c57a458e359fa531b7c not found: ID does not exist" containerID="aec6030df0a59ad8902d5c694d5cd30d249ca1e895155c57a458e359fa531b7c" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.595777 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aec6030df0a59ad8902d5c694d5cd30d249ca1e895155c57a458e359fa531b7c"} err="failed to get container status \"aec6030df0a59ad8902d5c694d5cd30d249ca1e895155c57a458e359fa531b7c\": rpc error: code = NotFound desc = could not find container \"aec6030df0a59ad8902d5c694d5cd30d249ca1e895155c57a458e359fa531b7c\": container with ID starting with aec6030df0a59ad8902d5c694d5cd30d249ca1e895155c57a458e359fa531b7c not found: ID does not exist" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.595793 4556 scope.go:117] "RemoveContainer" containerID="4cd6e758ee396c9000b1bdb6ded2176ec803fa9da6201cdd9b8146940a2c00cb" Feb 18 09:11:59 crc kubenswrapper[4556]: E0218 09:11:59.595980 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4cd6e758ee396c9000b1bdb6ded2176ec803fa9da6201cdd9b8146940a2c00cb\": container with ID starting with 4cd6e758ee396c9000b1bdb6ded2176ec803fa9da6201cdd9b8146940a2c00cb not found: ID does not exist" containerID="4cd6e758ee396c9000b1bdb6ded2176ec803fa9da6201cdd9b8146940a2c00cb" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.596011 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4cd6e758ee396c9000b1bdb6ded2176ec803fa9da6201cdd9b8146940a2c00cb"} err="failed to get container status \"4cd6e758ee396c9000b1bdb6ded2176ec803fa9da6201cdd9b8146940a2c00cb\": rpc error: code = NotFound desc = could not find container \"4cd6e758ee396c9000b1bdb6ded2176ec803fa9da6201cdd9b8146940a2c00cb\": container with ID starting with 4cd6e758ee396c9000b1bdb6ded2176ec803fa9da6201cdd9b8146940a2c00cb not found: ID does not exist" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.596024 4556 scope.go:117] "RemoveContainer" containerID="9b0412f3330999c3348ac726ac5a0c2d541571f4b13786ca24cb3109a44aa653" Feb 18 09:11:59 crc kubenswrapper[4556]: E0218 09:11:59.596232 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b0412f3330999c3348ac726ac5a0c2d541571f4b13786ca24cb3109a44aa653\": container with ID starting with 9b0412f3330999c3348ac726ac5a0c2d541571f4b13786ca24cb3109a44aa653 not found: ID does not exist" containerID="9b0412f3330999c3348ac726ac5a0c2d541571f4b13786ca24cb3109a44aa653" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.596267 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b0412f3330999c3348ac726ac5a0c2d541571f4b13786ca24cb3109a44aa653"} err="failed to get container status \"9b0412f3330999c3348ac726ac5a0c2d541571f4b13786ca24cb3109a44aa653\": rpc error: code = NotFound desc = could not find container \"9b0412f3330999c3348ac726ac5a0c2d541571f4b13786ca24cb3109a44aa653\": container with ID starting with 9b0412f3330999c3348ac726ac5a0c2d541571f4b13786ca24cb3109a44aa653 not found: ID does not exist" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.596291 4556 scope.go:117] "RemoveContainer" containerID="01c526c0613f19e1a1a584b7129bd56001faf08d0b97fd77feeb69f764ed01f8" Feb 18 09:11:59 crc kubenswrapper[4556]: E0218 09:11:59.596475 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01c526c0613f19e1a1a584b7129bd56001faf08d0b97fd77feeb69f764ed01f8\": container with ID starting with 01c526c0613f19e1a1a584b7129bd56001faf08d0b97fd77feeb69f764ed01f8 not found: ID does not exist" containerID="01c526c0613f19e1a1a584b7129bd56001faf08d0b97fd77feeb69f764ed01f8" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.596498 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01c526c0613f19e1a1a584b7129bd56001faf08d0b97fd77feeb69f764ed01f8"} err="failed to get container status \"01c526c0613f19e1a1a584b7129bd56001faf08d0b97fd77feeb69f764ed01f8\": rpc error: code = NotFound desc = could not find container \"01c526c0613f19e1a1a584b7129bd56001faf08d0b97fd77feeb69f764ed01f8\": container with ID starting with 01c526c0613f19e1a1a584b7129bd56001faf08d0b97fd77feeb69f764ed01f8 not found: ID does not exist" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.596512 4556 scope.go:117] "RemoveContainer" containerID="9170bdcbaf1065995d03e6448e40bdced0741c716c9e81f37a72b015132f1161" Feb 18 09:11:59 crc kubenswrapper[4556]: E0218 09:11:59.596681 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9170bdcbaf1065995d03e6448e40bdced0741c716c9e81f37a72b015132f1161\": container with ID starting with 9170bdcbaf1065995d03e6448e40bdced0741c716c9e81f37a72b015132f1161 not found: ID does not exist" containerID="9170bdcbaf1065995d03e6448e40bdced0741c716c9e81f37a72b015132f1161" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.596701 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9170bdcbaf1065995d03e6448e40bdced0741c716c9e81f37a72b015132f1161"} err="failed to get container status \"9170bdcbaf1065995d03e6448e40bdced0741c716c9e81f37a72b015132f1161\": rpc error: code = NotFound desc = could not find container \"9170bdcbaf1065995d03e6448e40bdced0741c716c9e81f37a72b015132f1161\": container with ID starting with 9170bdcbaf1065995d03e6448e40bdced0741c716c9e81f37a72b015132f1161 not found: ID does not exist" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.596716 4556 scope.go:117] "RemoveContainer" containerID="318cb69e50df4c14df9bf5ef57d5c381421a9bdb72135c9b9620f55c0547b410" Feb 18 09:11:59 crc kubenswrapper[4556]: E0218 09:11:59.596865 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"318cb69e50df4c14df9bf5ef57d5c381421a9bdb72135c9b9620f55c0547b410\": container with ID starting with 318cb69e50df4c14df9bf5ef57d5c381421a9bdb72135c9b9620f55c0547b410 not found: ID does not exist" containerID="318cb69e50df4c14df9bf5ef57d5c381421a9bdb72135c9b9620f55c0547b410" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.596884 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"318cb69e50df4c14df9bf5ef57d5c381421a9bdb72135c9b9620f55c0547b410"} err="failed to get container status \"318cb69e50df4c14df9bf5ef57d5c381421a9bdb72135c9b9620f55c0547b410\": rpc error: code = NotFound desc = could not find container \"318cb69e50df4c14df9bf5ef57d5c381421a9bdb72135c9b9620f55c0547b410\": container with ID starting with 318cb69e50df4c14df9bf5ef57d5c381421a9bdb72135c9b9620f55c0547b410 not found: ID does not exist" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.596898 4556 scope.go:117] "RemoveContainer" containerID="0c0f15d331af9bc2ca3082c1573dd167c66773eb390593844e681f8d0b455d3d" Feb 18 09:11:59 crc kubenswrapper[4556]: E0218 09:11:59.597064 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c0f15d331af9bc2ca3082c1573dd167c66773eb390593844e681f8d0b455d3d\": container with ID starting with 0c0f15d331af9bc2ca3082c1573dd167c66773eb390593844e681f8d0b455d3d not found: ID does not exist" containerID="0c0f15d331af9bc2ca3082c1573dd167c66773eb390593844e681f8d0b455d3d" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.597090 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c0f15d331af9bc2ca3082c1573dd167c66773eb390593844e681f8d0b455d3d"} err="failed to get container status \"0c0f15d331af9bc2ca3082c1573dd167c66773eb390593844e681f8d0b455d3d\": rpc error: code = NotFound desc = could not find container \"0c0f15d331af9bc2ca3082c1573dd167c66773eb390593844e681f8d0b455d3d\": container with ID starting with 0c0f15d331af9bc2ca3082c1573dd167c66773eb390593844e681f8d0b455d3d not found: ID does not exist" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.597103 4556 scope.go:117] "RemoveContainer" containerID="323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e" Feb 18 09:11:59 crc kubenswrapper[4556]: E0218 09:11:59.597389 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\": container with ID starting with 323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e not found: ID does not exist" containerID="323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.597411 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e"} err="failed to get container status \"323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\": rpc error: code = NotFound desc = could not find container \"323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\": container with ID starting with 323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e not found: ID does not exist" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.597423 4556 scope.go:117] "RemoveContainer" containerID="ad37b48f74a075b5562885db93abefd92df9f91f81a3b2ff685ee97a13b5e403" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.597566 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad37b48f74a075b5562885db93abefd92df9f91f81a3b2ff685ee97a13b5e403"} err="failed to get container status \"ad37b48f74a075b5562885db93abefd92df9f91f81a3b2ff685ee97a13b5e403\": rpc error: code = NotFound desc = could not find container \"ad37b48f74a075b5562885db93abefd92df9f91f81a3b2ff685ee97a13b5e403\": container with ID starting with ad37b48f74a075b5562885db93abefd92df9f91f81a3b2ff685ee97a13b5e403 not found: ID does not exist" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.597583 4556 scope.go:117] "RemoveContainer" containerID="c2f16fe4bf2d659398ec141ba12d2dda32727810b03b96bfaf491602656027cb" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.597726 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2f16fe4bf2d659398ec141ba12d2dda32727810b03b96bfaf491602656027cb"} err="failed to get container status \"c2f16fe4bf2d659398ec141ba12d2dda32727810b03b96bfaf491602656027cb\": rpc error: code = NotFound desc = could not find container \"c2f16fe4bf2d659398ec141ba12d2dda32727810b03b96bfaf491602656027cb\": container with ID starting with c2f16fe4bf2d659398ec141ba12d2dda32727810b03b96bfaf491602656027cb not found: ID does not exist" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.597749 4556 scope.go:117] "RemoveContainer" containerID="aec6030df0a59ad8902d5c694d5cd30d249ca1e895155c57a458e359fa531b7c" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.597879 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aec6030df0a59ad8902d5c694d5cd30d249ca1e895155c57a458e359fa531b7c"} err="failed to get container status \"aec6030df0a59ad8902d5c694d5cd30d249ca1e895155c57a458e359fa531b7c\": rpc error: code = NotFound desc = could not find container \"aec6030df0a59ad8902d5c694d5cd30d249ca1e895155c57a458e359fa531b7c\": container with ID starting with aec6030df0a59ad8902d5c694d5cd30d249ca1e895155c57a458e359fa531b7c not found: ID does not exist" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.597892 4556 scope.go:117] "RemoveContainer" containerID="4cd6e758ee396c9000b1bdb6ded2176ec803fa9da6201cdd9b8146940a2c00cb" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.598040 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4cd6e758ee396c9000b1bdb6ded2176ec803fa9da6201cdd9b8146940a2c00cb"} err="failed to get container status \"4cd6e758ee396c9000b1bdb6ded2176ec803fa9da6201cdd9b8146940a2c00cb\": rpc error: code = NotFound desc = could not find container \"4cd6e758ee396c9000b1bdb6ded2176ec803fa9da6201cdd9b8146940a2c00cb\": container with ID starting with 4cd6e758ee396c9000b1bdb6ded2176ec803fa9da6201cdd9b8146940a2c00cb not found: ID does not exist" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.598055 4556 scope.go:117] "RemoveContainer" containerID="9b0412f3330999c3348ac726ac5a0c2d541571f4b13786ca24cb3109a44aa653" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.598200 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b0412f3330999c3348ac726ac5a0c2d541571f4b13786ca24cb3109a44aa653"} err="failed to get container status \"9b0412f3330999c3348ac726ac5a0c2d541571f4b13786ca24cb3109a44aa653\": rpc error: code = NotFound desc = could not find container \"9b0412f3330999c3348ac726ac5a0c2d541571f4b13786ca24cb3109a44aa653\": container with ID starting with 9b0412f3330999c3348ac726ac5a0c2d541571f4b13786ca24cb3109a44aa653 not found: ID does not exist" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.598217 4556 scope.go:117] "RemoveContainer" containerID="01c526c0613f19e1a1a584b7129bd56001faf08d0b97fd77feeb69f764ed01f8" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.598359 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01c526c0613f19e1a1a584b7129bd56001faf08d0b97fd77feeb69f764ed01f8"} err="failed to get container status \"01c526c0613f19e1a1a584b7129bd56001faf08d0b97fd77feeb69f764ed01f8\": rpc error: code = NotFound desc = could not find container \"01c526c0613f19e1a1a584b7129bd56001faf08d0b97fd77feeb69f764ed01f8\": container with ID starting with 01c526c0613f19e1a1a584b7129bd56001faf08d0b97fd77feeb69f764ed01f8 not found: ID does not exist" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.598374 4556 scope.go:117] "RemoveContainer" containerID="9170bdcbaf1065995d03e6448e40bdced0741c716c9e81f37a72b015132f1161" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.598513 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9170bdcbaf1065995d03e6448e40bdced0741c716c9e81f37a72b015132f1161"} err="failed to get container status \"9170bdcbaf1065995d03e6448e40bdced0741c716c9e81f37a72b015132f1161\": rpc error: code = NotFound desc = could not find container \"9170bdcbaf1065995d03e6448e40bdced0741c716c9e81f37a72b015132f1161\": container with ID starting with 9170bdcbaf1065995d03e6448e40bdced0741c716c9e81f37a72b015132f1161 not found: ID does not exist" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.598528 4556 scope.go:117] "RemoveContainer" containerID="318cb69e50df4c14df9bf5ef57d5c381421a9bdb72135c9b9620f55c0547b410" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.598667 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"318cb69e50df4c14df9bf5ef57d5c381421a9bdb72135c9b9620f55c0547b410"} err="failed to get container status \"318cb69e50df4c14df9bf5ef57d5c381421a9bdb72135c9b9620f55c0547b410\": rpc error: code = NotFound desc = could not find container \"318cb69e50df4c14df9bf5ef57d5c381421a9bdb72135c9b9620f55c0547b410\": container with ID starting with 318cb69e50df4c14df9bf5ef57d5c381421a9bdb72135c9b9620f55c0547b410 not found: ID does not exist" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.598682 4556 scope.go:117] "RemoveContainer" containerID="0c0f15d331af9bc2ca3082c1573dd167c66773eb390593844e681f8d0b455d3d" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.598836 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c0f15d331af9bc2ca3082c1573dd167c66773eb390593844e681f8d0b455d3d"} err="failed to get container status \"0c0f15d331af9bc2ca3082c1573dd167c66773eb390593844e681f8d0b455d3d\": rpc error: code = NotFound desc = could not find container \"0c0f15d331af9bc2ca3082c1573dd167c66773eb390593844e681f8d0b455d3d\": container with ID starting with 0c0f15d331af9bc2ca3082c1573dd167c66773eb390593844e681f8d0b455d3d not found: ID does not exist" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.598851 4556 scope.go:117] "RemoveContainer" containerID="323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.599051 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e"} err="failed to get container status \"323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\": rpc error: code = NotFound desc = could not find container \"323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\": container with ID starting with 323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e not found: ID does not exist" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.599067 4556 scope.go:117] "RemoveContainer" containerID="ad37b48f74a075b5562885db93abefd92df9f91f81a3b2ff685ee97a13b5e403" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.599247 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad37b48f74a075b5562885db93abefd92df9f91f81a3b2ff685ee97a13b5e403"} err="failed to get container status \"ad37b48f74a075b5562885db93abefd92df9f91f81a3b2ff685ee97a13b5e403\": rpc error: code = NotFound desc = could not find container \"ad37b48f74a075b5562885db93abefd92df9f91f81a3b2ff685ee97a13b5e403\": container with ID starting with ad37b48f74a075b5562885db93abefd92df9f91f81a3b2ff685ee97a13b5e403 not found: ID does not exist" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.599261 4556 scope.go:117] "RemoveContainer" containerID="c2f16fe4bf2d659398ec141ba12d2dda32727810b03b96bfaf491602656027cb" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.599426 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2f16fe4bf2d659398ec141ba12d2dda32727810b03b96bfaf491602656027cb"} err="failed to get container status \"c2f16fe4bf2d659398ec141ba12d2dda32727810b03b96bfaf491602656027cb\": rpc error: code = NotFound desc = could not find container \"c2f16fe4bf2d659398ec141ba12d2dda32727810b03b96bfaf491602656027cb\": container with ID starting with c2f16fe4bf2d659398ec141ba12d2dda32727810b03b96bfaf491602656027cb not found: ID does not exist" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.599440 4556 scope.go:117] "RemoveContainer" containerID="aec6030df0a59ad8902d5c694d5cd30d249ca1e895155c57a458e359fa531b7c" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.599603 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aec6030df0a59ad8902d5c694d5cd30d249ca1e895155c57a458e359fa531b7c"} err="failed to get container status \"aec6030df0a59ad8902d5c694d5cd30d249ca1e895155c57a458e359fa531b7c\": rpc error: code = NotFound desc = could not find container \"aec6030df0a59ad8902d5c694d5cd30d249ca1e895155c57a458e359fa531b7c\": container with ID starting with aec6030df0a59ad8902d5c694d5cd30d249ca1e895155c57a458e359fa531b7c not found: ID does not exist" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.599618 4556 scope.go:117] "RemoveContainer" containerID="4cd6e758ee396c9000b1bdb6ded2176ec803fa9da6201cdd9b8146940a2c00cb" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.599783 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4cd6e758ee396c9000b1bdb6ded2176ec803fa9da6201cdd9b8146940a2c00cb"} err="failed to get container status \"4cd6e758ee396c9000b1bdb6ded2176ec803fa9da6201cdd9b8146940a2c00cb\": rpc error: code = NotFound desc = could not find container \"4cd6e758ee396c9000b1bdb6ded2176ec803fa9da6201cdd9b8146940a2c00cb\": container with ID starting with 4cd6e758ee396c9000b1bdb6ded2176ec803fa9da6201cdd9b8146940a2c00cb not found: ID does not exist" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.599797 4556 scope.go:117] "RemoveContainer" containerID="9b0412f3330999c3348ac726ac5a0c2d541571f4b13786ca24cb3109a44aa653" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.599958 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b0412f3330999c3348ac726ac5a0c2d541571f4b13786ca24cb3109a44aa653"} err="failed to get container status \"9b0412f3330999c3348ac726ac5a0c2d541571f4b13786ca24cb3109a44aa653\": rpc error: code = NotFound desc = could not find container \"9b0412f3330999c3348ac726ac5a0c2d541571f4b13786ca24cb3109a44aa653\": container with ID starting with 9b0412f3330999c3348ac726ac5a0c2d541571f4b13786ca24cb3109a44aa653 not found: ID does not exist" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.599973 4556 scope.go:117] "RemoveContainer" containerID="01c526c0613f19e1a1a584b7129bd56001faf08d0b97fd77feeb69f764ed01f8" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.600175 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01c526c0613f19e1a1a584b7129bd56001faf08d0b97fd77feeb69f764ed01f8"} err="failed to get container status \"01c526c0613f19e1a1a584b7129bd56001faf08d0b97fd77feeb69f764ed01f8\": rpc error: code = NotFound desc = could not find container \"01c526c0613f19e1a1a584b7129bd56001faf08d0b97fd77feeb69f764ed01f8\": container with ID starting with 01c526c0613f19e1a1a584b7129bd56001faf08d0b97fd77feeb69f764ed01f8 not found: ID does not exist" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.600192 4556 scope.go:117] "RemoveContainer" containerID="9170bdcbaf1065995d03e6448e40bdced0741c716c9e81f37a72b015132f1161" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.600345 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9170bdcbaf1065995d03e6448e40bdced0741c716c9e81f37a72b015132f1161"} err="failed to get container status \"9170bdcbaf1065995d03e6448e40bdced0741c716c9e81f37a72b015132f1161\": rpc error: code = NotFound desc = could not find container \"9170bdcbaf1065995d03e6448e40bdced0741c716c9e81f37a72b015132f1161\": container with ID starting with 9170bdcbaf1065995d03e6448e40bdced0741c716c9e81f37a72b015132f1161 not found: ID does not exist" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.600361 4556 scope.go:117] "RemoveContainer" containerID="318cb69e50df4c14df9bf5ef57d5c381421a9bdb72135c9b9620f55c0547b410" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.600522 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"318cb69e50df4c14df9bf5ef57d5c381421a9bdb72135c9b9620f55c0547b410"} err="failed to get container status \"318cb69e50df4c14df9bf5ef57d5c381421a9bdb72135c9b9620f55c0547b410\": rpc error: code = NotFound desc = could not find container \"318cb69e50df4c14df9bf5ef57d5c381421a9bdb72135c9b9620f55c0547b410\": container with ID starting with 318cb69e50df4c14df9bf5ef57d5c381421a9bdb72135c9b9620f55c0547b410 not found: ID does not exist" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.600536 4556 scope.go:117] "RemoveContainer" containerID="0c0f15d331af9bc2ca3082c1573dd167c66773eb390593844e681f8d0b455d3d" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.600696 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c0f15d331af9bc2ca3082c1573dd167c66773eb390593844e681f8d0b455d3d"} err="failed to get container status \"0c0f15d331af9bc2ca3082c1573dd167c66773eb390593844e681f8d0b455d3d\": rpc error: code = NotFound desc = could not find container \"0c0f15d331af9bc2ca3082c1573dd167c66773eb390593844e681f8d0b455d3d\": container with ID starting with 0c0f15d331af9bc2ca3082c1573dd167c66773eb390593844e681f8d0b455d3d not found: ID does not exist" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.600711 4556 scope.go:117] "RemoveContainer" containerID="323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.600873 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e"} err="failed to get container status \"323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\": rpc error: code = NotFound desc = could not find container \"323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\": container with ID starting with 323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e not found: ID does not exist" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.600887 4556 scope.go:117] "RemoveContainer" containerID="ad37b48f74a075b5562885db93abefd92df9f91f81a3b2ff685ee97a13b5e403" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.601068 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad37b48f74a075b5562885db93abefd92df9f91f81a3b2ff685ee97a13b5e403"} err="failed to get container status \"ad37b48f74a075b5562885db93abefd92df9f91f81a3b2ff685ee97a13b5e403\": rpc error: code = NotFound desc = could not find container \"ad37b48f74a075b5562885db93abefd92df9f91f81a3b2ff685ee97a13b5e403\": container with ID starting with ad37b48f74a075b5562885db93abefd92df9f91f81a3b2ff685ee97a13b5e403 not found: ID does not exist" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.601082 4556 scope.go:117] "RemoveContainer" containerID="c2f16fe4bf2d659398ec141ba12d2dda32727810b03b96bfaf491602656027cb" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.601307 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2f16fe4bf2d659398ec141ba12d2dda32727810b03b96bfaf491602656027cb"} err="failed to get container status \"c2f16fe4bf2d659398ec141ba12d2dda32727810b03b96bfaf491602656027cb\": rpc error: code = NotFound desc = could not find container \"c2f16fe4bf2d659398ec141ba12d2dda32727810b03b96bfaf491602656027cb\": container with ID starting with c2f16fe4bf2d659398ec141ba12d2dda32727810b03b96bfaf491602656027cb not found: ID does not exist" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.601322 4556 scope.go:117] "RemoveContainer" containerID="aec6030df0a59ad8902d5c694d5cd30d249ca1e895155c57a458e359fa531b7c" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.601484 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aec6030df0a59ad8902d5c694d5cd30d249ca1e895155c57a458e359fa531b7c"} err="failed to get container status \"aec6030df0a59ad8902d5c694d5cd30d249ca1e895155c57a458e359fa531b7c\": rpc error: code = NotFound desc = could not find container \"aec6030df0a59ad8902d5c694d5cd30d249ca1e895155c57a458e359fa531b7c\": container with ID starting with aec6030df0a59ad8902d5c694d5cd30d249ca1e895155c57a458e359fa531b7c not found: ID does not exist" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.601499 4556 scope.go:117] "RemoveContainer" containerID="4cd6e758ee396c9000b1bdb6ded2176ec803fa9da6201cdd9b8146940a2c00cb" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.601665 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4cd6e758ee396c9000b1bdb6ded2176ec803fa9da6201cdd9b8146940a2c00cb"} err="failed to get container status \"4cd6e758ee396c9000b1bdb6ded2176ec803fa9da6201cdd9b8146940a2c00cb\": rpc error: code = NotFound desc = could not find container \"4cd6e758ee396c9000b1bdb6ded2176ec803fa9da6201cdd9b8146940a2c00cb\": container with ID starting with 4cd6e758ee396c9000b1bdb6ded2176ec803fa9da6201cdd9b8146940a2c00cb not found: ID does not exist" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.601677 4556 scope.go:117] "RemoveContainer" containerID="9b0412f3330999c3348ac726ac5a0c2d541571f4b13786ca24cb3109a44aa653" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.601844 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b0412f3330999c3348ac726ac5a0c2d541571f4b13786ca24cb3109a44aa653"} err="failed to get container status \"9b0412f3330999c3348ac726ac5a0c2d541571f4b13786ca24cb3109a44aa653\": rpc error: code = NotFound desc = could not find container \"9b0412f3330999c3348ac726ac5a0c2d541571f4b13786ca24cb3109a44aa653\": container with ID starting with 9b0412f3330999c3348ac726ac5a0c2d541571f4b13786ca24cb3109a44aa653 not found: ID does not exist" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.601859 4556 scope.go:117] "RemoveContainer" containerID="01c526c0613f19e1a1a584b7129bd56001faf08d0b97fd77feeb69f764ed01f8" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.602058 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01c526c0613f19e1a1a584b7129bd56001faf08d0b97fd77feeb69f764ed01f8"} err="failed to get container status \"01c526c0613f19e1a1a584b7129bd56001faf08d0b97fd77feeb69f764ed01f8\": rpc error: code = NotFound desc = could not find container \"01c526c0613f19e1a1a584b7129bd56001faf08d0b97fd77feeb69f764ed01f8\": container with ID starting with 01c526c0613f19e1a1a584b7129bd56001faf08d0b97fd77feeb69f764ed01f8 not found: ID does not exist" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.602073 4556 scope.go:117] "RemoveContainer" containerID="9170bdcbaf1065995d03e6448e40bdced0741c716c9e81f37a72b015132f1161" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.602251 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9170bdcbaf1065995d03e6448e40bdced0741c716c9e81f37a72b015132f1161"} err="failed to get container status \"9170bdcbaf1065995d03e6448e40bdced0741c716c9e81f37a72b015132f1161\": rpc error: code = NotFound desc = could not find container \"9170bdcbaf1065995d03e6448e40bdced0741c716c9e81f37a72b015132f1161\": container with ID starting with 9170bdcbaf1065995d03e6448e40bdced0741c716c9e81f37a72b015132f1161 not found: ID does not exist" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.602267 4556 scope.go:117] "RemoveContainer" containerID="318cb69e50df4c14df9bf5ef57d5c381421a9bdb72135c9b9620f55c0547b410" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.602439 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"318cb69e50df4c14df9bf5ef57d5c381421a9bdb72135c9b9620f55c0547b410"} err="failed to get container status \"318cb69e50df4c14df9bf5ef57d5c381421a9bdb72135c9b9620f55c0547b410\": rpc error: code = NotFound desc = could not find container \"318cb69e50df4c14df9bf5ef57d5c381421a9bdb72135c9b9620f55c0547b410\": container with ID starting with 318cb69e50df4c14df9bf5ef57d5c381421a9bdb72135c9b9620f55c0547b410 not found: ID does not exist" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.602477 4556 scope.go:117] "RemoveContainer" containerID="0c0f15d331af9bc2ca3082c1573dd167c66773eb390593844e681f8d0b455d3d" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.602605 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c0f15d331af9bc2ca3082c1573dd167c66773eb390593844e681f8d0b455d3d"} err="failed to get container status \"0c0f15d331af9bc2ca3082c1573dd167c66773eb390593844e681f8d0b455d3d\": rpc error: code = NotFound desc = could not find container \"0c0f15d331af9bc2ca3082c1573dd167c66773eb390593844e681f8d0b455d3d\": container with ID starting with 0c0f15d331af9bc2ca3082c1573dd167c66773eb390593844e681f8d0b455d3d not found: ID does not exist" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.602651 4556 scope.go:117] "RemoveContainer" containerID="323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.602777 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e"} err="failed to get container status \"323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\": rpc error: code = NotFound desc = could not find container \"323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e\": container with ID starting with 323cd8fe2e1fadc677e83e775d1e842747b975dcfc380d253dcbb81da9ff466e not found: ID does not exist" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.602810 4556 scope.go:117] "RemoveContainer" containerID="ad37b48f74a075b5562885db93abefd92df9f91f81a3b2ff685ee97a13b5e403" Feb 18 09:11:59 crc kubenswrapper[4556]: I0218 09:11:59.602947 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad37b48f74a075b5562885db93abefd92df9f91f81a3b2ff685ee97a13b5e403"} err="failed to get container status \"ad37b48f74a075b5562885db93abefd92df9f91f81a3b2ff685ee97a13b5e403\": rpc error: code = NotFound desc = could not find container \"ad37b48f74a075b5562885db93abefd92df9f91f81a3b2ff685ee97a13b5e403\": container with ID starting with ad37b48f74a075b5562885db93abefd92df9f91f81a3b2ff685ee97a13b5e403 not found: ID does not exist" Feb 18 09:12:00 crc kubenswrapper[4556]: I0218 09:12:00.444262 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" event={"ID":"1c39975c-67c8-47f3-bd1e-c2deb655d009","Type":"ContainerStarted","Data":"6cc46a63e49e9bcba9ba8447e324741e28f2e2c2c3ece382b22e5df1e17c3ca2"} Feb 18 09:12:00 crc kubenswrapper[4556]: I0218 09:12:00.444302 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" event={"ID":"1c39975c-67c8-47f3-bd1e-c2deb655d009","Type":"ContainerStarted","Data":"a2ceaf99d0b55348b1d11bc8e10e73da18747ea187ce14e9243093ff03026ed0"} Feb 18 09:12:00 crc kubenswrapper[4556]: I0218 09:12:00.444313 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" event={"ID":"1c39975c-67c8-47f3-bd1e-c2deb655d009","Type":"ContainerStarted","Data":"510e4762449897a29d841291081c96afdbdf6a84a83bc5364f8d8099d698cd0e"} Feb 18 09:12:00 crc kubenswrapper[4556]: I0218 09:12:00.444322 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" event={"ID":"1c39975c-67c8-47f3-bd1e-c2deb655d009","Type":"ContainerStarted","Data":"1f2c3af8fb50356f3427730de93f64ad2733b478307067398895b326cfac095a"} Feb 18 09:12:00 crc kubenswrapper[4556]: I0218 09:12:00.444330 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" event={"ID":"1c39975c-67c8-47f3-bd1e-c2deb655d009","Type":"ContainerStarted","Data":"9243774b5be1ff76c7bc96c29057d11d02f0f42f59a5e210039a11a1ce92ca63"} Feb 18 09:12:00 crc kubenswrapper[4556]: I0218 09:12:00.444339 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" event={"ID":"1c39975c-67c8-47f3-bd1e-c2deb655d009","Type":"ContainerStarted","Data":"298f22e76b309ce39e0372be37419ae5925628e186a4c44b27eb63e2069856ad"} Feb 18 09:12:01 crc kubenswrapper[4556]: I0218 09:12:01.289283 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09971522-31bc-4e10-8831-4fb62675e220" path="/var/lib/kubelet/pods/09971522-31bc-4e10-8831-4fb62675e220/volumes" Feb 18 09:12:02 crc kubenswrapper[4556]: I0218 09:12:02.454900 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" event={"ID":"1c39975c-67c8-47f3-bd1e-c2deb655d009","Type":"ContainerStarted","Data":"c428352ef2d2300ec6788c31d3a6e7071775952666fb193cf92c76493309579a"} Feb 18 09:12:04 crc kubenswrapper[4556]: I0218 09:12:04.472775 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" event={"ID":"1c39975c-67c8-47f3-bd1e-c2deb655d009","Type":"ContainerStarted","Data":"bf3395c3b916dc676bb03c1df4f7163c9a2255fa15fd0fbe4150441421db25bb"} Feb 18 09:12:04 crc kubenswrapper[4556]: I0218 09:12:04.473328 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:12:04 crc kubenswrapper[4556]: I0218 09:12:04.473340 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:12:04 crc kubenswrapper[4556]: I0218 09:12:04.473349 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:12:04 crc kubenswrapper[4556]: I0218 09:12:04.494336 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" podStartSLOduration=6.494321589 podStartE2EDuration="6.494321589s" podCreationTimestamp="2026-02-18 09:11:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:12:04.491365407 +0000 UTC m=+481.508326387" watchObservedRunningTime="2026-02-18 09:12:04.494321589 +0000 UTC m=+481.511282569" Feb 18 09:12:04 crc kubenswrapper[4556]: I0218 09:12:04.495369 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:12:04 crc kubenswrapper[4556]: I0218 09:12:04.497131 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:12:12 crc kubenswrapper[4556]: I0218 09:12:12.282149 4556 scope.go:117] "RemoveContainer" containerID="23f488cd416c6d49acf193e27beba9e9f1dcfb36430f9ae540742c2e2bd0f144" Feb 18 09:12:12 crc kubenswrapper[4556]: I0218 09:12:12.503687 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-qfksk_e226b189-cfa9-47fc-b94d-19b5cbe0859f/kube-multus/1.log" Feb 18 09:12:12 crc kubenswrapper[4556]: I0218 09:12:12.503846 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-qfksk" event={"ID":"e226b189-cfa9-47fc-b94d-19b5cbe0859f","Type":"ContainerStarted","Data":"1330e16d396e5fecece850d115de3b3fca0bc812b407eff54e9398a722bf0150"} Feb 18 09:12:21 crc kubenswrapper[4556]: I0218 09:12:21.200877 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecal8bm8"] Feb 18 09:12:21 crc kubenswrapper[4556]: I0218 09:12:21.202360 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecal8bm8" Feb 18 09:12:21 crc kubenswrapper[4556]: I0218 09:12:21.203923 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Feb 18 09:12:21 crc kubenswrapper[4556]: I0218 09:12:21.208138 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecal8bm8"] Feb 18 09:12:21 crc kubenswrapper[4556]: I0218 09:12:21.287997 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6b5ce21b-7d77-4d8c-b2d6-272ab408d929-bundle\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecal8bm8\" (UID: \"6b5ce21b-7d77-4d8c-b2d6-272ab408d929\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecal8bm8" Feb 18 09:12:21 crc kubenswrapper[4556]: I0218 09:12:21.288045 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2vxd\" (UniqueName: \"kubernetes.io/projected/6b5ce21b-7d77-4d8c-b2d6-272ab408d929-kube-api-access-x2vxd\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecal8bm8\" (UID: \"6b5ce21b-7d77-4d8c-b2d6-272ab408d929\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecal8bm8" Feb 18 09:12:21 crc kubenswrapper[4556]: I0218 09:12:21.288109 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6b5ce21b-7d77-4d8c-b2d6-272ab408d929-util\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecal8bm8\" (UID: \"6b5ce21b-7d77-4d8c-b2d6-272ab408d929\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecal8bm8" Feb 18 09:12:21 crc kubenswrapper[4556]: I0218 09:12:21.388904 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6b5ce21b-7d77-4d8c-b2d6-272ab408d929-util\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecal8bm8\" (UID: \"6b5ce21b-7d77-4d8c-b2d6-272ab408d929\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecal8bm8" Feb 18 09:12:21 crc kubenswrapper[4556]: I0218 09:12:21.388999 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6b5ce21b-7d77-4d8c-b2d6-272ab408d929-bundle\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecal8bm8\" (UID: \"6b5ce21b-7d77-4d8c-b2d6-272ab408d929\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecal8bm8" Feb 18 09:12:21 crc kubenswrapper[4556]: I0218 09:12:21.389041 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2vxd\" (UniqueName: \"kubernetes.io/projected/6b5ce21b-7d77-4d8c-b2d6-272ab408d929-kube-api-access-x2vxd\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecal8bm8\" (UID: \"6b5ce21b-7d77-4d8c-b2d6-272ab408d929\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecal8bm8" Feb 18 09:12:21 crc kubenswrapper[4556]: I0218 09:12:21.389552 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6b5ce21b-7d77-4d8c-b2d6-272ab408d929-util\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecal8bm8\" (UID: \"6b5ce21b-7d77-4d8c-b2d6-272ab408d929\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecal8bm8" Feb 18 09:12:21 crc kubenswrapper[4556]: I0218 09:12:21.389592 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6b5ce21b-7d77-4d8c-b2d6-272ab408d929-bundle\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecal8bm8\" (UID: \"6b5ce21b-7d77-4d8c-b2d6-272ab408d929\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecal8bm8" Feb 18 09:12:21 crc kubenswrapper[4556]: I0218 09:12:21.407017 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2vxd\" (UniqueName: \"kubernetes.io/projected/6b5ce21b-7d77-4d8c-b2d6-272ab408d929-kube-api-access-x2vxd\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecal8bm8\" (UID: \"6b5ce21b-7d77-4d8c-b2d6-272ab408d929\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecal8bm8" Feb 18 09:12:21 crc kubenswrapper[4556]: I0218 09:12:21.518117 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecal8bm8" Feb 18 09:12:21 crc kubenswrapper[4556]: I0218 09:12:21.880431 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecal8bm8"] Feb 18 09:12:21 crc kubenswrapper[4556]: W0218 09:12:21.884461 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6b5ce21b_7d77_4d8c_b2d6_272ab408d929.slice/crio-a3883ad67a7c3f625874cf4ca2294007bbd17a867668fa44ab8a064653affc62 WatchSource:0}: Error finding container a3883ad67a7c3f625874cf4ca2294007bbd17a867668fa44ab8a064653affc62: Status 404 returned error can't find the container with id a3883ad67a7c3f625874cf4ca2294007bbd17a867668fa44ab8a064653affc62 Feb 18 09:12:22 crc kubenswrapper[4556]: I0218 09:12:22.546237 4556 generic.go:334] "Generic (PLEG): container finished" podID="6b5ce21b-7d77-4d8c-b2d6-272ab408d929" containerID="dfbecdfe50927cb2428217e3fb1e6e2c2702fc4d89ac481a58f0c960efcbeea9" exitCode=0 Feb 18 09:12:22 crc kubenswrapper[4556]: I0218 09:12:22.546356 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecal8bm8" event={"ID":"6b5ce21b-7d77-4d8c-b2d6-272ab408d929","Type":"ContainerDied","Data":"dfbecdfe50927cb2428217e3fb1e6e2c2702fc4d89ac481a58f0c960efcbeea9"} Feb 18 09:12:22 crc kubenswrapper[4556]: I0218 09:12:22.546654 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecal8bm8" event={"ID":"6b5ce21b-7d77-4d8c-b2d6-272ab408d929","Type":"ContainerStarted","Data":"a3883ad67a7c3f625874cf4ca2294007bbd17a867668fa44ab8a064653affc62"} Feb 18 09:12:24 crc kubenswrapper[4556]: I0218 09:12:24.556755 4556 generic.go:334] "Generic (PLEG): container finished" podID="6b5ce21b-7d77-4d8c-b2d6-272ab408d929" containerID="5ef0ee21806ae116841ffbe12645e842e1b863184dc6522228b690f1a0a5da28" exitCode=0 Feb 18 09:12:24 crc kubenswrapper[4556]: I0218 09:12:24.556809 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecal8bm8" event={"ID":"6b5ce21b-7d77-4d8c-b2d6-272ab408d929","Type":"ContainerDied","Data":"5ef0ee21806ae116841ffbe12645e842e1b863184dc6522228b690f1a0a5da28"} Feb 18 09:12:25 crc kubenswrapper[4556]: I0218 09:12:25.562883 4556 generic.go:334] "Generic (PLEG): container finished" podID="6b5ce21b-7d77-4d8c-b2d6-272ab408d929" containerID="7c6edd3d3631d334173a343aab4c6e48f3f83e1a4fa7061fe0b950453bcd92d6" exitCode=0 Feb 18 09:12:25 crc kubenswrapper[4556]: I0218 09:12:25.562932 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecal8bm8" event={"ID":"6b5ce21b-7d77-4d8c-b2d6-272ab408d929","Type":"ContainerDied","Data":"7c6edd3d3631d334173a343aab4c6e48f3f83e1a4fa7061fe0b950453bcd92d6"} Feb 18 09:12:26 crc kubenswrapper[4556]: I0218 09:12:26.752873 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecal8bm8" Feb 18 09:12:26 crc kubenswrapper[4556]: I0218 09:12:26.955389 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2vxd\" (UniqueName: \"kubernetes.io/projected/6b5ce21b-7d77-4d8c-b2d6-272ab408d929-kube-api-access-x2vxd\") pod \"6b5ce21b-7d77-4d8c-b2d6-272ab408d929\" (UID: \"6b5ce21b-7d77-4d8c-b2d6-272ab408d929\") " Feb 18 09:12:26 crc kubenswrapper[4556]: I0218 09:12:26.955520 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6b5ce21b-7d77-4d8c-b2d6-272ab408d929-util\") pod \"6b5ce21b-7d77-4d8c-b2d6-272ab408d929\" (UID: \"6b5ce21b-7d77-4d8c-b2d6-272ab408d929\") " Feb 18 09:12:26 crc kubenswrapper[4556]: I0218 09:12:26.955569 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6b5ce21b-7d77-4d8c-b2d6-272ab408d929-bundle\") pod \"6b5ce21b-7d77-4d8c-b2d6-272ab408d929\" (UID: \"6b5ce21b-7d77-4d8c-b2d6-272ab408d929\") " Feb 18 09:12:26 crc kubenswrapper[4556]: I0218 09:12:26.956171 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b5ce21b-7d77-4d8c-b2d6-272ab408d929-bundle" (OuterVolumeSpecName: "bundle") pod "6b5ce21b-7d77-4d8c-b2d6-272ab408d929" (UID: "6b5ce21b-7d77-4d8c-b2d6-272ab408d929"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:12:26 crc kubenswrapper[4556]: I0218 09:12:26.956694 4556 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6b5ce21b-7d77-4d8c-b2d6-272ab408d929-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:12:26 crc kubenswrapper[4556]: I0218 09:12:26.961347 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b5ce21b-7d77-4d8c-b2d6-272ab408d929-kube-api-access-x2vxd" (OuterVolumeSpecName: "kube-api-access-x2vxd") pod "6b5ce21b-7d77-4d8c-b2d6-272ab408d929" (UID: "6b5ce21b-7d77-4d8c-b2d6-272ab408d929"). InnerVolumeSpecName "kube-api-access-x2vxd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:12:26 crc kubenswrapper[4556]: I0218 09:12:26.965797 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b5ce21b-7d77-4d8c-b2d6-272ab408d929-util" (OuterVolumeSpecName: "util") pod "6b5ce21b-7d77-4d8c-b2d6-272ab408d929" (UID: "6b5ce21b-7d77-4d8c-b2d6-272ab408d929"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:12:27 crc kubenswrapper[4556]: I0218 09:12:27.057974 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2vxd\" (UniqueName: \"kubernetes.io/projected/6b5ce21b-7d77-4d8c-b2d6-272ab408d929-kube-api-access-x2vxd\") on node \"crc\" DevicePath \"\"" Feb 18 09:12:27 crc kubenswrapper[4556]: I0218 09:12:27.058021 4556 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6b5ce21b-7d77-4d8c-b2d6-272ab408d929-util\") on node \"crc\" DevicePath \"\"" Feb 18 09:12:27 crc kubenswrapper[4556]: I0218 09:12:27.574004 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecal8bm8" event={"ID":"6b5ce21b-7d77-4d8c-b2d6-272ab408d929","Type":"ContainerDied","Data":"a3883ad67a7c3f625874cf4ca2294007bbd17a867668fa44ab8a064653affc62"} Feb 18 09:12:27 crc kubenswrapper[4556]: I0218 09:12:27.574050 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a3883ad67a7c3f625874cf4ca2294007bbd17a867668fa44ab8a064653affc62" Feb 18 09:12:27 crc kubenswrapper[4556]: I0218 09:12:27.574058 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecal8bm8" Feb 18 09:12:28 crc kubenswrapper[4556]: I0218 09:12:28.943360 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-wdg5l" Feb 18 09:12:29 crc kubenswrapper[4556]: I0218 09:12:29.123914 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-694c9596b7-mrrmv"] Feb 18 09:12:29 crc kubenswrapper[4556]: E0218 09:12:29.124135 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b5ce21b-7d77-4d8c-b2d6-272ab408d929" containerName="extract" Feb 18 09:12:29 crc kubenswrapper[4556]: I0218 09:12:29.124164 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b5ce21b-7d77-4d8c-b2d6-272ab408d929" containerName="extract" Feb 18 09:12:29 crc kubenswrapper[4556]: E0218 09:12:29.124178 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b5ce21b-7d77-4d8c-b2d6-272ab408d929" containerName="util" Feb 18 09:12:29 crc kubenswrapper[4556]: I0218 09:12:29.124184 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b5ce21b-7d77-4d8c-b2d6-272ab408d929" containerName="util" Feb 18 09:12:29 crc kubenswrapper[4556]: E0218 09:12:29.124198 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b5ce21b-7d77-4d8c-b2d6-272ab408d929" containerName="pull" Feb 18 09:12:29 crc kubenswrapper[4556]: I0218 09:12:29.124203 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b5ce21b-7d77-4d8c-b2d6-272ab408d929" containerName="pull" Feb 18 09:12:29 crc kubenswrapper[4556]: I0218 09:12:29.124295 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b5ce21b-7d77-4d8c-b2d6-272ab408d929" containerName="extract" Feb 18 09:12:29 crc kubenswrapper[4556]: I0218 09:12:29.124677 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-694c9596b7-mrrmv" Feb 18 09:12:29 crc kubenswrapper[4556]: I0218 09:12:29.126339 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Feb 18 09:12:29 crc kubenswrapper[4556]: I0218 09:12:29.126480 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Feb 18 09:12:29 crc kubenswrapper[4556]: I0218 09:12:29.126568 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-g67hd" Feb 18 09:12:29 crc kubenswrapper[4556]: I0218 09:12:29.139282 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-694c9596b7-mrrmv"] Feb 18 09:12:29 crc kubenswrapper[4556]: I0218 09:12:29.189836 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whswl\" (UniqueName: \"kubernetes.io/projected/b2ae390e-7b87-4f49-b95b-c39e89bbf523-kube-api-access-whswl\") pod \"nmstate-operator-694c9596b7-mrrmv\" (UID: \"b2ae390e-7b87-4f49-b95b-c39e89bbf523\") " pod="openshift-nmstate/nmstate-operator-694c9596b7-mrrmv" Feb 18 09:12:29 crc kubenswrapper[4556]: I0218 09:12:29.290680 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whswl\" (UniqueName: \"kubernetes.io/projected/b2ae390e-7b87-4f49-b95b-c39e89bbf523-kube-api-access-whswl\") pod \"nmstate-operator-694c9596b7-mrrmv\" (UID: \"b2ae390e-7b87-4f49-b95b-c39e89bbf523\") " pod="openshift-nmstate/nmstate-operator-694c9596b7-mrrmv" Feb 18 09:12:29 crc kubenswrapper[4556]: I0218 09:12:29.308588 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whswl\" (UniqueName: \"kubernetes.io/projected/b2ae390e-7b87-4f49-b95b-c39e89bbf523-kube-api-access-whswl\") pod \"nmstate-operator-694c9596b7-mrrmv\" (UID: \"b2ae390e-7b87-4f49-b95b-c39e89bbf523\") " pod="openshift-nmstate/nmstate-operator-694c9596b7-mrrmv" Feb 18 09:12:29 crc kubenswrapper[4556]: I0218 09:12:29.436935 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-694c9596b7-mrrmv" Feb 18 09:12:29 crc kubenswrapper[4556]: I0218 09:12:29.598974 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-694c9596b7-mrrmv"] Feb 18 09:12:29 crc kubenswrapper[4556]: W0218 09:12:29.602687 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb2ae390e_7b87_4f49_b95b_c39e89bbf523.slice/crio-3b5de09c4a93993576f69b754f4af13809b2fc36b07b2ecf3a47ccd64c9490bd WatchSource:0}: Error finding container 3b5de09c4a93993576f69b754f4af13809b2fc36b07b2ecf3a47ccd64c9490bd: Status 404 returned error can't find the container with id 3b5de09c4a93993576f69b754f4af13809b2fc36b07b2ecf3a47ccd64c9490bd Feb 18 09:12:30 crc kubenswrapper[4556]: I0218 09:12:30.600697 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-694c9596b7-mrrmv" event={"ID":"b2ae390e-7b87-4f49-b95b-c39e89bbf523","Type":"ContainerStarted","Data":"3b5de09c4a93993576f69b754f4af13809b2fc36b07b2ecf3a47ccd64c9490bd"} Feb 18 09:12:31 crc kubenswrapper[4556]: I0218 09:12:31.609977 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-694c9596b7-mrrmv" event={"ID":"b2ae390e-7b87-4f49-b95b-c39e89bbf523","Type":"ContainerStarted","Data":"5769b64f4e1a6b01242515e305546c3415917cdc60a133f8f9aef63bc3f1ce6d"} Feb 18 09:12:31 crc kubenswrapper[4556]: I0218 09:12:31.627404 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-694c9596b7-mrrmv" podStartSLOduration=0.847392917 podStartE2EDuration="2.627388661s" podCreationTimestamp="2026-02-18 09:12:29 +0000 UTC" firstStartedPulling="2026-02-18 09:12:29.605216617 +0000 UTC m=+506.622177596" lastFinishedPulling="2026-02-18 09:12:31.385212361 +0000 UTC m=+508.402173340" observedRunningTime="2026-02-18 09:12:31.625311157 +0000 UTC m=+508.642272137" watchObservedRunningTime="2026-02-18 09:12:31.627388661 +0000 UTC m=+508.644349641" Feb 18 09:12:31 crc kubenswrapper[4556]: I0218 09:12:31.727285 4556 patch_prober.go:28] interesting pod/machine-config-daemon-f76hs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 09:12:31 crc kubenswrapper[4556]: I0218 09:12:31.727337 4556 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.433448 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-58c85c668d-nhd7k"] Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.434479 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-58c85c668d-nhd7k" Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.436939 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-7czds" Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.437970 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b64z7\" (UniqueName: \"kubernetes.io/projected/0933417f-4853-4022-b87e-7f3584341e8d-kube-api-access-b64z7\") pod \"nmstate-metrics-58c85c668d-nhd7k\" (UID: \"0933417f-4853-4022-b87e-7f3584341e8d\") " pod="openshift-nmstate/nmstate-metrics-58c85c668d-nhd7k" Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.450034 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-58c85c668d-nhd7k"] Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.453695 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-866bcb46dc-cm8rm"] Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.454512 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-cm8rm" Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.455783 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.460830 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-7dmwd"] Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.461723 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-7dmwd" Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.484556 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-866bcb46dc-cm8rm"] Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.539696 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b64z7\" (UniqueName: \"kubernetes.io/projected/0933417f-4853-4022-b87e-7f3584341e8d-kube-api-access-b64z7\") pod \"nmstate-metrics-58c85c668d-nhd7k\" (UID: \"0933417f-4853-4022-b87e-7f3584341e8d\") " pod="openshift-nmstate/nmstate-metrics-58c85c668d-nhd7k" Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.539775 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/df346433-edc0-4b09-b028-ef60228567d1-tls-key-pair\") pod \"nmstate-webhook-866bcb46dc-cm8rm\" (UID: \"df346433-edc0-4b09-b028-ef60228567d1\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-cm8rm" Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.539806 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wzdj4\" (UniqueName: \"kubernetes.io/projected/df346433-edc0-4b09-b028-ef60228567d1-kube-api-access-wzdj4\") pod \"nmstate-webhook-866bcb46dc-cm8rm\" (UID: \"df346433-edc0-4b09-b028-ef60228567d1\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-cm8rm" Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.539831 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/9a59cd87-e29a-4cf4-a407-3de0680bc1dc-nmstate-lock\") pod \"nmstate-handler-7dmwd\" (UID: \"9a59cd87-e29a-4cf4-a407-3de0680bc1dc\") " pod="openshift-nmstate/nmstate-handler-7dmwd" Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.539868 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/9a59cd87-e29a-4cf4-a407-3de0680bc1dc-ovs-socket\") pod \"nmstate-handler-7dmwd\" (UID: \"9a59cd87-e29a-4cf4-a407-3de0680bc1dc\") " pod="openshift-nmstate/nmstate-handler-7dmwd" Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.539884 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24nhn\" (UniqueName: \"kubernetes.io/projected/9a59cd87-e29a-4cf4-a407-3de0680bc1dc-kube-api-access-24nhn\") pod \"nmstate-handler-7dmwd\" (UID: \"9a59cd87-e29a-4cf4-a407-3de0680bc1dc\") " pod="openshift-nmstate/nmstate-handler-7dmwd" Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.539905 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/9a59cd87-e29a-4cf4-a407-3de0680bc1dc-dbus-socket\") pod \"nmstate-handler-7dmwd\" (UID: \"9a59cd87-e29a-4cf4-a407-3de0680bc1dc\") " pod="openshift-nmstate/nmstate-handler-7dmwd" Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.551626 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5c78fc5d65-lzgcp"] Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.552422 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-lzgcp" Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.554119 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.554281 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-pg2p2" Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.561975 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b64z7\" (UniqueName: \"kubernetes.io/projected/0933417f-4853-4022-b87e-7f3584341e8d-kube-api-access-b64z7\") pod \"nmstate-metrics-58c85c668d-nhd7k\" (UID: \"0933417f-4853-4022-b87e-7f3584341e8d\") " pod="openshift-nmstate/nmstate-metrics-58c85c668d-nhd7k" Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.561980 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5c78fc5d65-lzgcp"] Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.571011 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.641336 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rz6mt\" (UniqueName: \"kubernetes.io/projected/bccd4d34-9d89-40ef-9d4e-4c6a58d2571b-kube-api-access-rz6mt\") pod \"nmstate-console-plugin-5c78fc5d65-lzgcp\" (UID: \"bccd4d34-9d89-40ef-9d4e-4c6a58d2571b\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-lzgcp" Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.641941 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/bccd4d34-9d89-40ef-9d4e-4c6a58d2571b-nginx-conf\") pod \"nmstate-console-plugin-5c78fc5d65-lzgcp\" (UID: \"bccd4d34-9d89-40ef-9d4e-4c6a58d2571b\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-lzgcp" Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.642162 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/df346433-edc0-4b09-b028-ef60228567d1-tls-key-pair\") pod \"nmstate-webhook-866bcb46dc-cm8rm\" (UID: \"df346433-edc0-4b09-b028-ef60228567d1\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-cm8rm" Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.642207 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/bccd4d34-9d89-40ef-9d4e-4c6a58d2571b-plugin-serving-cert\") pod \"nmstate-console-plugin-5c78fc5d65-lzgcp\" (UID: \"bccd4d34-9d89-40ef-9d4e-4c6a58d2571b\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-lzgcp" Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.642253 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wzdj4\" (UniqueName: \"kubernetes.io/projected/df346433-edc0-4b09-b028-ef60228567d1-kube-api-access-wzdj4\") pod \"nmstate-webhook-866bcb46dc-cm8rm\" (UID: \"df346433-edc0-4b09-b028-ef60228567d1\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-cm8rm" Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.642291 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/9a59cd87-e29a-4cf4-a407-3de0680bc1dc-nmstate-lock\") pod \"nmstate-handler-7dmwd\" (UID: \"9a59cd87-e29a-4cf4-a407-3de0680bc1dc\") " pod="openshift-nmstate/nmstate-handler-7dmwd" Feb 18 09:12:32 crc kubenswrapper[4556]: E0218 09:12:32.642346 4556 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.642400 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/9a59cd87-e29a-4cf4-a407-3de0680bc1dc-ovs-socket\") pod \"nmstate-handler-7dmwd\" (UID: \"9a59cd87-e29a-4cf4-a407-3de0680bc1dc\") " pod="openshift-nmstate/nmstate-handler-7dmwd" Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.642360 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/9a59cd87-e29a-4cf4-a407-3de0680bc1dc-ovs-socket\") pod \"nmstate-handler-7dmwd\" (UID: \"9a59cd87-e29a-4cf4-a407-3de0680bc1dc\") " pod="openshift-nmstate/nmstate-handler-7dmwd" Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.642353 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/9a59cd87-e29a-4cf4-a407-3de0680bc1dc-nmstate-lock\") pod \"nmstate-handler-7dmwd\" (UID: \"9a59cd87-e29a-4cf4-a407-3de0680bc1dc\") " pod="openshift-nmstate/nmstate-handler-7dmwd" Feb 18 09:12:32 crc kubenswrapper[4556]: E0218 09:12:32.642434 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/df346433-edc0-4b09-b028-ef60228567d1-tls-key-pair podName:df346433-edc0-4b09-b028-ef60228567d1 nodeName:}" failed. No retries permitted until 2026-02-18 09:12:33.142408707 +0000 UTC m=+510.159369688 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/df346433-edc0-4b09-b028-ef60228567d1-tls-key-pair") pod "nmstate-webhook-866bcb46dc-cm8rm" (UID: "df346433-edc0-4b09-b028-ef60228567d1") : secret "openshift-nmstate-webhook" not found Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.642489 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24nhn\" (UniqueName: \"kubernetes.io/projected/9a59cd87-e29a-4cf4-a407-3de0680bc1dc-kube-api-access-24nhn\") pod \"nmstate-handler-7dmwd\" (UID: \"9a59cd87-e29a-4cf4-a407-3de0680bc1dc\") " pod="openshift-nmstate/nmstate-handler-7dmwd" Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.642522 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/9a59cd87-e29a-4cf4-a407-3de0680bc1dc-dbus-socket\") pod \"nmstate-handler-7dmwd\" (UID: \"9a59cd87-e29a-4cf4-a407-3de0680bc1dc\") " pod="openshift-nmstate/nmstate-handler-7dmwd" Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.642767 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/9a59cd87-e29a-4cf4-a407-3de0680bc1dc-dbus-socket\") pod \"nmstate-handler-7dmwd\" (UID: \"9a59cd87-e29a-4cf4-a407-3de0680bc1dc\") " pod="openshift-nmstate/nmstate-handler-7dmwd" Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.660264 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wzdj4\" (UniqueName: \"kubernetes.io/projected/df346433-edc0-4b09-b028-ef60228567d1-kube-api-access-wzdj4\") pod \"nmstate-webhook-866bcb46dc-cm8rm\" (UID: \"df346433-edc0-4b09-b028-ef60228567d1\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-cm8rm" Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.675676 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24nhn\" (UniqueName: \"kubernetes.io/projected/9a59cd87-e29a-4cf4-a407-3de0680bc1dc-kube-api-access-24nhn\") pod \"nmstate-handler-7dmwd\" (UID: \"9a59cd87-e29a-4cf4-a407-3de0680bc1dc\") " pod="openshift-nmstate/nmstate-handler-7dmwd" Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.734453 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-86f477f8bf-p8wlp"] Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.735424 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-86f477f8bf-p8wlp" Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.743873 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e4247d8b-6532-4c57-a260-d4c098e917c7-service-ca\") pod \"console-86f477f8bf-p8wlp\" (UID: \"e4247d8b-6532-4c57-a260-d4c098e917c7\") " pod="openshift-console/console-86f477f8bf-p8wlp" Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.743920 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e4247d8b-6532-4c57-a260-d4c098e917c7-oauth-serving-cert\") pod \"console-86f477f8bf-p8wlp\" (UID: \"e4247d8b-6532-4c57-a260-d4c098e917c7\") " pod="openshift-console/console-86f477f8bf-p8wlp" Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.743998 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/bccd4d34-9d89-40ef-9d4e-4c6a58d2571b-plugin-serving-cert\") pod \"nmstate-console-plugin-5c78fc5d65-lzgcp\" (UID: \"bccd4d34-9d89-40ef-9d4e-4c6a58d2571b\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-lzgcp" Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.744074 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e4247d8b-6532-4c57-a260-d4c098e917c7-trusted-ca-bundle\") pod \"console-86f477f8bf-p8wlp\" (UID: \"e4247d8b-6532-4c57-a260-d4c098e917c7\") " pod="openshift-console/console-86f477f8bf-p8wlp" Feb 18 09:12:32 crc kubenswrapper[4556]: E0218 09:12:32.744112 4556 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.744138 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e4247d8b-6532-4c57-a260-d4c098e917c7-console-config\") pod \"console-86f477f8bf-p8wlp\" (UID: \"e4247d8b-6532-4c57-a260-d4c098e917c7\") " pod="openshift-console/console-86f477f8bf-p8wlp" Feb 18 09:12:32 crc kubenswrapper[4556]: E0218 09:12:32.744176 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bccd4d34-9d89-40ef-9d4e-4c6a58d2571b-plugin-serving-cert podName:bccd4d34-9d89-40ef-9d4e-4c6a58d2571b nodeName:}" failed. No retries permitted until 2026-02-18 09:12:33.244143986 +0000 UTC m=+510.261104967 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/bccd4d34-9d89-40ef-9d4e-4c6a58d2571b-plugin-serving-cert") pod "nmstate-console-plugin-5c78fc5d65-lzgcp" (UID: "bccd4d34-9d89-40ef-9d4e-4c6a58d2571b") : secret "plugin-serving-cert" not found Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.744214 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e4247d8b-6532-4c57-a260-d4c098e917c7-console-oauth-config\") pod \"console-86f477f8bf-p8wlp\" (UID: \"e4247d8b-6532-4c57-a260-d4c098e917c7\") " pod="openshift-console/console-86f477f8bf-p8wlp" Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.744240 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rz6mt\" (UniqueName: \"kubernetes.io/projected/bccd4d34-9d89-40ef-9d4e-4c6a58d2571b-kube-api-access-rz6mt\") pod \"nmstate-console-plugin-5c78fc5d65-lzgcp\" (UID: \"bccd4d34-9d89-40ef-9d4e-4c6a58d2571b\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-lzgcp" Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.744292 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e4247d8b-6532-4c57-a260-d4c098e917c7-console-serving-cert\") pod \"console-86f477f8bf-p8wlp\" (UID: \"e4247d8b-6532-4c57-a260-d4c098e917c7\") " pod="openshift-console/console-86f477f8bf-p8wlp" Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.744312 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/bccd4d34-9d89-40ef-9d4e-4c6a58d2571b-nginx-conf\") pod \"nmstate-console-plugin-5c78fc5d65-lzgcp\" (UID: \"bccd4d34-9d89-40ef-9d4e-4c6a58d2571b\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-lzgcp" Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.744334 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdpkp\" (UniqueName: \"kubernetes.io/projected/e4247d8b-6532-4c57-a260-d4c098e917c7-kube-api-access-sdpkp\") pod \"console-86f477f8bf-p8wlp\" (UID: \"e4247d8b-6532-4c57-a260-d4c098e917c7\") " pod="openshift-console/console-86f477f8bf-p8wlp" Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.745240 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/bccd4d34-9d89-40ef-9d4e-4c6a58d2571b-nginx-conf\") pod \"nmstate-console-plugin-5c78fc5d65-lzgcp\" (UID: \"bccd4d34-9d89-40ef-9d4e-4c6a58d2571b\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-lzgcp" Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.748403 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-58c85c668d-nhd7k" Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.748426 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-86f477f8bf-p8wlp"] Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.776300 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rz6mt\" (UniqueName: \"kubernetes.io/projected/bccd4d34-9d89-40ef-9d4e-4c6a58d2571b-kube-api-access-rz6mt\") pod \"nmstate-console-plugin-5c78fc5d65-lzgcp\" (UID: \"bccd4d34-9d89-40ef-9d4e-4c6a58d2571b\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-lzgcp" Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.782546 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-7dmwd" Feb 18 09:12:32 crc kubenswrapper[4556]: W0218 09:12:32.813764 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9a59cd87_e29a_4cf4_a407_3de0680bc1dc.slice/crio-50a65b783413c19926d1e81a28d981129c04abd9bf7a8be49d5056ac6c5e177a WatchSource:0}: Error finding container 50a65b783413c19926d1e81a28d981129c04abd9bf7a8be49d5056ac6c5e177a: Status 404 returned error can't find the container with id 50a65b783413c19926d1e81a28d981129c04abd9bf7a8be49d5056ac6c5e177a Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.847786 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e4247d8b-6532-4c57-a260-d4c098e917c7-console-oauth-config\") pod \"console-86f477f8bf-p8wlp\" (UID: \"e4247d8b-6532-4c57-a260-d4c098e917c7\") " pod="openshift-console/console-86f477f8bf-p8wlp" Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.847961 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e4247d8b-6532-4c57-a260-d4c098e917c7-console-serving-cert\") pod \"console-86f477f8bf-p8wlp\" (UID: \"e4247d8b-6532-4c57-a260-d4c098e917c7\") " pod="openshift-console/console-86f477f8bf-p8wlp" Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.848041 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdpkp\" (UniqueName: \"kubernetes.io/projected/e4247d8b-6532-4c57-a260-d4c098e917c7-kube-api-access-sdpkp\") pod \"console-86f477f8bf-p8wlp\" (UID: \"e4247d8b-6532-4c57-a260-d4c098e917c7\") " pod="openshift-console/console-86f477f8bf-p8wlp" Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.848120 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e4247d8b-6532-4c57-a260-d4c098e917c7-service-ca\") pod \"console-86f477f8bf-p8wlp\" (UID: \"e4247d8b-6532-4c57-a260-d4c098e917c7\") " pod="openshift-console/console-86f477f8bf-p8wlp" Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.848222 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e4247d8b-6532-4c57-a260-d4c098e917c7-oauth-serving-cert\") pod \"console-86f477f8bf-p8wlp\" (UID: \"e4247d8b-6532-4c57-a260-d4c098e917c7\") " pod="openshift-console/console-86f477f8bf-p8wlp" Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.848319 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e4247d8b-6532-4c57-a260-d4c098e917c7-trusted-ca-bundle\") pod \"console-86f477f8bf-p8wlp\" (UID: \"e4247d8b-6532-4c57-a260-d4c098e917c7\") " pod="openshift-console/console-86f477f8bf-p8wlp" Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.848408 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e4247d8b-6532-4c57-a260-d4c098e917c7-console-config\") pod \"console-86f477f8bf-p8wlp\" (UID: \"e4247d8b-6532-4c57-a260-d4c098e917c7\") " pod="openshift-console/console-86f477f8bf-p8wlp" Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.854623 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e4247d8b-6532-4c57-a260-d4c098e917c7-console-serving-cert\") pod \"console-86f477f8bf-p8wlp\" (UID: \"e4247d8b-6532-4c57-a260-d4c098e917c7\") " pod="openshift-console/console-86f477f8bf-p8wlp" Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.857452 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e4247d8b-6532-4c57-a260-d4c098e917c7-console-config\") pod \"console-86f477f8bf-p8wlp\" (UID: \"e4247d8b-6532-4c57-a260-d4c098e917c7\") " pod="openshift-console/console-86f477f8bf-p8wlp" Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.858270 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e4247d8b-6532-4c57-a260-d4c098e917c7-trusted-ca-bundle\") pod \"console-86f477f8bf-p8wlp\" (UID: \"e4247d8b-6532-4c57-a260-d4c098e917c7\") " pod="openshift-console/console-86f477f8bf-p8wlp" Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.858661 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e4247d8b-6532-4c57-a260-d4c098e917c7-console-oauth-config\") pod \"console-86f477f8bf-p8wlp\" (UID: \"e4247d8b-6532-4c57-a260-d4c098e917c7\") " pod="openshift-console/console-86f477f8bf-p8wlp" Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.860411 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e4247d8b-6532-4c57-a260-d4c098e917c7-service-ca\") pod \"console-86f477f8bf-p8wlp\" (UID: \"e4247d8b-6532-4c57-a260-d4c098e917c7\") " pod="openshift-console/console-86f477f8bf-p8wlp" Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.861712 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e4247d8b-6532-4c57-a260-d4c098e917c7-oauth-serving-cert\") pod \"console-86f477f8bf-p8wlp\" (UID: \"e4247d8b-6532-4c57-a260-d4c098e917c7\") " pod="openshift-console/console-86f477f8bf-p8wlp" Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.872721 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdpkp\" (UniqueName: \"kubernetes.io/projected/e4247d8b-6532-4c57-a260-d4c098e917c7-kube-api-access-sdpkp\") pod \"console-86f477f8bf-p8wlp\" (UID: \"e4247d8b-6532-4c57-a260-d4c098e917c7\") " pod="openshift-console/console-86f477f8bf-p8wlp" Feb 18 09:12:32 crc kubenswrapper[4556]: I0218 09:12:32.950604 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-58c85c668d-nhd7k"] Feb 18 09:12:32 crc kubenswrapper[4556]: W0218 09:12:32.956947 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0933417f_4853_4022_b87e_7f3584341e8d.slice/crio-ef46a3d9236e3fe51d33cdf15654e32f6fc9cbbe258929b4a29599c24c4d9ea5 WatchSource:0}: Error finding container ef46a3d9236e3fe51d33cdf15654e32f6fc9cbbe258929b4a29599c24c4d9ea5: Status 404 returned error can't find the container with id ef46a3d9236e3fe51d33cdf15654e32f6fc9cbbe258929b4a29599c24c4d9ea5 Feb 18 09:12:33 crc kubenswrapper[4556]: I0218 09:12:33.049300 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-86f477f8bf-p8wlp" Feb 18 09:12:33 crc kubenswrapper[4556]: I0218 09:12:33.152634 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/df346433-edc0-4b09-b028-ef60228567d1-tls-key-pair\") pod \"nmstate-webhook-866bcb46dc-cm8rm\" (UID: \"df346433-edc0-4b09-b028-ef60228567d1\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-cm8rm" Feb 18 09:12:33 crc kubenswrapper[4556]: I0218 09:12:33.156578 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/df346433-edc0-4b09-b028-ef60228567d1-tls-key-pair\") pod \"nmstate-webhook-866bcb46dc-cm8rm\" (UID: \"df346433-edc0-4b09-b028-ef60228567d1\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-cm8rm" Feb 18 09:12:33 crc kubenswrapper[4556]: I0218 09:12:33.211461 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-86f477f8bf-p8wlp"] Feb 18 09:12:33 crc kubenswrapper[4556]: W0218 09:12:33.213722 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode4247d8b_6532_4c57_a260_d4c098e917c7.slice/crio-de8f42173727cc50a030911d1ddab6b3053aec1a9d07daed5163b5ca623af5d6 WatchSource:0}: Error finding container de8f42173727cc50a030911d1ddab6b3053aec1a9d07daed5163b5ca623af5d6: Status 404 returned error can't find the container with id de8f42173727cc50a030911d1ddab6b3053aec1a9d07daed5163b5ca623af5d6 Feb 18 09:12:33 crc kubenswrapper[4556]: I0218 09:12:33.253590 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/bccd4d34-9d89-40ef-9d4e-4c6a58d2571b-plugin-serving-cert\") pod \"nmstate-console-plugin-5c78fc5d65-lzgcp\" (UID: \"bccd4d34-9d89-40ef-9d4e-4c6a58d2571b\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-lzgcp" Feb 18 09:12:33 crc kubenswrapper[4556]: I0218 09:12:33.257279 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/bccd4d34-9d89-40ef-9d4e-4c6a58d2571b-plugin-serving-cert\") pod \"nmstate-console-plugin-5c78fc5d65-lzgcp\" (UID: \"bccd4d34-9d89-40ef-9d4e-4c6a58d2571b\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-lzgcp" Feb 18 09:12:33 crc kubenswrapper[4556]: I0218 09:12:33.368143 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-cm8rm" Feb 18 09:12:33 crc kubenswrapper[4556]: I0218 09:12:33.479365 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-lzgcp" Feb 18 09:12:33 crc kubenswrapper[4556]: I0218 09:12:33.621425 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-7dmwd" event={"ID":"9a59cd87-e29a-4cf4-a407-3de0680bc1dc","Type":"ContainerStarted","Data":"50a65b783413c19926d1e81a28d981129c04abd9bf7a8be49d5056ac6c5e177a"} Feb 18 09:12:33 crc kubenswrapper[4556]: I0218 09:12:33.623444 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-86f477f8bf-p8wlp" event={"ID":"e4247d8b-6532-4c57-a260-d4c098e917c7","Type":"ContainerStarted","Data":"c6c894c11284ca93f3c0f6a67163ff891d15e94486d9784f5a95ab477b2c562f"} Feb 18 09:12:33 crc kubenswrapper[4556]: I0218 09:12:33.623494 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-86f477f8bf-p8wlp" event={"ID":"e4247d8b-6532-4c57-a260-d4c098e917c7","Type":"ContainerStarted","Data":"de8f42173727cc50a030911d1ddab6b3053aec1a9d07daed5163b5ca623af5d6"} Feb 18 09:12:33 crc kubenswrapper[4556]: I0218 09:12:33.624683 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58c85c668d-nhd7k" event={"ID":"0933417f-4853-4022-b87e-7f3584341e8d","Type":"ContainerStarted","Data":"ef46a3d9236e3fe51d33cdf15654e32f6fc9cbbe258929b4a29599c24c4d9ea5"} Feb 18 09:12:33 crc kubenswrapper[4556]: I0218 09:12:33.637921 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-86f477f8bf-p8wlp" podStartSLOduration=1.637903772 podStartE2EDuration="1.637903772s" podCreationTimestamp="2026-02-18 09:12:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:12:33.634281392 +0000 UTC m=+510.651242372" watchObservedRunningTime="2026-02-18 09:12:33.637903772 +0000 UTC m=+510.654864753" Feb 18 09:12:33 crc kubenswrapper[4556]: I0218 09:12:33.645293 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5c78fc5d65-lzgcp"] Feb 18 09:12:33 crc kubenswrapper[4556]: W0218 09:12:33.647960 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbccd4d34_9d89_40ef_9d4e_4c6a58d2571b.slice/crio-ed3128583cec0c6f2a25959657a40a1e613be112a4f2e5beb4fedebfd74d6a5f WatchSource:0}: Error finding container ed3128583cec0c6f2a25959657a40a1e613be112a4f2e5beb4fedebfd74d6a5f: Status 404 returned error can't find the container with id ed3128583cec0c6f2a25959657a40a1e613be112a4f2e5beb4fedebfd74d6a5f Feb 18 09:12:33 crc kubenswrapper[4556]: I0218 09:12:33.734210 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-866bcb46dc-cm8rm"] Feb 18 09:12:34 crc kubenswrapper[4556]: I0218 09:12:34.632318 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-cm8rm" event={"ID":"df346433-edc0-4b09-b028-ef60228567d1","Type":"ContainerStarted","Data":"50ab9637d2347745b50ce920d8a5a7bbe36526105e25db90a5e79e0c3bf3cf31"} Feb 18 09:12:34 crc kubenswrapper[4556]: I0218 09:12:34.634422 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-lzgcp" event={"ID":"bccd4d34-9d89-40ef-9d4e-4c6a58d2571b","Type":"ContainerStarted","Data":"ed3128583cec0c6f2a25959657a40a1e613be112a4f2e5beb4fedebfd74d6a5f"} Feb 18 09:12:35 crc kubenswrapper[4556]: I0218 09:12:35.641609 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58c85c668d-nhd7k" event={"ID":"0933417f-4853-4022-b87e-7f3584341e8d","Type":"ContainerStarted","Data":"26a5c2a45ebac7e46895312781e21fdca6057f231a973ea6f0c7c6c47bb1eefc"} Feb 18 09:12:35 crc kubenswrapper[4556]: I0218 09:12:35.644844 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-7dmwd" event={"ID":"9a59cd87-e29a-4cf4-a407-3de0680bc1dc","Type":"ContainerStarted","Data":"474d245c669571cbcb459363cc8e05ce980aa7342438363195b03a3325e18b3d"} Feb 18 09:12:35 crc kubenswrapper[4556]: I0218 09:12:35.645012 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-7dmwd" Feb 18 09:12:35 crc kubenswrapper[4556]: I0218 09:12:35.654014 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-cm8rm" event={"ID":"df346433-edc0-4b09-b028-ef60228567d1","Type":"ContainerStarted","Data":"5151d1dcd061ad3fad20a46b3e76295d3a6999c0df812c68a8dd7ce9b5d39b8d"} Feb 18 09:12:35 crc kubenswrapper[4556]: I0218 09:12:35.654166 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-cm8rm" Feb 18 09:12:35 crc kubenswrapper[4556]: I0218 09:12:35.671220 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-7dmwd" podStartSLOduration=1.313385265 podStartE2EDuration="3.671201988s" podCreationTimestamp="2026-02-18 09:12:32 +0000 UTC" firstStartedPulling="2026-02-18 09:12:32.816294899 +0000 UTC m=+509.833255879" lastFinishedPulling="2026-02-18 09:12:35.174111622 +0000 UTC m=+512.191072602" observedRunningTime="2026-02-18 09:12:35.670454496 +0000 UTC m=+512.687415477" watchObservedRunningTime="2026-02-18 09:12:35.671201988 +0000 UTC m=+512.688162969" Feb 18 09:12:35 crc kubenswrapper[4556]: I0218 09:12:35.686398 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-cm8rm" podStartSLOduration=2.257055492 podStartE2EDuration="3.686385s" podCreationTimestamp="2026-02-18 09:12:32 +0000 UTC" firstStartedPulling="2026-02-18 09:12:33.739016406 +0000 UTC m=+510.755977386" lastFinishedPulling="2026-02-18 09:12:35.168345914 +0000 UTC m=+512.185306894" observedRunningTime="2026-02-18 09:12:35.684517734 +0000 UTC m=+512.701478714" watchObservedRunningTime="2026-02-18 09:12:35.686385 +0000 UTC m=+512.703345980" Feb 18 09:12:36 crc kubenswrapper[4556]: I0218 09:12:36.661855 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-lzgcp" event={"ID":"bccd4d34-9d89-40ef-9d4e-4c6a58d2571b","Type":"ContainerStarted","Data":"53e9b2c4ef261707ee905e848578424dbe9f95975dfacd878bf20b2234dd718a"} Feb 18 09:12:36 crc kubenswrapper[4556]: I0218 09:12:36.684270 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-lzgcp" podStartSLOduration=2.424120542 podStartE2EDuration="4.684210707s" podCreationTimestamp="2026-02-18 09:12:32 +0000 UTC" firstStartedPulling="2026-02-18 09:12:33.650805065 +0000 UTC m=+510.667766046" lastFinishedPulling="2026-02-18 09:12:35.910895231 +0000 UTC m=+512.927856211" observedRunningTime="2026-02-18 09:12:36.678563944 +0000 UTC m=+513.695524924" watchObservedRunningTime="2026-02-18 09:12:36.684210707 +0000 UTC m=+513.701171687" Feb 18 09:12:37 crc kubenswrapper[4556]: I0218 09:12:37.671792 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58c85c668d-nhd7k" event={"ID":"0933417f-4853-4022-b87e-7f3584341e8d","Type":"ContainerStarted","Data":"078f38445de20dca5a92bdee533ae7497a3851f966c3ac039c48e2d1b36ca0ed"} Feb 18 09:12:37 crc kubenswrapper[4556]: I0218 09:12:37.687082 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-58c85c668d-nhd7k" podStartSLOduration=1.078445869 podStartE2EDuration="5.687052726s" podCreationTimestamp="2026-02-18 09:12:32 +0000 UTC" firstStartedPulling="2026-02-18 09:12:32.958922139 +0000 UTC m=+509.975883119" lastFinishedPulling="2026-02-18 09:12:37.567528995 +0000 UTC m=+514.584489976" observedRunningTime="2026-02-18 09:12:37.684937623 +0000 UTC m=+514.701898602" watchObservedRunningTime="2026-02-18 09:12:37.687052726 +0000 UTC m=+514.704013706" Feb 18 09:12:42 crc kubenswrapper[4556]: I0218 09:12:42.800192 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-7dmwd" Feb 18 09:12:43 crc kubenswrapper[4556]: I0218 09:12:43.050012 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-86f477f8bf-p8wlp" Feb 18 09:12:43 crc kubenswrapper[4556]: I0218 09:12:43.050073 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-86f477f8bf-p8wlp" Feb 18 09:12:43 crc kubenswrapper[4556]: I0218 09:12:43.054796 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-86f477f8bf-p8wlp" Feb 18 09:12:43 crc kubenswrapper[4556]: I0218 09:12:43.703627 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-86f477f8bf-p8wlp" Feb 18 09:12:43 crc kubenswrapper[4556]: I0218 09:12:43.738850 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-8gpk5"] Feb 18 09:12:53 crc kubenswrapper[4556]: I0218 09:12:53.373407 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-cm8rm" Feb 18 09:13:01 crc kubenswrapper[4556]: I0218 09:13:01.728781 4556 patch_prober.go:28] interesting pod/machine-config-daemon-f76hs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 09:13:01 crc kubenswrapper[4556]: I0218 09:13:01.730102 4556 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 09:13:05 crc kubenswrapper[4556]: I0218 09:13:05.053780 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gp95f"] Feb 18 09:13:05 crc kubenswrapper[4556]: I0218 09:13:05.055836 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gp95f" Feb 18 09:13:05 crc kubenswrapper[4556]: I0218 09:13:05.058842 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Feb 18 09:13:05 crc kubenswrapper[4556]: I0218 09:13:05.066883 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gp95f"] Feb 18 09:13:05 crc kubenswrapper[4556]: I0218 09:13:05.082950 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3adcbc7e-17be-4546-bd1b-38b6ddee82b6-bundle\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gp95f\" (UID: \"3adcbc7e-17be-4546-bd1b-38b6ddee82b6\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gp95f" Feb 18 09:13:05 crc kubenswrapper[4556]: I0218 09:13:05.083012 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhs7h\" (UniqueName: \"kubernetes.io/projected/3adcbc7e-17be-4546-bd1b-38b6ddee82b6-kube-api-access-xhs7h\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gp95f\" (UID: \"3adcbc7e-17be-4546-bd1b-38b6ddee82b6\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gp95f" Feb 18 09:13:05 crc kubenswrapper[4556]: I0218 09:13:05.083063 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3adcbc7e-17be-4546-bd1b-38b6ddee82b6-util\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gp95f\" (UID: \"3adcbc7e-17be-4546-bd1b-38b6ddee82b6\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gp95f" Feb 18 09:13:05 crc kubenswrapper[4556]: I0218 09:13:05.185209 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3adcbc7e-17be-4546-bd1b-38b6ddee82b6-bundle\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gp95f\" (UID: \"3adcbc7e-17be-4546-bd1b-38b6ddee82b6\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gp95f" Feb 18 09:13:05 crc kubenswrapper[4556]: I0218 09:13:05.185335 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhs7h\" (UniqueName: \"kubernetes.io/projected/3adcbc7e-17be-4546-bd1b-38b6ddee82b6-kube-api-access-xhs7h\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gp95f\" (UID: \"3adcbc7e-17be-4546-bd1b-38b6ddee82b6\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gp95f" Feb 18 09:13:05 crc kubenswrapper[4556]: I0218 09:13:05.185415 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3adcbc7e-17be-4546-bd1b-38b6ddee82b6-util\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gp95f\" (UID: \"3adcbc7e-17be-4546-bd1b-38b6ddee82b6\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gp95f" Feb 18 09:13:05 crc kubenswrapper[4556]: I0218 09:13:05.185670 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3adcbc7e-17be-4546-bd1b-38b6ddee82b6-bundle\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gp95f\" (UID: \"3adcbc7e-17be-4546-bd1b-38b6ddee82b6\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gp95f" Feb 18 09:13:05 crc kubenswrapper[4556]: I0218 09:13:05.186036 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3adcbc7e-17be-4546-bd1b-38b6ddee82b6-util\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gp95f\" (UID: \"3adcbc7e-17be-4546-bd1b-38b6ddee82b6\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gp95f" Feb 18 09:13:05 crc kubenswrapper[4556]: I0218 09:13:05.203508 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhs7h\" (UniqueName: \"kubernetes.io/projected/3adcbc7e-17be-4546-bd1b-38b6ddee82b6-kube-api-access-xhs7h\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gp95f\" (UID: \"3adcbc7e-17be-4546-bd1b-38b6ddee82b6\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gp95f" Feb 18 09:13:05 crc kubenswrapper[4556]: I0218 09:13:05.373912 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gp95f" Feb 18 09:13:05 crc kubenswrapper[4556]: I0218 09:13:05.531862 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gp95f"] Feb 18 09:13:05 crc kubenswrapper[4556]: I0218 09:13:05.831197 4556 generic.go:334] "Generic (PLEG): container finished" podID="3adcbc7e-17be-4546-bd1b-38b6ddee82b6" containerID="bb363bd0b1637dfb9ad683ef5dcd1bddb0d39751867bfe70ea2ca302c08d3d0a" exitCode=0 Feb 18 09:13:05 crc kubenswrapper[4556]: I0218 09:13:05.831289 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gp95f" event={"ID":"3adcbc7e-17be-4546-bd1b-38b6ddee82b6","Type":"ContainerDied","Data":"bb363bd0b1637dfb9ad683ef5dcd1bddb0d39751867bfe70ea2ca302c08d3d0a"} Feb 18 09:13:05 crc kubenswrapper[4556]: I0218 09:13:05.831803 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gp95f" event={"ID":"3adcbc7e-17be-4546-bd1b-38b6ddee82b6","Type":"ContainerStarted","Data":"f39ea0897fa13118f78ae59edc8491ee98d667e5c5ca5a7b22e461c5122719ac"} Feb 18 09:13:07 crc kubenswrapper[4556]: I0218 09:13:07.845594 4556 generic.go:334] "Generic (PLEG): container finished" podID="3adcbc7e-17be-4546-bd1b-38b6ddee82b6" containerID="7c02a75b578833ae2fc6eabbab196072bbe205bd700a69b4e789c8268fbe220d" exitCode=0 Feb 18 09:13:07 crc kubenswrapper[4556]: I0218 09:13:07.845657 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gp95f" event={"ID":"3adcbc7e-17be-4546-bd1b-38b6ddee82b6","Type":"ContainerDied","Data":"7c02a75b578833ae2fc6eabbab196072bbe205bd700a69b4e789c8268fbe220d"} Feb 18 09:13:08 crc kubenswrapper[4556]: I0218 09:13:08.766116 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-8gpk5" podUID="f34c2f1a-5d5b-4719-82e3-d7534e16f7e6" containerName="console" containerID="cri-o://b101476ee7791820a00adaa4bd676ed0412c9e2d157dee254b414bd6c62e163c" gracePeriod=15 Feb 18 09:13:08 crc kubenswrapper[4556]: I0218 09:13:08.851924 4556 generic.go:334] "Generic (PLEG): container finished" podID="3adcbc7e-17be-4546-bd1b-38b6ddee82b6" containerID="2cefcda31f74d968ae9bd2efa3156d8a0a92f8189fe4f40ebf3b3e09711f70d9" exitCode=0 Feb 18 09:13:08 crc kubenswrapper[4556]: I0218 09:13:08.851974 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gp95f" event={"ID":"3adcbc7e-17be-4546-bd1b-38b6ddee82b6","Type":"ContainerDied","Data":"2cefcda31f74d968ae9bd2efa3156d8a0a92f8189fe4f40ebf3b3e09711f70d9"} Feb 18 09:13:09 crc kubenswrapper[4556]: I0218 09:13:09.091413 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-8gpk5_f34c2f1a-5d5b-4719-82e3-d7534e16f7e6/console/0.log" Feb 18 09:13:09 crc kubenswrapper[4556]: I0218 09:13:09.091694 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-8gpk5" Feb 18 09:13:09 crc kubenswrapper[4556]: I0218 09:13:09.232146 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f34c2f1a-5d5b-4719-82e3-d7534e16f7e6-console-config\") pod \"f34c2f1a-5d5b-4719-82e3-d7534e16f7e6\" (UID: \"f34c2f1a-5d5b-4719-82e3-d7534e16f7e6\") " Feb 18 09:13:09 crc kubenswrapper[4556]: I0218 09:13:09.232246 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f34c2f1a-5d5b-4719-82e3-d7534e16f7e6-trusted-ca-bundle\") pod \"f34c2f1a-5d5b-4719-82e3-d7534e16f7e6\" (UID: \"f34c2f1a-5d5b-4719-82e3-d7534e16f7e6\") " Feb 18 09:13:09 crc kubenswrapper[4556]: I0218 09:13:09.232288 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f34c2f1a-5d5b-4719-82e3-d7534e16f7e6-console-serving-cert\") pod \"f34c2f1a-5d5b-4719-82e3-d7534e16f7e6\" (UID: \"f34c2f1a-5d5b-4719-82e3-d7534e16f7e6\") " Feb 18 09:13:09 crc kubenswrapper[4556]: I0218 09:13:09.232313 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f34c2f1a-5d5b-4719-82e3-d7534e16f7e6-service-ca\") pod \"f34c2f1a-5d5b-4719-82e3-d7534e16f7e6\" (UID: \"f34c2f1a-5d5b-4719-82e3-d7534e16f7e6\") " Feb 18 09:13:09 crc kubenswrapper[4556]: I0218 09:13:09.232384 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f34c2f1a-5d5b-4719-82e3-d7534e16f7e6-console-oauth-config\") pod \"f34c2f1a-5d5b-4719-82e3-d7534e16f7e6\" (UID: \"f34c2f1a-5d5b-4719-82e3-d7534e16f7e6\") " Feb 18 09:13:09 crc kubenswrapper[4556]: I0218 09:13:09.232446 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f34c2f1a-5d5b-4719-82e3-d7534e16f7e6-oauth-serving-cert\") pod \"f34c2f1a-5d5b-4719-82e3-d7534e16f7e6\" (UID: \"f34c2f1a-5d5b-4719-82e3-d7534e16f7e6\") " Feb 18 09:13:09 crc kubenswrapper[4556]: I0218 09:13:09.232470 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ftlsd\" (UniqueName: \"kubernetes.io/projected/f34c2f1a-5d5b-4719-82e3-d7534e16f7e6-kube-api-access-ftlsd\") pod \"f34c2f1a-5d5b-4719-82e3-d7534e16f7e6\" (UID: \"f34c2f1a-5d5b-4719-82e3-d7534e16f7e6\") " Feb 18 09:13:09 crc kubenswrapper[4556]: I0218 09:13:09.232994 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f34c2f1a-5d5b-4719-82e3-d7534e16f7e6-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "f34c2f1a-5d5b-4719-82e3-d7534e16f7e6" (UID: "f34c2f1a-5d5b-4719-82e3-d7534e16f7e6"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:13:09 crc kubenswrapper[4556]: I0218 09:13:09.233008 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f34c2f1a-5d5b-4719-82e3-d7534e16f7e6-console-config" (OuterVolumeSpecName: "console-config") pod "f34c2f1a-5d5b-4719-82e3-d7534e16f7e6" (UID: "f34c2f1a-5d5b-4719-82e3-d7534e16f7e6"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:13:09 crc kubenswrapper[4556]: I0218 09:13:09.233091 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f34c2f1a-5d5b-4719-82e3-d7534e16f7e6-service-ca" (OuterVolumeSpecName: "service-ca") pod "f34c2f1a-5d5b-4719-82e3-d7534e16f7e6" (UID: "f34c2f1a-5d5b-4719-82e3-d7534e16f7e6"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:13:09 crc kubenswrapper[4556]: I0218 09:13:09.233639 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f34c2f1a-5d5b-4719-82e3-d7534e16f7e6-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "f34c2f1a-5d5b-4719-82e3-d7534e16f7e6" (UID: "f34c2f1a-5d5b-4719-82e3-d7534e16f7e6"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:13:09 crc kubenswrapper[4556]: I0218 09:13:09.238387 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f34c2f1a-5d5b-4719-82e3-d7534e16f7e6-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "f34c2f1a-5d5b-4719-82e3-d7534e16f7e6" (UID: "f34c2f1a-5d5b-4719-82e3-d7534e16f7e6"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:13:09 crc kubenswrapper[4556]: I0218 09:13:09.238425 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f34c2f1a-5d5b-4719-82e3-d7534e16f7e6-kube-api-access-ftlsd" (OuterVolumeSpecName: "kube-api-access-ftlsd") pod "f34c2f1a-5d5b-4719-82e3-d7534e16f7e6" (UID: "f34c2f1a-5d5b-4719-82e3-d7534e16f7e6"). InnerVolumeSpecName "kube-api-access-ftlsd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:13:09 crc kubenswrapper[4556]: I0218 09:13:09.238609 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f34c2f1a-5d5b-4719-82e3-d7534e16f7e6-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "f34c2f1a-5d5b-4719-82e3-d7534e16f7e6" (UID: "f34c2f1a-5d5b-4719-82e3-d7534e16f7e6"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:13:09 crc kubenswrapper[4556]: I0218 09:13:09.333909 4556 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f34c2f1a-5d5b-4719-82e3-d7534e16f7e6-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:13:09 crc kubenswrapper[4556]: I0218 09:13:09.333941 4556 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f34c2f1a-5d5b-4719-82e3-d7534e16f7e6-console-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 09:13:09 crc kubenswrapper[4556]: I0218 09:13:09.333951 4556 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f34c2f1a-5d5b-4719-82e3-d7534e16f7e6-service-ca\") on node \"crc\" DevicePath \"\"" Feb 18 09:13:09 crc kubenswrapper[4556]: I0218 09:13:09.333961 4556 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f34c2f1a-5d5b-4719-82e3-d7534e16f7e6-console-oauth-config\") on node \"crc\" DevicePath \"\"" Feb 18 09:13:09 crc kubenswrapper[4556]: I0218 09:13:09.333970 4556 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f34c2f1a-5d5b-4719-82e3-d7534e16f7e6-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 09:13:09 crc kubenswrapper[4556]: I0218 09:13:09.333981 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ftlsd\" (UniqueName: \"kubernetes.io/projected/f34c2f1a-5d5b-4719-82e3-d7534e16f7e6-kube-api-access-ftlsd\") on node \"crc\" DevicePath \"\"" Feb 18 09:13:09 crc kubenswrapper[4556]: I0218 09:13:09.333993 4556 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f34c2f1a-5d5b-4719-82e3-d7534e16f7e6-console-config\") on node \"crc\" DevicePath \"\"" Feb 18 09:13:09 crc kubenswrapper[4556]: I0218 09:13:09.857796 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-8gpk5_f34c2f1a-5d5b-4719-82e3-d7534e16f7e6/console/0.log" Feb 18 09:13:09 crc kubenswrapper[4556]: I0218 09:13:09.858405 4556 generic.go:334] "Generic (PLEG): container finished" podID="f34c2f1a-5d5b-4719-82e3-d7534e16f7e6" containerID="b101476ee7791820a00adaa4bd676ed0412c9e2d157dee254b414bd6c62e163c" exitCode=2 Feb 18 09:13:09 crc kubenswrapper[4556]: I0218 09:13:09.858484 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-8gpk5" Feb 18 09:13:09 crc kubenswrapper[4556]: I0218 09:13:09.858450 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-8gpk5" event={"ID":"f34c2f1a-5d5b-4719-82e3-d7534e16f7e6","Type":"ContainerDied","Data":"b101476ee7791820a00adaa4bd676ed0412c9e2d157dee254b414bd6c62e163c"} Feb 18 09:13:09 crc kubenswrapper[4556]: I0218 09:13:09.858580 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-8gpk5" event={"ID":"f34c2f1a-5d5b-4719-82e3-d7534e16f7e6","Type":"ContainerDied","Data":"7601dbdf99c7d9a7ad22314d5e2ee6774b6e2db5f7605c3cd3e6915f61d264dd"} Feb 18 09:13:09 crc kubenswrapper[4556]: I0218 09:13:09.858600 4556 scope.go:117] "RemoveContainer" containerID="b101476ee7791820a00adaa4bd676ed0412c9e2d157dee254b414bd6c62e163c" Feb 18 09:13:09 crc kubenswrapper[4556]: I0218 09:13:09.875759 4556 scope.go:117] "RemoveContainer" containerID="b101476ee7791820a00adaa4bd676ed0412c9e2d157dee254b414bd6c62e163c" Feb 18 09:13:09 crc kubenswrapper[4556]: E0218 09:13:09.876682 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b101476ee7791820a00adaa4bd676ed0412c9e2d157dee254b414bd6c62e163c\": container with ID starting with b101476ee7791820a00adaa4bd676ed0412c9e2d157dee254b414bd6c62e163c not found: ID does not exist" containerID="b101476ee7791820a00adaa4bd676ed0412c9e2d157dee254b414bd6c62e163c" Feb 18 09:13:09 crc kubenswrapper[4556]: I0218 09:13:09.876726 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b101476ee7791820a00adaa4bd676ed0412c9e2d157dee254b414bd6c62e163c"} err="failed to get container status \"b101476ee7791820a00adaa4bd676ed0412c9e2d157dee254b414bd6c62e163c\": rpc error: code = NotFound desc = could not find container \"b101476ee7791820a00adaa4bd676ed0412c9e2d157dee254b414bd6c62e163c\": container with ID starting with b101476ee7791820a00adaa4bd676ed0412c9e2d157dee254b414bd6c62e163c not found: ID does not exist" Feb 18 09:13:09 crc kubenswrapper[4556]: I0218 09:13:09.876742 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-8gpk5"] Feb 18 09:13:09 crc kubenswrapper[4556]: I0218 09:13:09.883039 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-8gpk5"] Feb 18 09:13:10 crc kubenswrapper[4556]: I0218 09:13:10.067804 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gp95f" Feb 18 09:13:10 crc kubenswrapper[4556]: I0218 09:13:10.244749 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xhs7h\" (UniqueName: \"kubernetes.io/projected/3adcbc7e-17be-4546-bd1b-38b6ddee82b6-kube-api-access-xhs7h\") pod \"3adcbc7e-17be-4546-bd1b-38b6ddee82b6\" (UID: \"3adcbc7e-17be-4546-bd1b-38b6ddee82b6\") " Feb 18 09:13:10 crc kubenswrapper[4556]: I0218 09:13:10.244807 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3adcbc7e-17be-4546-bd1b-38b6ddee82b6-bundle\") pod \"3adcbc7e-17be-4546-bd1b-38b6ddee82b6\" (UID: \"3adcbc7e-17be-4546-bd1b-38b6ddee82b6\") " Feb 18 09:13:10 crc kubenswrapper[4556]: I0218 09:13:10.244898 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3adcbc7e-17be-4546-bd1b-38b6ddee82b6-util\") pod \"3adcbc7e-17be-4546-bd1b-38b6ddee82b6\" (UID: \"3adcbc7e-17be-4546-bd1b-38b6ddee82b6\") " Feb 18 09:13:10 crc kubenswrapper[4556]: I0218 09:13:10.245970 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3adcbc7e-17be-4546-bd1b-38b6ddee82b6-bundle" (OuterVolumeSpecName: "bundle") pod "3adcbc7e-17be-4546-bd1b-38b6ddee82b6" (UID: "3adcbc7e-17be-4546-bd1b-38b6ddee82b6"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:13:10 crc kubenswrapper[4556]: I0218 09:13:10.248275 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3adcbc7e-17be-4546-bd1b-38b6ddee82b6-kube-api-access-xhs7h" (OuterVolumeSpecName: "kube-api-access-xhs7h") pod "3adcbc7e-17be-4546-bd1b-38b6ddee82b6" (UID: "3adcbc7e-17be-4546-bd1b-38b6ddee82b6"). InnerVolumeSpecName "kube-api-access-xhs7h". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:13:10 crc kubenswrapper[4556]: I0218 09:13:10.255135 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3adcbc7e-17be-4546-bd1b-38b6ddee82b6-util" (OuterVolumeSpecName: "util") pod "3adcbc7e-17be-4546-bd1b-38b6ddee82b6" (UID: "3adcbc7e-17be-4546-bd1b-38b6ddee82b6"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:13:10 crc kubenswrapper[4556]: I0218 09:13:10.346137 4556 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3adcbc7e-17be-4546-bd1b-38b6ddee82b6-util\") on node \"crc\" DevicePath \"\"" Feb 18 09:13:10 crc kubenswrapper[4556]: I0218 09:13:10.346210 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xhs7h\" (UniqueName: \"kubernetes.io/projected/3adcbc7e-17be-4546-bd1b-38b6ddee82b6-kube-api-access-xhs7h\") on node \"crc\" DevicePath \"\"" Feb 18 09:13:10 crc kubenswrapper[4556]: I0218 09:13:10.346223 4556 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3adcbc7e-17be-4546-bd1b-38b6ddee82b6-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:13:10 crc kubenswrapper[4556]: I0218 09:13:10.867766 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gp95f" event={"ID":"3adcbc7e-17be-4546-bd1b-38b6ddee82b6","Type":"ContainerDied","Data":"f39ea0897fa13118f78ae59edc8491ee98d667e5c5ca5a7b22e461c5122719ac"} Feb 18 09:13:10 crc kubenswrapper[4556]: I0218 09:13:10.867814 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f39ea0897fa13118f78ae59edc8491ee98d667e5c5ca5a7b22e461c5122719ac" Feb 18 09:13:10 crc kubenswrapper[4556]: I0218 09:13:10.867826 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gp95f" Feb 18 09:13:11 crc kubenswrapper[4556]: I0218 09:13:11.291380 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f34c2f1a-5d5b-4719-82e3-d7534e16f7e6" path="/var/lib/kubelet/pods/f34c2f1a-5d5b-4719-82e3-d7534e16f7e6/volumes" Feb 18 09:13:19 crc kubenswrapper[4556]: I0218 09:13:19.127769 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-7cccb48f48-gbb9s"] Feb 18 09:13:19 crc kubenswrapper[4556]: E0218 09:13:19.129650 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f34c2f1a-5d5b-4719-82e3-d7534e16f7e6" containerName="console" Feb 18 09:13:19 crc kubenswrapper[4556]: I0218 09:13:19.129724 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="f34c2f1a-5d5b-4719-82e3-d7534e16f7e6" containerName="console" Feb 18 09:13:19 crc kubenswrapper[4556]: E0218 09:13:19.129775 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3adcbc7e-17be-4546-bd1b-38b6ddee82b6" containerName="pull" Feb 18 09:13:19 crc kubenswrapper[4556]: I0218 09:13:19.129825 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="3adcbc7e-17be-4546-bd1b-38b6ddee82b6" containerName="pull" Feb 18 09:13:19 crc kubenswrapper[4556]: E0218 09:13:19.129875 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3adcbc7e-17be-4546-bd1b-38b6ddee82b6" containerName="util" Feb 18 09:13:19 crc kubenswrapper[4556]: I0218 09:13:19.129916 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="3adcbc7e-17be-4546-bd1b-38b6ddee82b6" containerName="util" Feb 18 09:13:19 crc kubenswrapper[4556]: E0218 09:13:19.129958 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3adcbc7e-17be-4546-bd1b-38b6ddee82b6" containerName="extract" Feb 18 09:13:19 crc kubenswrapper[4556]: I0218 09:13:19.130005 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="3adcbc7e-17be-4546-bd1b-38b6ddee82b6" containerName="extract" Feb 18 09:13:19 crc kubenswrapper[4556]: I0218 09:13:19.130165 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="f34c2f1a-5d5b-4719-82e3-d7534e16f7e6" containerName="console" Feb 18 09:13:19 crc kubenswrapper[4556]: I0218 09:13:19.130252 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="3adcbc7e-17be-4546-bd1b-38b6ddee82b6" containerName="extract" Feb 18 09:13:19 crc kubenswrapper[4556]: I0218 09:13:19.130641 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-7cccb48f48-gbb9s" Feb 18 09:13:19 crc kubenswrapper[4556]: I0218 09:13:19.135532 4556 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Feb 18 09:13:19 crc kubenswrapper[4556]: I0218 09:13:19.135545 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Feb 18 09:13:19 crc kubenswrapper[4556]: I0218 09:13:19.135816 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Feb 18 09:13:19 crc kubenswrapper[4556]: I0218 09:13:19.135913 4556 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-x7hxh" Feb 18 09:13:19 crc kubenswrapper[4556]: I0218 09:13:19.135921 4556 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Feb 18 09:13:19 crc kubenswrapper[4556]: I0218 09:13:19.146141 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-7cccb48f48-gbb9s"] Feb 18 09:13:19 crc kubenswrapper[4556]: I0218 09:13:19.262267 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k54zv\" (UniqueName: \"kubernetes.io/projected/6750b0ea-8f96-404d-bd38-e04e8be95127-kube-api-access-k54zv\") pod \"metallb-operator-controller-manager-7cccb48f48-gbb9s\" (UID: \"6750b0ea-8f96-404d-bd38-e04e8be95127\") " pod="metallb-system/metallb-operator-controller-manager-7cccb48f48-gbb9s" Feb 18 09:13:19 crc kubenswrapper[4556]: I0218 09:13:19.262362 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6750b0ea-8f96-404d-bd38-e04e8be95127-webhook-cert\") pod \"metallb-operator-controller-manager-7cccb48f48-gbb9s\" (UID: \"6750b0ea-8f96-404d-bd38-e04e8be95127\") " pod="metallb-system/metallb-operator-controller-manager-7cccb48f48-gbb9s" Feb 18 09:13:19 crc kubenswrapper[4556]: I0218 09:13:19.262405 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6750b0ea-8f96-404d-bd38-e04e8be95127-apiservice-cert\") pod \"metallb-operator-controller-manager-7cccb48f48-gbb9s\" (UID: \"6750b0ea-8f96-404d-bd38-e04e8be95127\") " pod="metallb-system/metallb-operator-controller-manager-7cccb48f48-gbb9s" Feb 18 09:13:19 crc kubenswrapper[4556]: I0218 09:13:19.364624 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k54zv\" (UniqueName: \"kubernetes.io/projected/6750b0ea-8f96-404d-bd38-e04e8be95127-kube-api-access-k54zv\") pod \"metallb-operator-controller-manager-7cccb48f48-gbb9s\" (UID: \"6750b0ea-8f96-404d-bd38-e04e8be95127\") " pod="metallb-system/metallb-operator-controller-manager-7cccb48f48-gbb9s" Feb 18 09:13:19 crc kubenswrapper[4556]: I0218 09:13:19.364729 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6750b0ea-8f96-404d-bd38-e04e8be95127-webhook-cert\") pod \"metallb-operator-controller-manager-7cccb48f48-gbb9s\" (UID: \"6750b0ea-8f96-404d-bd38-e04e8be95127\") " pod="metallb-system/metallb-operator-controller-manager-7cccb48f48-gbb9s" Feb 18 09:13:19 crc kubenswrapper[4556]: I0218 09:13:19.364770 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6750b0ea-8f96-404d-bd38-e04e8be95127-apiservice-cert\") pod \"metallb-operator-controller-manager-7cccb48f48-gbb9s\" (UID: \"6750b0ea-8f96-404d-bd38-e04e8be95127\") " pod="metallb-system/metallb-operator-controller-manager-7cccb48f48-gbb9s" Feb 18 09:13:19 crc kubenswrapper[4556]: I0218 09:13:19.370851 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6750b0ea-8f96-404d-bd38-e04e8be95127-apiservice-cert\") pod \"metallb-operator-controller-manager-7cccb48f48-gbb9s\" (UID: \"6750b0ea-8f96-404d-bd38-e04e8be95127\") " pod="metallb-system/metallb-operator-controller-manager-7cccb48f48-gbb9s" Feb 18 09:13:19 crc kubenswrapper[4556]: I0218 09:13:19.382755 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6750b0ea-8f96-404d-bd38-e04e8be95127-webhook-cert\") pod \"metallb-operator-controller-manager-7cccb48f48-gbb9s\" (UID: \"6750b0ea-8f96-404d-bd38-e04e8be95127\") " pod="metallb-system/metallb-operator-controller-manager-7cccb48f48-gbb9s" Feb 18 09:13:19 crc kubenswrapper[4556]: I0218 09:13:19.389817 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k54zv\" (UniqueName: \"kubernetes.io/projected/6750b0ea-8f96-404d-bd38-e04e8be95127-kube-api-access-k54zv\") pod \"metallb-operator-controller-manager-7cccb48f48-gbb9s\" (UID: \"6750b0ea-8f96-404d-bd38-e04e8be95127\") " pod="metallb-system/metallb-operator-controller-manager-7cccb48f48-gbb9s" Feb 18 09:13:19 crc kubenswrapper[4556]: I0218 09:13:19.401826 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-749d56bf4-c8bg9"] Feb 18 09:13:19 crc kubenswrapper[4556]: I0218 09:13:19.402444 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-749d56bf4-c8bg9" Feb 18 09:13:19 crc kubenswrapper[4556]: I0218 09:13:19.404176 4556 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Feb 18 09:13:19 crc kubenswrapper[4556]: I0218 09:13:19.404575 4556 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Feb 18 09:13:19 crc kubenswrapper[4556]: I0218 09:13:19.404726 4556 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-tc94h" Feb 18 09:13:19 crc kubenswrapper[4556]: I0218 09:13:19.445887 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-7cccb48f48-gbb9s" Feb 18 09:13:19 crc kubenswrapper[4556]: I0218 09:13:19.474227 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-749d56bf4-c8bg9"] Feb 18 09:13:19 crc kubenswrapper[4556]: I0218 09:13:19.571995 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/414515f8-ef03-41cd-a1c8-7811efcca959-webhook-cert\") pod \"metallb-operator-webhook-server-749d56bf4-c8bg9\" (UID: \"414515f8-ef03-41cd-a1c8-7811efcca959\") " pod="metallb-system/metallb-operator-webhook-server-749d56bf4-c8bg9" Feb 18 09:13:19 crc kubenswrapper[4556]: I0218 09:13:19.572373 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/414515f8-ef03-41cd-a1c8-7811efcca959-apiservice-cert\") pod \"metallb-operator-webhook-server-749d56bf4-c8bg9\" (UID: \"414515f8-ef03-41cd-a1c8-7811efcca959\") " pod="metallb-system/metallb-operator-webhook-server-749d56bf4-c8bg9" Feb 18 09:13:19 crc kubenswrapper[4556]: I0218 09:13:19.572540 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95jlb\" (UniqueName: \"kubernetes.io/projected/414515f8-ef03-41cd-a1c8-7811efcca959-kube-api-access-95jlb\") pod \"metallb-operator-webhook-server-749d56bf4-c8bg9\" (UID: \"414515f8-ef03-41cd-a1c8-7811efcca959\") " pod="metallb-system/metallb-operator-webhook-server-749d56bf4-c8bg9" Feb 18 09:13:19 crc kubenswrapper[4556]: I0218 09:13:19.662834 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-7cccb48f48-gbb9s"] Feb 18 09:13:19 crc kubenswrapper[4556]: W0218 09:13:19.672786 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6750b0ea_8f96_404d_bd38_e04e8be95127.slice/crio-861a7fc31aa040cc83423376fe11d824263788198cab0f1f9f924793b29985a1 WatchSource:0}: Error finding container 861a7fc31aa040cc83423376fe11d824263788198cab0f1f9f924793b29985a1: Status 404 returned error can't find the container with id 861a7fc31aa040cc83423376fe11d824263788198cab0f1f9f924793b29985a1 Feb 18 09:13:19 crc kubenswrapper[4556]: I0218 09:13:19.673381 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95jlb\" (UniqueName: \"kubernetes.io/projected/414515f8-ef03-41cd-a1c8-7811efcca959-kube-api-access-95jlb\") pod \"metallb-operator-webhook-server-749d56bf4-c8bg9\" (UID: \"414515f8-ef03-41cd-a1c8-7811efcca959\") " pod="metallb-system/metallb-operator-webhook-server-749d56bf4-c8bg9" Feb 18 09:13:19 crc kubenswrapper[4556]: I0218 09:13:19.673456 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/414515f8-ef03-41cd-a1c8-7811efcca959-webhook-cert\") pod \"metallb-operator-webhook-server-749d56bf4-c8bg9\" (UID: \"414515f8-ef03-41cd-a1c8-7811efcca959\") " pod="metallb-system/metallb-operator-webhook-server-749d56bf4-c8bg9" Feb 18 09:13:19 crc kubenswrapper[4556]: I0218 09:13:19.673539 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/414515f8-ef03-41cd-a1c8-7811efcca959-apiservice-cert\") pod \"metallb-operator-webhook-server-749d56bf4-c8bg9\" (UID: \"414515f8-ef03-41cd-a1c8-7811efcca959\") " pod="metallb-system/metallb-operator-webhook-server-749d56bf4-c8bg9" Feb 18 09:13:19 crc kubenswrapper[4556]: I0218 09:13:19.677047 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/414515f8-ef03-41cd-a1c8-7811efcca959-apiservice-cert\") pod \"metallb-operator-webhook-server-749d56bf4-c8bg9\" (UID: \"414515f8-ef03-41cd-a1c8-7811efcca959\") " pod="metallb-system/metallb-operator-webhook-server-749d56bf4-c8bg9" Feb 18 09:13:19 crc kubenswrapper[4556]: I0218 09:13:19.677469 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/414515f8-ef03-41cd-a1c8-7811efcca959-webhook-cert\") pod \"metallb-operator-webhook-server-749d56bf4-c8bg9\" (UID: \"414515f8-ef03-41cd-a1c8-7811efcca959\") " pod="metallb-system/metallb-operator-webhook-server-749d56bf4-c8bg9" Feb 18 09:13:19 crc kubenswrapper[4556]: I0218 09:13:19.690436 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95jlb\" (UniqueName: \"kubernetes.io/projected/414515f8-ef03-41cd-a1c8-7811efcca959-kube-api-access-95jlb\") pod \"metallb-operator-webhook-server-749d56bf4-c8bg9\" (UID: \"414515f8-ef03-41cd-a1c8-7811efcca959\") " pod="metallb-system/metallb-operator-webhook-server-749d56bf4-c8bg9" Feb 18 09:13:19 crc kubenswrapper[4556]: I0218 09:13:19.741574 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-749d56bf4-c8bg9" Feb 18 09:13:19 crc kubenswrapper[4556]: I0218 09:13:19.897244 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-749d56bf4-c8bg9"] Feb 18 09:13:19 crc kubenswrapper[4556]: W0218 09:13:19.908118 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod414515f8_ef03_41cd_a1c8_7811efcca959.slice/crio-017881be1ccaf14fba331b26d76818037b33d740fb7b0d11fd70ca6e33bfbe42 WatchSource:0}: Error finding container 017881be1ccaf14fba331b26d76818037b33d740fb7b0d11fd70ca6e33bfbe42: Status 404 returned error can't find the container with id 017881be1ccaf14fba331b26d76818037b33d740fb7b0d11fd70ca6e33bfbe42 Feb 18 09:13:19 crc kubenswrapper[4556]: I0218 09:13:19.924537 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-749d56bf4-c8bg9" event={"ID":"414515f8-ef03-41cd-a1c8-7811efcca959","Type":"ContainerStarted","Data":"017881be1ccaf14fba331b26d76818037b33d740fb7b0d11fd70ca6e33bfbe42"} Feb 18 09:13:19 crc kubenswrapper[4556]: I0218 09:13:19.926080 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-7cccb48f48-gbb9s" event={"ID":"6750b0ea-8f96-404d-bd38-e04e8be95127","Type":"ContainerStarted","Data":"861a7fc31aa040cc83423376fe11d824263788198cab0f1f9f924793b29985a1"} Feb 18 09:13:23 crc kubenswrapper[4556]: I0218 09:13:23.955318 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-7cccb48f48-gbb9s" event={"ID":"6750b0ea-8f96-404d-bd38-e04e8be95127","Type":"ContainerStarted","Data":"a1205eac48049dc19e8ba2e23b5c8e011e0bb4d1817e50a1fb4642889fd9a913"} Feb 18 09:13:23 crc kubenswrapper[4556]: I0218 09:13:23.955981 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-7cccb48f48-gbb9s" Feb 18 09:13:23 crc kubenswrapper[4556]: I0218 09:13:23.957073 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-749d56bf4-c8bg9" event={"ID":"414515f8-ef03-41cd-a1c8-7811efcca959","Type":"ContainerStarted","Data":"75f75de0b84119a4e04f27abfe7284cea92d8d8c101a33d0671c83aa49987f54"} Feb 18 09:13:23 crc kubenswrapper[4556]: I0218 09:13:23.957452 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-749d56bf4-c8bg9" Feb 18 09:13:23 crc kubenswrapper[4556]: I0218 09:13:23.970626 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-7cccb48f48-gbb9s" podStartSLOduration=1.114395432 podStartE2EDuration="4.9706135s" podCreationTimestamp="2026-02-18 09:13:19 +0000 UTC" firstStartedPulling="2026-02-18 09:13:19.675357807 +0000 UTC m=+556.692318787" lastFinishedPulling="2026-02-18 09:13:23.531575874 +0000 UTC m=+560.548536855" observedRunningTime="2026-02-18 09:13:23.969282989 +0000 UTC m=+560.986243968" watchObservedRunningTime="2026-02-18 09:13:23.9706135 +0000 UTC m=+560.987574480" Feb 18 09:13:23 crc kubenswrapper[4556]: I0218 09:13:23.985387 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-749d56bf4-c8bg9" podStartSLOduration=1.355643277 podStartE2EDuration="4.985376751s" podCreationTimestamp="2026-02-18 09:13:19 +0000 UTC" firstStartedPulling="2026-02-18 09:13:19.915071596 +0000 UTC m=+556.932032576" lastFinishedPulling="2026-02-18 09:13:23.54480507 +0000 UTC m=+560.561766050" observedRunningTime="2026-02-18 09:13:23.984375982 +0000 UTC m=+561.001336962" watchObservedRunningTime="2026-02-18 09:13:23.985376751 +0000 UTC m=+561.002337731" Feb 18 09:13:31 crc kubenswrapper[4556]: I0218 09:13:31.727646 4556 patch_prober.go:28] interesting pod/machine-config-daemon-f76hs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 09:13:31 crc kubenswrapper[4556]: I0218 09:13:31.728027 4556 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 09:13:31 crc kubenswrapper[4556]: I0218 09:13:31.728077 4556 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" Feb 18 09:13:31 crc kubenswrapper[4556]: I0218 09:13:31.728689 4556 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"dc67a95c5ea909b9cea5fe464915331393bc298bd4bcd198fcb5ebe8f8ba8a4b"} pod="openshift-machine-config-operator/machine-config-daemon-f76hs" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 18 09:13:31 crc kubenswrapper[4556]: I0218 09:13:31.728742 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerName="machine-config-daemon" containerID="cri-o://dc67a95c5ea909b9cea5fe464915331393bc298bd4bcd198fcb5ebe8f8ba8a4b" gracePeriod=600 Feb 18 09:13:32 crc kubenswrapper[4556]: I0218 09:13:32.010211 4556 generic.go:334] "Generic (PLEG): container finished" podID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerID="dc67a95c5ea909b9cea5fe464915331393bc298bd4bcd198fcb5ebe8f8ba8a4b" exitCode=0 Feb 18 09:13:32 crc kubenswrapper[4556]: I0218 09:13:32.010280 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" event={"ID":"8dac7f27-d3d1-4778-9e54-f273035a1d37","Type":"ContainerDied","Data":"dc67a95c5ea909b9cea5fe464915331393bc298bd4bcd198fcb5ebe8f8ba8a4b"} Feb 18 09:13:32 crc kubenswrapper[4556]: I0218 09:13:32.010543 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" event={"ID":"8dac7f27-d3d1-4778-9e54-f273035a1d37","Type":"ContainerStarted","Data":"5eaeb58ab20518b6122148754032f26a9e6729291f4bb5d77b32e55450f0b0ee"} Feb 18 09:13:32 crc kubenswrapper[4556]: I0218 09:13:32.010566 4556 scope.go:117] "RemoveContainer" containerID="8b54607784a82541a6af0430808ac511ecebb51bf04484e5e4e70f755a71305e" Feb 18 09:13:39 crc kubenswrapper[4556]: I0218 09:13:39.746252 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-749d56bf4-c8bg9" Feb 18 09:13:59 crc kubenswrapper[4556]: I0218 09:13:59.448364 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-7cccb48f48-gbb9s" Feb 18 09:13:59 crc kubenswrapper[4556]: I0218 09:13:59.963746 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-qltwk"] Feb 18 09:13:59 crc kubenswrapper[4556]: I0218 09:13:59.965725 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-qltwk" Feb 18 09:13:59 crc kubenswrapper[4556]: I0218 09:13:59.967671 4556 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Feb 18 09:13:59 crc kubenswrapper[4556]: I0218 09:13:59.967714 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Feb 18 09:13:59 crc kubenswrapper[4556]: I0218 09:13:59.969535 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-78b44bf5bb-spppg"] Feb 18 09:13:59 crc kubenswrapper[4556]: I0218 09:13:59.970060 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-spppg" Feb 18 09:13:59 crc kubenswrapper[4556]: I0218 09:13:59.972625 4556 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-6kk7c" Feb 18 09:13:59 crc kubenswrapper[4556]: I0218 09:13:59.972688 4556 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Feb 18 09:13:59 crc kubenswrapper[4556]: I0218 09:13:59.975437 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfkgm\" (UniqueName: \"kubernetes.io/projected/04e46782-3a29-4a3d-8b82-ff9f8d34090c-kube-api-access-vfkgm\") pod \"frr-k8s-webhook-server-78b44bf5bb-spppg\" (UID: \"04e46782-3a29-4a3d-8b82-ff9f8d34090c\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-spppg" Feb 18 09:13:59 crc kubenswrapper[4556]: I0218 09:13:59.975481 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/7cb5a650-ed14-42c7-8655-c5dfcaeb9d40-frr-conf\") pod \"frr-k8s-qltwk\" (UID: \"7cb5a650-ed14-42c7-8655-c5dfcaeb9d40\") " pod="metallb-system/frr-k8s-qltwk" Feb 18 09:13:59 crc kubenswrapper[4556]: I0218 09:13:59.975549 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/04e46782-3a29-4a3d-8b82-ff9f8d34090c-cert\") pod \"frr-k8s-webhook-server-78b44bf5bb-spppg\" (UID: \"04e46782-3a29-4a3d-8b82-ff9f8d34090c\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-spppg" Feb 18 09:13:59 crc kubenswrapper[4556]: I0218 09:13:59.975581 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qr6gf\" (UniqueName: \"kubernetes.io/projected/7cb5a650-ed14-42c7-8655-c5dfcaeb9d40-kube-api-access-qr6gf\") pod \"frr-k8s-qltwk\" (UID: \"7cb5a650-ed14-42c7-8655-c5dfcaeb9d40\") " pod="metallb-system/frr-k8s-qltwk" Feb 18 09:13:59 crc kubenswrapper[4556]: I0218 09:13:59.975659 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/7cb5a650-ed14-42c7-8655-c5dfcaeb9d40-frr-startup\") pod \"frr-k8s-qltwk\" (UID: \"7cb5a650-ed14-42c7-8655-c5dfcaeb9d40\") " pod="metallb-system/frr-k8s-qltwk" Feb 18 09:13:59 crc kubenswrapper[4556]: I0218 09:13:59.975753 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7cb5a650-ed14-42c7-8655-c5dfcaeb9d40-metrics-certs\") pod \"frr-k8s-qltwk\" (UID: \"7cb5a650-ed14-42c7-8655-c5dfcaeb9d40\") " pod="metallb-system/frr-k8s-qltwk" Feb 18 09:13:59 crc kubenswrapper[4556]: I0218 09:13:59.975802 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/7cb5a650-ed14-42c7-8655-c5dfcaeb9d40-frr-sockets\") pod \"frr-k8s-qltwk\" (UID: \"7cb5a650-ed14-42c7-8655-c5dfcaeb9d40\") " pod="metallb-system/frr-k8s-qltwk" Feb 18 09:13:59 crc kubenswrapper[4556]: I0218 09:13:59.975943 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/7cb5a650-ed14-42c7-8655-c5dfcaeb9d40-reloader\") pod \"frr-k8s-qltwk\" (UID: \"7cb5a650-ed14-42c7-8655-c5dfcaeb9d40\") " pod="metallb-system/frr-k8s-qltwk" Feb 18 09:13:59 crc kubenswrapper[4556]: I0218 09:13:59.976010 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/7cb5a650-ed14-42c7-8655-c5dfcaeb9d40-metrics\") pod \"frr-k8s-qltwk\" (UID: \"7cb5a650-ed14-42c7-8655-c5dfcaeb9d40\") " pod="metallb-system/frr-k8s-qltwk" Feb 18 09:13:59 crc kubenswrapper[4556]: I0218 09:13:59.980783 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-78b44bf5bb-spppg"] Feb 18 09:14:00 crc kubenswrapper[4556]: I0218 09:14:00.035339 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-cjpmm"] Feb 18 09:14:00 crc kubenswrapper[4556]: I0218 09:14:00.036112 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-cjpmm" Feb 18 09:14:00 crc kubenswrapper[4556]: I0218 09:14:00.037511 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Feb 18 09:14:00 crc kubenswrapper[4556]: I0218 09:14:00.037700 4556 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Feb 18 09:14:00 crc kubenswrapper[4556]: I0218 09:14:00.037852 4556 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-hs88t" Feb 18 09:14:00 crc kubenswrapper[4556]: I0218 09:14:00.038294 4556 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Feb 18 09:14:00 crc kubenswrapper[4556]: I0218 09:14:00.061212 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-69bbfbf88f-l7dg7"] Feb 18 09:14:00 crc kubenswrapper[4556]: I0218 09:14:00.062266 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-69bbfbf88f-l7dg7" Feb 18 09:14:00 crc kubenswrapper[4556]: I0218 09:14:00.067934 4556 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Feb 18 09:14:00 crc kubenswrapper[4556]: I0218 09:14:00.072209 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-69bbfbf88f-l7dg7"] Feb 18 09:14:00 crc kubenswrapper[4556]: I0218 09:14:00.077553 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/90e14ccf-fe48-4f1a-85af-5e7aa2cc1874-metrics-certs\") pod \"speaker-cjpmm\" (UID: \"90e14ccf-fe48-4f1a-85af-5e7aa2cc1874\") " pod="metallb-system/speaker-cjpmm" Feb 18 09:14:00 crc kubenswrapper[4556]: I0218 09:14:00.077618 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cb943421-d850-4010-817d-15920051ce04-metrics-certs\") pod \"controller-69bbfbf88f-l7dg7\" (UID: \"cb943421-d850-4010-817d-15920051ce04\") " pod="metallb-system/controller-69bbfbf88f-l7dg7" Feb 18 09:14:00 crc kubenswrapper[4556]: I0218 09:14:00.077660 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/04e46782-3a29-4a3d-8b82-ff9f8d34090c-cert\") pod \"frr-k8s-webhook-server-78b44bf5bb-spppg\" (UID: \"04e46782-3a29-4a3d-8b82-ff9f8d34090c\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-spppg" Feb 18 09:14:00 crc kubenswrapper[4556]: I0218 09:14:00.077689 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/90e14ccf-fe48-4f1a-85af-5e7aa2cc1874-memberlist\") pod \"speaker-cjpmm\" (UID: \"90e14ccf-fe48-4f1a-85af-5e7aa2cc1874\") " pod="metallb-system/speaker-cjpmm" Feb 18 09:14:00 crc kubenswrapper[4556]: I0218 09:14:00.077725 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qr6gf\" (UniqueName: \"kubernetes.io/projected/7cb5a650-ed14-42c7-8655-c5dfcaeb9d40-kube-api-access-qr6gf\") pod \"frr-k8s-qltwk\" (UID: \"7cb5a650-ed14-42c7-8655-c5dfcaeb9d40\") " pod="metallb-system/frr-k8s-qltwk" Feb 18 09:14:00 crc kubenswrapper[4556]: I0218 09:14:00.077772 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/7cb5a650-ed14-42c7-8655-c5dfcaeb9d40-frr-startup\") pod \"frr-k8s-qltwk\" (UID: \"7cb5a650-ed14-42c7-8655-c5dfcaeb9d40\") " pod="metallb-system/frr-k8s-qltwk" Feb 18 09:14:00 crc kubenswrapper[4556]: I0218 09:14:00.077818 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cb943421-d850-4010-817d-15920051ce04-cert\") pod \"controller-69bbfbf88f-l7dg7\" (UID: \"cb943421-d850-4010-817d-15920051ce04\") " pod="metallb-system/controller-69bbfbf88f-l7dg7" Feb 18 09:14:00 crc kubenswrapper[4556]: I0218 09:14:00.077844 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kcxc2\" (UniqueName: \"kubernetes.io/projected/90e14ccf-fe48-4f1a-85af-5e7aa2cc1874-kube-api-access-kcxc2\") pod \"speaker-cjpmm\" (UID: \"90e14ccf-fe48-4f1a-85af-5e7aa2cc1874\") " pod="metallb-system/speaker-cjpmm" Feb 18 09:14:00 crc kubenswrapper[4556]: E0218 09:14:00.077861 4556 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Feb 18 09:14:00 crc kubenswrapper[4556]: E0218 09:14:00.077930 4556 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Feb 18 09:14:00 crc kubenswrapper[4556]: E0218 09:14:00.077930 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/04e46782-3a29-4a3d-8b82-ff9f8d34090c-cert podName:04e46782-3a29-4a3d-8b82-ff9f8d34090c nodeName:}" failed. No retries permitted until 2026-02-18 09:14:00.577909842 +0000 UTC m=+597.594870823 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/04e46782-3a29-4a3d-8b82-ff9f8d34090c-cert") pod "frr-k8s-webhook-server-78b44bf5bb-spppg" (UID: "04e46782-3a29-4a3d-8b82-ff9f8d34090c") : secret "frr-k8s-webhook-server-cert" not found Feb 18 09:14:00 crc kubenswrapper[4556]: I0218 09:14:00.077866 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7cb5a650-ed14-42c7-8655-c5dfcaeb9d40-metrics-certs\") pod \"frr-k8s-qltwk\" (UID: \"7cb5a650-ed14-42c7-8655-c5dfcaeb9d40\") " pod="metallb-system/frr-k8s-qltwk" Feb 18 09:14:00 crc kubenswrapper[4556]: E0218 09:14:00.077978 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7cb5a650-ed14-42c7-8655-c5dfcaeb9d40-metrics-certs podName:7cb5a650-ed14-42c7-8655-c5dfcaeb9d40 nodeName:}" failed. No retries permitted until 2026-02-18 09:14:00.57796668 +0000 UTC m=+597.594927660 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7cb5a650-ed14-42c7-8655-c5dfcaeb9d40-metrics-certs") pod "frr-k8s-qltwk" (UID: "7cb5a650-ed14-42c7-8655-c5dfcaeb9d40") : secret "frr-k8s-certs-secret" not found Feb 18 09:14:00 crc kubenswrapper[4556]: I0218 09:14:00.078035 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/7cb5a650-ed14-42c7-8655-c5dfcaeb9d40-frr-sockets\") pod \"frr-k8s-qltwk\" (UID: \"7cb5a650-ed14-42c7-8655-c5dfcaeb9d40\") " pod="metallb-system/frr-k8s-qltwk" Feb 18 09:14:00 crc kubenswrapper[4556]: I0218 09:14:00.078116 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/90e14ccf-fe48-4f1a-85af-5e7aa2cc1874-metallb-excludel2\") pod \"speaker-cjpmm\" (UID: \"90e14ccf-fe48-4f1a-85af-5e7aa2cc1874\") " pod="metallb-system/speaker-cjpmm" Feb 18 09:14:00 crc kubenswrapper[4556]: I0218 09:14:00.078234 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/7cb5a650-ed14-42c7-8655-c5dfcaeb9d40-reloader\") pod \"frr-k8s-qltwk\" (UID: \"7cb5a650-ed14-42c7-8655-c5dfcaeb9d40\") " pod="metallb-system/frr-k8s-qltwk" Feb 18 09:14:00 crc kubenswrapper[4556]: I0218 09:14:00.078326 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/7cb5a650-ed14-42c7-8655-c5dfcaeb9d40-metrics\") pod \"frr-k8s-qltwk\" (UID: \"7cb5a650-ed14-42c7-8655-c5dfcaeb9d40\") " pod="metallb-system/frr-k8s-qltwk" Feb 18 09:14:00 crc kubenswrapper[4556]: I0218 09:14:00.078360 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfkgm\" (UniqueName: \"kubernetes.io/projected/04e46782-3a29-4a3d-8b82-ff9f8d34090c-kube-api-access-vfkgm\") pod \"frr-k8s-webhook-server-78b44bf5bb-spppg\" (UID: \"04e46782-3a29-4a3d-8b82-ff9f8d34090c\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-spppg" Feb 18 09:14:00 crc kubenswrapper[4556]: I0218 09:14:00.078380 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ssn7p\" (UniqueName: \"kubernetes.io/projected/cb943421-d850-4010-817d-15920051ce04-kube-api-access-ssn7p\") pod \"controller-69bbfbf88f-l7dg7\" (UID: \"cb943421-d850-4010-817d-15920051ce04\") " pod="metallb-system/controller-69bbfbf88f-l7dg7" Feb 18 09:14:00 crc kubenswrapper[4556]: I0218 09:14:00.078416 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/7cb5a650-ed14-42c7-8655-c5dfcaeb9d40-frr-conf\") pod \"frr-k8s-qltwk\" (UID: \"7cb5a650-ed14-42c7-8655-c5dfcaeb9d40\") " pod="metallb-system/frr-k8s-qltwk" Feb 18 09:14:00 crc kubenswrapper[4556]: I0218 09:14:00.078603 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/7cb5a650-ed14-42c7-8655-c5dfcaeb9d40-frr-sockets\") pod \"frr-k8s-qltwk\" (UID: \"7cb5a650-ed14-42c7-8655-c5dfcaeb9d40\") " pod="metallb-system/frr-k8s-qltwk" Feb 18 09:14:00 crc kubenswrapper[4556]: I0218 09:14:00.078652 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/7cb5a650-ed14-42c7-8655-c5dfcaeb9d40-reloader\") pod \"frr-k8s-qltwk\" (UID: \"7cb5a650-ed14-42c7-8655-c5dfcaeb9d40\") " pod="metallb-system/frr-k8s-qltwk" Feb 18 09:14:00 crc kubenswrapper[4556]: I0218 09:14:00.078678 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/7cb5a650-ed14-42c7-8655-c5dfcaeb9d40-frr-startup\") pod \"frr-k8s-qltwk\" (UID: \"7cb5a650-ed14-42c7-8655-c5dfcaeb9d40\") " pod="metallb-system/frr-k8s-qltwk" Feb 18 09:14:00 crc kubenswrapper[4556]: I0218 09:14:00.078716 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/7cb5a650-ed14-42c7-8655-c5dfcaeb9d40-frr-conf\") pod \"frr-k8s-qltwk\" (UID: \"7cb5a650-ed14-42c7-8655-c5dfcaeb9d40\") " pod="metallb-system/frr-k8s-qltwk" Feb 18 09:14:00 crc kubenswrapper[4556]: I0218 09:14:00.078814 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/7cb5a650-ed14-42c7-8655-c5dfcaeb9d40-metrics\") pod \"frr-k8s-qltwk\" (UID: \"7cb5a650-ed14-42c7-8655-c5dfcaeb9d40\") " pod="metallb-system/frr-k8s-qltwk" Feb 18 09:14:00 crc kubenswrapper[4556]: I0218 09:14:00.104436 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfkgm\" (UniqueName: \"kubernetes.io/projected/04e46782-3a29-4a3d-8b82-ff9f8d34090c-kube-api-access-vfkgm\") pod \"frr-k8s-webhook-server-78b44bf5bb-spppg\" (UID: \"04e46782-3a29-4a3d-8b82-ff9f8d34090c\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-spppg" Feb 18 09:14:00 crc kubenswrapper[4556]: I0218 09:14:00.120859 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qr6gf\" (UniqueName: \"kubernetes.io/projected/7cb5a650-ed14-42c7-8655-c5dfcaeb9d40-kube-api-access-qr6gf\") pod \"frr-k8s-qltwk\" (UID: \"7cb5a650-ed14-42c7-8655-c5dfcaeb9d40\") " pod="metallb-system/frr-k8s-qltwk" Feb 18 09:14:00 crc kubenswrapper[4556]: I0218 09:14:00.179121 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cb943421-d850-4010-817d-15920051ce04-cert\") pod \"controller-69bbfbf88f-l7dg7\" (UID: \"cb943421-d850-4010-817d-15920051ce04\") " pod="metallb-system/controller-69bbfbf88f-l7dg7" Feb 18 09:14:00 crc kubenswrapper[4556]: I0218 09:14:00.179181 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kcxc2\" (UniqueName: \"kubernetes.io/projected/90e14ccf-fe48-4f1a-85af-5e7aa2cc1874-kube-api-access-kcxc2\") pod \"speaker-cjpmm\" (UID: \"90e14ccf-fe48-4f1a-85af-5e7aa2cc1874\") " pod="metallb-system/speaker-cjpmm" Feb 18 09:14:00 crc kubenswrapper[4556]: I0218 09:14:00.179236 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/90e14ccf-fe48-4f1a-85af-5e7aa2cc1874-metallb-excludel2\") pod \"speaker-cjpmm\" (UID: \"90e14ccf-fe48-4f1a-85af-5e7aa2cc1874\") " pod="metallb-system/speaker-cjpmm" Feb 18 09:14:00 crc kubenswrapper[4556]: I0218 09:14:00.179338 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ssn7p\" (UniqueName: \"kubernetes.io/projected/cb943421-d850-4010-817d-15920051ce04-kube-api-access-ssn7p\") pod \"controller-69bbfbf88f-l7dg7\" (UID: \"cb943421-d850-4010-817d-15920051ce04\") " pod="metallb-system/controller-69bbfbf88f-l7dg7" Feb 18 09:14:00 crc kubenswrapper[4556]: I0218 09:14:00.179383 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/90e14ccf-fe48-4f1a-85af-5e7aa2cc1874-metrics-certs\") pod \"speaker-cjpmm\" (UID: \"90e14ccf-fe48-4f1a-85af-5e7aa2cc1874\") " pod="metallb-system/speaker-cjpmm" Feb 18 09:14:00 crc kubenswrapper[4556]: I0218 09:14:00.179403 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cb943421-d850-4010-817d-15920051ce04-metrics-certs\") pod \"controller-69bbfbf88f-l7dg7\" (UID: \"cb943421-d850-4010-817d-15920051ce04\") " pod="metallb-system/controller-69bbfbf88f-l7dg7" Feb 18 09:14:00 crc kubenswrapper[4556]: I0218 09:14:00.179444 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/90e14ccf-fe48-4f1a-85af-5e7aa2cc1874-memberlist\") pod \"speaker-cjpmm\" (UID: \"90e14ccf-fe48-4f1a-85af-5e7aa2cc1874\") " pod="metallb-system/speaker-cjpmm" Feb 18 09:14:00 crc kubenswrapper[4556]: E0218 09:14:00.179541 4556 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Feb 18 09:14:00 crc kubenswrapper[4556]: E0218 09:14:00.180099 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/90e14ccf-fe48-4f1a-85af-5e7aa2cc1874-memberlist podName:90e14ccf-fe48-4f1a-85af-5e7aa2cc1874 nodeName:}" failed. No retries permitted until 2026-02-18 09:14:00.680083121 +0000 UTC m=+597.697044101 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/90e14ccf-fe48-4f1a-85af-5e7aa2cc1874-memberlist") pod "speaker-cjpmm" (UID: "90e14ccf-fe48-4f1a-85af-5e7aa2cc1874") : secret "metallb-memberlist" not found Feb 18 09:14:00 crc kubenswrapper[4556]: I0218 09:14:00.180121 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/90e14ccf-fe48-4f1a-85af-5e7aa2cc1874-metallb-excludel2\") pod \"speaker-cjpmm\" (UID: \"90e14ccf-fe48-4f1a-85af-5e7aa2cc1874\") " pod="metallb-system/speaker-cjpmm" Feb 18 09:14:00 crc kubenswrapper[4556]: I0218 09:14:00.182876 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cb943421-d850-4010-817d-15920051ce04-metrics-certs\") pod \"controller-69bbfbf88f-l7dg7\" (UID: \"cb943421-d850-4010-817d-15920051ce04\") " pod="metallb-system/controller-69bbfbf88f-l7dg7" Feb 18 09:14:00 crc kubenswrapper[4556]: I0218 09:14:00.182968 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/90e14ccf-fe48-4f1a-85af-5e7aa2cc1874-metrics-certs\") pod \"speaker-cjpmm\" (UID: \"90e14ccf-fe48-4f1a-85af-5e7aa2cc1874\") " pod="metallb-system/speaker-cjpmm" Feb 18 09:14:00 crc kubenswrapper[4556]: I0218 09:14:00.183178 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cb943421-d850-4010-817d-15920051ce04-cert\") pod \"controller-69bbfbf88f-l7dg7\" (UID: \"cb943421-d850-4010-817d-15920051ce04\") " pod="metallb-system/controller-69bbfbf88f-l7dg7" Feb 18 09:14:00 crc kubenswrapper[4556]: I0218 09:14:00.192369 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kcxc2\" (UniqueName: \"kubernetes.io/projected/90e14ccf-fe48-4f1a-85af-5e7aa2cc1874-kube-api-access-kcxc2\") pod \"speaker-cjpmm\" (UID: \"90e14ccf-fe48-4f1a-85af-5e7aa2cc1874\") " pod="metallb-system/speaker-cjpmm" Feb 18 09:14:00 crc kubenswrapper[4556]: I0218 09:14:00.194846 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ssn7p\" (UniqueName: \"kubernetes.io/projected/cb943421-d850-4010-817d-15920051ce04-kube-api-access-ssn7p\") pod \"controller-69bbfbf88f-l7dg7\" (UID: \"cb943421-d850-4010-817d-15920051ce04\") " pod="metallb-system/controller-69bbfbf88f-l7dg7" Feb 18 09:14:00 crc kubenswrapper[4556]: I0218 09:14:00.374039 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-69bbfbf88f-l7dg7" Feb 18 09:14:00 crc kubenswrapper[4556]: I0218 09:14:00.522738 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-69bbfbf88f-l7dg7"] Feb 18 09:14:00 crc kubenswrapper[4556]: I0218 09:14:00.589746 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/04e46782-3a29-4a3d-8b82-ff9f8d34090c-cert\") pod \"frr-k8s-webhook-server-78b44bf5bb-spppg\" (UID: \"04e46782-3a29-4a3d-8b82-ff9f8d34090c\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-spppg" Feb 18 09:14:00 crc kubenswrapper[4556]: I0218 09:14:00.590006 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7cb5a650-ed14-42c7-8655-c5dfcaeb9d40-metrics-certs\") pod \"frr-k8s-qltwk\" (UID: \"7cb5a650-ed14-42c7-8655-c5dfcaeb9d40\") " pod="metallb-system/frr-k8s-qltwk" Feb 18 09:14:00 crc kubenswrapper[4556]: I0218 09:14:00.593001 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7cb5a650-ed14-42c7-8655-c5dfcaeb9d40-metrics-certs\") pod \"frr-k8s-qltwk\" (UID: \"7cb5a650-ed14-42c7-8655-c5dfcaeb9d40\") " pod="metallb-system/frr-k8s-qltwk" Feb 18 09:14:00 crc kubenswrapper[4556]: I0218 09:14:00.593286 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/04e46782-3a29-4a3d-8b82-ff9f8d34090c-cert\") pod \"frr-k8s-webhook-server-78b44bf5bb-spppg\" (UID: \"04e46782-3a29-4a3d-8b82-ff9f8d34090c\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-spppg" Feb 18 09:14:00 crc kubenswrapper[4556]: I0218 09:14:00.691035 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/90e14ccf-fe48-4f1a-85af-5e7aa2cc1874-memberlist\") pod \"speaker-cjpmm\" (UID: \"90e14ccf-fe48-4f1a-85af-5e7aa2cc1874\") " pod="metallb-system/speaker-cjpmm" Feb 18 09:14:00 crc kubenswrapper[4556]: I0218 09:14:00.694180 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/90e14ccf-fe48-4f1a-85af-5e7aa2cc1874-memberlist\") pod \"speaker-cjpmm\" (UID: \"90e14ccf-fe48-4f1a-85af-5e7aa2cc1874\") " pod="metallb-system/speaker-cjpmm" Feb 18 09:14:00 crc kubenswrapper[4556]: I0218 09:14:00.882770 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-qltwk" Feb 18 09:14:00 crc kubenswrapper[4556]: I0218 09:14:00.893847 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-spppg" Feb 18 09:14:00 crc kubenswrapper[4556]: I0218 09:14:00.949604 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-cjpmm" Feb 18 09:14:01 crc kubenswrapper[4556]: I0218 09:14:01.172721 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-69bbfbf88f-l7dg7" event={"ID":"cb943421-d850-4010-817d-15920051ce04","Type":"ContainerStarted","Data":"53710ae7f5f42ae8e40b862a52a01885d6f6f647d5d5245f05c45b6275c06adb"} Feb 18 09:14:01 crc kubenswrapper[4556]: I0218 09:14:01.173085 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-69bbfbf88f-l7dg7" Feb 18 09:14:01 crc kubenswrapper[4556]: I0218 09:14:01.173103 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-69bbfbf88f-l7dg7" event={"ID":"cb943421-d850-4010-817d-15920051ce04","Type":"ContainerStarted","Data":"58ca50c3cb17a9531bf70ff9462536c115f4ff3bf100b827c7e79b2dad95779f"} Feb 18 09:14:01 crc kubenswrapper[4556]: I0218 09:14:01.173119 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-69bbfbf88f-l7dg7" event={"ID":"cb943421-d850-4010-817d-15920051ce04","Type":"ContainerStarted","Data":"a784719ad00adde7e67e164acb1cd3e31c9f0e1b868d82b582a8e7494848d3a5"} Feb 18 09:14:01 crc kubenswrapper[4556]: I0218 09:14:01.174685 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-qltwk" event={"ID":"7cb5a650-ed14-42c7-8655-c5dfcaeb9d40","Type":"ContainerStarted","Data":"bffba56fe2ea103978304e2d30a1546ed43f026a1c0cddf495f53af28cdf0c68"} Feb 18 09:14:01 crc kubenswrapper[4556]: I0218 09:14:01.176203 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-cjpmm" event={"ID":"90e14ccf-fe48-4f1a-85af-5e7aa2cc1874","Type":"ContainerStarted","Data":"bf56c0e5dd061d78eb40723ae35682a47da72fec0fea590b7077a09f20cd64a7"} Feb 18 09:14:01 crc kubenswrapper[4556]: I0218 09:14:01.192076 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-69bbfbf88f-l7dg7" podStartSLOduration=1.192058826 podStartE2EDuration="1.192058826s" podCreationTimestamp="2026-02-18 09:14:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:14:01.187597333 +0000 UTC m=+598.204558313" watchObservedRunningTime="2026-02-18 09:14:01.192058826 +0000 UTC m=+598.209019805" Feb 18 09:14:01 crc kubenswrapper[4556]: I0218 09:14:01.343215 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-78b44bf5bb-spppg"] Feb 18 09:14:02 crc kubenswrapper[4556]: I0218 09:14:02.182663 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-cjpmm" event={"ID":"90e14ccf-fe48-4f1a-85af-5e7aa2cc1874","Type":"ContainerStarted","Data":"ee8da70d4c7e7d7a8656fb16a3adf55d7bf4219d85694efc7e2c1a38a9e06b19"} Feb 18 09:14:02 crc kubenswrapper[4556]: I0218 09:14:02.182701 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-cjpmm" event={"ID":"90e14ccf-fe48-4f1a-85af-5e7aa2cc1874","Type":"ContainerStarted","Data":"ff9666130dcbf044c0ae8f6bbea39cbe008078dee95f6d9d41aee93d37b6ad09"} Feb 18 09:14:02 crc kubenswrapper[4556]: I0218 09:14:02.182773 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-cjpmm" Feb 18 09:14:02 crc kubenswrapper[4556]: I0218 09:14:02.183338 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-spppg" event={"ID":"04e46782-3a29-4a3d-8b82-ff9f8d34090c","Type":"ContainerStarted","Data":"bb5df5812713b2df107e9f37eb6706f19e0b4bf82f741e8092e81e60fe6c271e"} Feb 18 09:14:02 crc kubenswrapper[4556]: I0218 09:14:02.196556 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-cjpmm" podStartSLOduration=2.196542637 podStartE2EDuration="2.196542637s" podCreationTimestamp="2026-02-18 09:14:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:14:02.194142165 +0000 UTC m=+599.211103145" watchObservedRunningTime="2026-02-18 09:14:02.196542637 +0000 UTC m=+599.213503617" Feb 18 09:14:07 crc kubenswrapper[4556]: I0218 09:14:07.216838 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-spppg" event={"ID":"04e46782-3a29-4a3d-8b82-ff9f8d34090c","Type":"ContainerStarted","Data":"a93fb109113eba522e40ee72d70dc461d6108670d02b082f526ecd9a40f85f0f"} Feb 18 09:14:07 crc kubenswrapper[4556]: I0218 09:14:07.217763 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-spppg" Feb 18 09:14:07 crc kubenswrapper[4556]: I0218 09:14:07.218827 4556 generic.go:334] "Generic (PLEG): container finished" podID="7cb5a650-ed14-42c7-8655-c5dfcaeb9d40" containerID="f73672d63a58110d8b88f7ec11dde503332065c848a8a4509992914e8704b615" exitCode=0 Feb 18 09:14:07 crc kubenswrapper[4556]: I0218 09:14:07.219074 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-qltwk" event={"ID":"7cb5a650-ed14-42c7-8655-c5dfcaeb9d40","Type":"ContainerDied","Data":"f73672d63a58110d8b88f7ec11dde503332065c848a8a4509992914e8704b615"} Feb 18 09:14:07 crc kubenswrapper[4556]: I0218 09:14:07.233683 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-spppg" podStartSLOduration=3.035486726 podStartE2EDuration="8.233666984s" podCreationTimestamp="2026-02-18 09:13:59 +0000 UTC" firstStartedPulling="2026-02-18 09:14:01.349364406 +0000 UTC m=+598.366325376" lastFinishedPulling="2026-02-18 09:14:06.547544654 +0000 UTC m=+603.564505634" observedRunningTime="2026-02-18 09:14:07.230902466 +0000 UTC m=+604.247863446" watchObservedRunningTime="2026-02-18 09:14:07.233666984 +0000 UTC m=+604.250627964" Feb 18 09:14:08 crc kubenswrapper[4556]: I0218 09:14:08.229007 4556 generic.go:334] "Generic (PLEG): container finished" podID="7cb5a650-ed14-42c7-8655-c5dfcaeb9d40" containerID="6b31c5ee5c8cdd4300c238774f1ec131b4ed0c774400b4793079a47dfa1fed12" exitCode=0 Feb 18 09:14:08 crc kubenswrapper[4556]: I0218 09:14:08.229077 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-qltwk" event={"ID":"7cb5a650-ed14-42c7-8655-c5dfcaeb9d40","Type":"ContainerDied","Data":"6b31c5ee5c8cdd4300c238774f1ec131b4ed0c774400b4793079a47dfa1fed12"} Feb 18 09:14:09 crc kubenswrapper[4556]: I0218 09:14:09.239515 4556 generic.go:334] "Generic (PLEG): container finished" podID="7cb5a650-ed14-42c7-8655-c5dfcaeb9d40" containerID="3554a364d409c0c7fc8c3054d23f4198487b27b5ad6cc54626bfd0984b5e50c6" exitCode=0 Feb 18 09:14:09 crc kubenswrapper[4556]: I0218 09:14:09.239583 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-qltwk" event={"ID":"7cb5a650-ed14-42c7-8655-c5dfcaeb9d40","Type":"ContainerDied","Data":"3554a364d409c0c7fc8c3054d23f4198487b27b5ad6cc54626bfd0984b5e50c6"} Feb 18 09:14:10 crc kubenswrapper[4556]: I0218 09:14:10.250352 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-qltwk" event={"ID":"7cb5a650-ed14-42c7-8655-c5dfcaeb9d40","Type":"ContainerStarted","Data":"ce3f6d764ee601c416f251c6252277e8192b01ac2cb30ce1930f6ac139168d11"} Feb 18 09:14:10 crc kubenswrapper[4556]: I0218 09:14:10.250751 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-qltwk" event={"ID":"7cb5a650-ed14-42c7-8655-c5dfcaeb9d40","Type":"ContainerStarted","Data":"063b184c72f9fa7c4a5c979dd2b9182d92054404441761e674d9c8f49ad7f709"} Feb 18 09:14:10 crc kubenswrapper[4556]: I0218 09:14:10.250766 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-qltwk" event={"ID":"7cb5a650-ed14-42c7-8655-c5dfcaeb9d40","Type":"ContainerStarted","Data":"c0df39de766c0420a53cf8a59f55396b87b61da1f5bb7acf8c8d064fd29ed782"} Feb 18 09:14:10 crc kubenswrapper[4556]: I0218 09:14:10.250779 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-qltwk" event={"ID":"7cb5a650-ed14-42c7-8655-c5dfcaeb9d40","Type":"ContainerStarted","Data":"604e7ad2f2d6e9f0e3bc9807eeb0c9a4ee3f51e287af135dd4e7eba44618e969"} Feb 18 09:14:10 crc kubenswrapper[4556]: I0218 09:14:10.250795 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-qltwk" Feb 18 09:14:10 crc kubenswrapper[4556]: I0218 09:14:10.250807 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-qltwk" event={"ID":"7cb5a650-ed14-42c7-8655-c5dfcaeb9d40","Type":"ContainerStarted","Data":"38f9927489cd4e3f1bceaf2e2768c2b9e97be07cb2301a4ed2c55c6b736dcea8"} Feb 18 09:14:10 crc kubenswrapper[4556]: I0218 09:14:10.250819 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-qltwk" event={"ID":"7cb5a650-ed14-42c7-8655-c5dfcaeb9d40","Type":"ContainerStarted","Data":"4a5dbffd1970248020827b9703dc485b7e0f309c2abf689d2399e03acf57a946"} Feb 18 09:14:10 crc kubenswrapper[4556]: I0218 09:14:10.276022 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-qltwk" podStartSLOduration=5.777151254 podStartE2EDuration="11.276006384s" podCreationTimestamp="2026-02-18 09:13:59 +0000 UTC" firstStartedPulling="2026-02-18 09:14:01.042649111 +0000 UTC m=+598.059610091" lastFinishedPulling="2026-02-18 09:14:06.541504241 +0000 UTC m=+603.558465221" observedRunningTime="2026-02-18 09:14:10.27199368 +0000 UTC m=+607.288954659" watchObservedRunningTime="2026-02-18 09:14:10.276006384 +0000 UTC m=+607.292967365" Feb 18 09:14:10 crc kubenswrapper[4556]: I0218 09:14:10.377736 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-69bbfbf88f-l7dg7" Feb 18 09:14:10 crc kubenswrapper[4556]: I0218 09:14:10.882965 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-qltwk" Feb 18 09:14:10 crc kubenswrapper[4556]: I0218 09:14:10.916343 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-qltwk" Feb 18 09:14:20 crc kubenswrapper[4556]: I0218 09:14:20.888506 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-qltwk" Feb 18 09:14:20 crc kubenswrapper[4556]: I0218 09:14:20.897628 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-spppg" Feb 18 09:14:20 crc kubenswrapper[4556]: I0218 09:14:20.953422 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-cjpmm" Feb 18 09:14:23 crc kubenswrapper[4556]: I0218 09:14:23.188780 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-kc59d"] Feb 18 09:14:23 crc kubenswrapper[4556]: I0218 09:14:23.189946 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-kc59d" Feb 18 09:14:23 crc kubenswrapper[4556]: I0218 09:14:23.196883 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-282r2" Feb 18 09:14:23 crc kubenswrapper[4556]: I0218 09:14:23.197242 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Feb 18 09:14:23 crc kubenswrapper[4556]: I0218 09:14:23.197202 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Feb 18 09:14:23 crc kubenswrapper[4556]: I0218 09:14:23.201118 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-kc59d"] Feb 18 09:14:23 crc kubenswrapper[4556]: I0218 09:14:23.220346 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rdvv\" (UniqueName: \"kubernetes.io/projected/6aedd285-bec7-471d-893c-3349e9b98e2c-kube-api-access-2rdvv\") pod \"openstack-operator-index-kc59d\" (UID: \"6aedd285-bec7-471d-893c-3349e9b98e2c\") " pod="openstack-operators/openstack-operator-index-kc59d" Feb 18 09:14:23 crc kubenswrapper[4556]: I0218 09:14:23.321222 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rdvv\" (UniqueName: \"kubernetes.io/projected/6aedd285-bec7-471d-893c-3349e9b98e2c-kube-api-access-2rdvv\") pod \"openstack-operator-index-kc59d\" (UID: \"6aedd285-bec7-471d-893c-3349e9b98e2c\") " pod="openstack-operators/openstack-operator-index-kc59d" Feb 18 09:14:23 crc kubenswrapper[4556]: I0218 09:14:23.342061 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rdvv\" (UniqueName: \"kubernetes.io/projected/6aedd285-bec7-471d-893c-3349e9b98e2c-kube-api-access-2rdvv\") pod \"openstack-operator-index-kc59d\" (UID: \"6aedd285-bec7-471d-893c-3349e9b98e2c\") " pod="openstack-operators/openstack-operator-index-kc59d" Feb 18 09:14:23 crc kubenswrapper[4556]: I0218 09:14:23.511719 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-kc59d" Feb 18 09:14:23 crc kubenswrapper[4556]: I0218 09:14:23.903021 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-kc59d"] Feb 18 09:14:24 crc kubenswrapper[4556]: I0218 09:14:24.331632 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-kc59d" event={"ID":"6aedd285-bec7-471d-893c-3349e9b98e2c","Type":"ContainerStarted","Data":"0d996b2d1f0b985499b540652065020503cf808ce001a1e54038e9983ecf717d"} Feb 18 09:14:25 crc kubenswrapper[4556]: I0218 09:14:25.339593 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-kc59d" event={"ID":"6aedd285-bec7-471d-893c-3349e9b98e2c","Type":"ContainerStarted","Data":"a1a294a503de9461171290a7dfc1db111db796d6cbc7e786fccbaf1fc9449177"} Feb 18 09:14:25 crc kubenswrapper[4556]: I0218 09:14:25.361496 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-kc59d" podStartSLOduration=1.435445425 podStartE2EDuration="2.361471769s" podCreationTimestamp="2026-02-18 09:14:23 +0000 UTC" firstStartedPulling="2026-02-18 09:14:23.909859033 +0000 UTC m=+620.926820003" lastFinishedPulling="2026-02-18 09:14:24.835885367 +0000 UTC m=+621.852846347" observedRunningTime="2026-02-18 09:14:25.355107374 +0000 UTC m=+622.372068355" watchObservedRunningTime="2026-02-18 09:14:25.361471769 +0000 UTC m=+622.378432739" Feb 18 09:14:26 crc kubenswrapper[4556]: I0218 09:14:26.562749 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-kc59d"] Feb 18 09:14:27 crc kubenswrapper[4556]: I0218 09:14:27.171828 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-tg2z4"] Feb 18 09:14:27 crc kubenswrapper[4556]: I0218 09:14:27.173118 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-tg2z4" Feb 18 09:14:27 crc kubenswrapper[4556]: I0218 09:14:27.179078 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-tg2z4"] Feb 18 09:14:27 crc kubenswrapper[4556]: I0218 09:14:27.277316 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qb9kr\" (UniqueName: \"kubernetes.io/projected/85052b3f-274e-437c-af72-fdd608552f7c-kube-api-access-qb9kr\") pod \"openstack-operator-index-tg2z4\" (UID: \"85052b3f-274e-437c-af72-fdd608552f7c\") " pod="openstack-operators/openstack-operator-index-tg2z4" Feb 18 09:14:27 crc kubenswrapper[4556]: I0218 09:14:27.348887 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-kc59d" podUID="6aedd285-bec7-471d-893c-3349e9b98e2c" containerName="registry-server" containerID="cri-o://a1a294a503de9461171290a7dfc1db111db796d6cbc7e786fccbaf1fc9449177" gracePeriod=2 Feb 18 09:14:27 crc kubenswrapper[4556]: I0218 09:14:27.378518 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qb9kr\" (UniqueName: \"kubernetes.io/projected/85052b3f-274e-437c-af72-fdd608552f7c-kube-api-access-qb9kr\") pod \"openstack-operator-index-tg2z4\" (UID: \"85052b3f-274e-437c-af72-fdd608552f7c\") " pod="openstack-operators/openstack-operator-index-tg2z4" Feb 18 09:14:27 crc kubenswrapper[4556]: I0218 09:14:27.395569 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qb9kr\" (UniqueName: \"kubernetes.io/projected/85052b3f-274e-437c-af72-fdd608552f7c-kube-api-access-qb9kr\") pod \"openstack-operator-index-tg2z4\" (UID: \"85052b3f-274e-437c-af72-fdd608552f7c\") " pod="openstack-operators/openstack-operator-index-tg2z4" Feb 18 09:14:27 crc kubenswrapper[4556]: I0218 09:14:27.487081 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-tg2z4" Feb 18 09:14:27 crc kubenswrapper[4556]: I0218 09:14:27.660654 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-kc59d" Feb 18 09:14:27 crc kubenswrapper[4556]: I0218 09:14:27.682849 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2rdvv\" (UniqueName: \"kubernetes.io/projected/6aedd285-bec7-471d-893c-3349e9b98e2c-kube-api-access-2rdvv\") pod \"6aedd285-bec7-471d-893c-3349e9b98e2c\" (UID: \"6aedd285-bec7-471d-893c-3349e9b98e2c\") " Feb 18 09:14:27 crc kubenswrapper[4556]: I0218 09:14:27.687752 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6aedd285-bec7-471d-893c-3349e9b98e2c-kube-api-access-2rdvv" (OuterVolumeSpecName: "kube-api-access-2rdvv") pod "6aedd285-bec7-471d-893c-3349e9b98e2c" (UID: "6aedd285-bec7-471d-893c-3349e9b98e2c"). InnerVolumeSpecName "kube-api-access-2rdvv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:14:27 crc kubenswrapper[4556]: I0218 09:14:27.784921 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2rdvv\" (UniqueName: \"kubernetes.io/projected/6aedd285-bec7-471d-893c-3349e9b98e2c-kube-api-access-2rdvv\") on node \"crc\" DevicePath \"\"" Feb 18 09:14:27 crc kubenswrapper[4556]: I0218 09:14:27.853879 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-tg2z4"] Feb 18 09:14:27 crc kubenswrapper[4556]: W0218 09:14:27.859443 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod85052b3f_274e_437c_af72_fdd608552f7c.slice/crio-2d2677087422e002795964a687037b4b4291e0d02bd9dac130abf911c601ef73 WatchSource:0}: Error finding container 2d2677087422e002795964a687037b4b4291e0d02bd9dac130abf911c601ef73: Status 404 returned error can't find the container with id 2d2677087422e002795964a687037b4b4291e0d02bd9dac130abf911c601ef73 Feb 18 09:14:28 crc kubenswrapper[4556]: I0218 09:14:28.355521 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-tg2z4" event={"ID":"85052b3f-274e-437c-af72-fdd608552f7c","Type":"ContainerStarted","Data":"2d2677087422e002795964a687037b4b4291e0d02bd9dac130abf911c601ef73"} Feb 18 09:14:28 crc kubenswrapper[4556]: I0218 09:14:28.357065 4556 generic.go:334] "Generic (PLEG): container finished" podID="6aedd285-bec7-471d-893c-3349e9b98e2c" containerID="a1a294a503de9461171290a7dfc1db111db796d6cbc7e786fccbaf1fc9449177" exitCode=0 Feb 18 09:14:28 crc kubenswrapper[4556]: I0218 09:14:28.357124 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-kc59d" event={"ID":"6aedd285-bec7-471d-893c-3349e9b98e2c","Type":"ContainerDied","Data":"a1a294a503de9461171290a7dfc1db111db796d6cbc7e786fccbaf1fc9449177"} Feb 18 09:14:28 crc kubenswrapper[4556]: I0218 09:14:28.357126 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-kc59d" Feb 18 09:14:28 crc kubenswrapper[4556]: I0218 09:14:28.357147 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-kc59d" event={"ID":"6aedd285-bec7-471d-893c-3349e9b98e2c","Type":"ContainerDied","Data":"0d996b2d1f0b985499b540652065020503cf808ce001a1e54038e9983ecf717d"} Feb 18 09:14:28 crc kubenswrapper[4556]: I0218 09:14:28.357209 4556 scope.go:117] "RemoveContainer" containerID="a1a294a503de9461171290a7dfc1db111db796d6cbc7e786fccbaf1fc9449177" Feb 18 09:14:28 crc kubenswrapper[4556]: I0218 09:14:28.387267 4556 scope.go:117] "RemoveContainer" containerID="a1a294a503de9461171290a7dfc1db111db796d6cbc7e786fccbaf1fc9449177" Feb 18 09:14:28 crc kubenswrapper[4556]: I0218 09:14:28.389916 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-kc59d"] Feb 18 09:14:28 crc kubenswrapper[4556]: E0218 09:14:28.391712 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1a294a503de9461171290a7dfc1db111db796d6cbc7e786fccbaf1fc9449177\": container with ID starting with a1a294a503de9461171290a7dfc1db111db796d6cbc7e786fccbaf1fc9449177 not found: ID does not exist" containerID="a1a294a503de9461171290a7dfc1db111db796d6cbc7e786fccbaf1fc9449177" Feb 18 09:14:28 crc kubenswrapper[4556]: I0218 09:14:28.391768 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1a294a503de9461171290a7dfc1db111db796d6cbc7e786fccbaf1fc9449177"} err="failed to get container status \"a1a294a503de9461171290a7dfc1db111db796d6cbc7e786fccbaf1fc9449177\": rpc error: code = NotFound desc = could not find container \"a1a294a503de9461171290a7dfc1db111db796d6cbc7e786fccbaf1fc9449177\": container with ID starting with a1a294a503de9461171290a7dfc1db111db796d6cbc7e786fccbaf1fc9449177 not found: ID does not exist" Feb 18 09:14:28 crc kubenswrapper[4556]: I0218 09:14:28.393175 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-kc59d"] Feb 18 09:14:29 crc kubenswrapper[4556]: I0218 09:14:29.289600 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6aedd285-bec7-471d-893c-3349e9b98e2c" path="/var/lib/kubelet/pods/6aedd285-bec7-471d-893c-3349e9b98e2c/volumes" Feb 18 09:14:29 crc kubenswrapper[4556]: I0218 09:14:29.367696 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-tg2z4" event={"ID":"85052b3f-274e-437c-af72-fdd608552f7c","Type":"ContainerStarted","Data":"fdc7134258182514072ffc0bdbdd40e7fc12823bcbb3c1fa98441ab0b7f0fc1d"} Feb 18 09:14:29 crc kubenswrapper[4556]: I0218 09:14:29.381921 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-tg2z4" podStartSLOduration=1.8500232730000001 podStartE2EDuration="2.381905921s" podCreationTimestamp="2026-02-18 09:14:27 +0000 UTC" firstStartedPulling="2026-02-18 09:14:27.863448752 +0000 UTC m=+624.880409732" lastFinishedPulling="2026-02-18 09:14:28.3953314 +0000 UTC m=+625.412292380" observedRunningTime="2026-02-18 09:14:29.379137235 +0000 UTC m=+626.396098215" watchObservedRunningTime="2026-02-18 09:14:29.381905921 +0000 UTC m=+626.398866901" Feb 18 09:14:37 crc kubenswrapper[4556]: I0218 09:14:37.487662 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-tg2z4" Feb 18 09:14:37 crc kubenswrapper[4556]: I0218 09:14:37.488282 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-tg2z4" Feb 18 09:14:37 crc kubenswrapper[4556]: I0218 09:14:37.513335 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-tg2z4" Feb 18 09:14:38 crc kubenswrapper[4556]: I0218 09:14:38.437121 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-tg2z4" Feb 18 09:14:43 crc kubenswrapper[4556]: I0218 09:14:43.797839 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967fk785"] Feb 18 09:14:43 crc kubenswrapper[4556]: E0218 09:14:43.798605 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6aedd285-bec7-471d-893c-3349e9b98e2c" containerName="registry-server" Feb 18 09:14:43 crc kubenswrapper[4556]: I0218 09:14:43.798619 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="6aedd285-bec7-471d-893c-3349e9b98e2c" containerName="registry-server" Feb 18 09:14:43 crc kubenswrapper[4556]: I0218 09:14:43.798740 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="6aedd285-bec7-471d-893c-3349e9b98e2c" containerName="registry-server" Feb 18 09:14:43 crc kubenswrapper[4556]: I0218 09:14:43.799436 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967fk785" Feb 18 09:14:43 crc kubenswrapper[4556]: I0218 09:14:43.801472 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-2tw24" Feb 18 09:14:43 crc kubenswrapper[4556]: I0218 09:14:43.805279 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967fk785"] Feb 18 09:14:43 crc kubenswrapper[4556]: I0218 09:14:43.878982 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b828da70-b43b-421e-a2b2-d1a2bd562ae4-util\") pod \"8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967fk785\" (UID: \"b828da70-b43b-421e-a2b2-d1a2bd562ae4\") " pod="openstack-operators/8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967fk785" Feb 18 09:14:43 crc kubenswrapper[4556]: I0218 09:14:43.879039 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b828da70-b43b-421e-a2b2-d1a2bd562ae4-bundle\") pod \"8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967fk785\" (UID: \"b828da70-b43b-421e-a2b2-d1a2bd562ae4\") " pod="openstack-operators/8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967fk785" Feb 18 09:14:43 crc kubenswrapper[4556]: I0218 09:14:43.879086 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8v9lq\" (UniqueName: \"kubernetes.io/projected/b828da70-b43b-421e-a2b2-d1a2bd562ae4-kube-api-access-8v9lq\") pod \"8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967fk785\" (UID: \"b828da70-b43b-421e-a2b2-d1a2bd562ae4\") " pod="openstack-operators/8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967fk785" Feb 18 09:14:43 crc kubenswrapper[4556]: I0218 09:14:43.979961 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b828da70-b43b-421e-a2b2-d1a2bd562ae4-util\") pod \"8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967fk785\" (UID: \"b828da70-b43b-421e-a2b2-d1a2bd562ae4\") " pod="openstack-operators/8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967fk785" Feb 18 09:14:43 crc kubenswrapper[4556]: I0218 09:14:43.980007 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b828da70-b43b-421e-a2b2-d1a2bd562ae4-bundle\") pod \"8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967fk785\" (UID: \"b828da70-b43b-421e-a2b2-d1a2bd562ae4\") " pod="openstack-operators/8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967fk785" Feb 18 09:14:43 crc kubenswrapper[4556]: I0218 09:14:43.980048 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8v9lq\" (UniqueName: \"kubernetes.io/projected/b828da70-b43b-421e-a2b2-d1a2bd562ae4-kube-api-access-8v9lq\") pod \"8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967fk785\" (UID: \"b828da70-b43b-421e-a2b2-d1a2bd562ae4\") " pod="openstack-operators/8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967fk785" Feb 18 09:14:43 crc kubenswrapper[4556]: I0218 09:14:43.980473 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b828da70-b43b-421e-a2b2-d1a2bd562ae4-util\") pod \"8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967fk785\" (UID: \"b828da70-b43b-421e-a2b2-d1a2bd562ae4\") " pod="openstack-operators/8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967fk785" Feb 18 09:14:43 crc kubenswrapper[4556]: I0218 09:14:43.980528 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b828da70-b43b-421e-a2b2-d1a2bd562ae4-bundle\") pod \"8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967fk785\" (UID: \"b828da70-b43b-421e-a2b2-d1a2bd562ae4\") " pod="openstack-operators/8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967fk785" Feb 18 09:14:43 crc kubenswrapper[4556]: I0218 09:14:43.995228 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8v9lq\" (UniqueName: \"kubernetes.io/projected/b828da70-b43b-421e-a2b2-d1a2bd562ae4-kube-api-access-8v9lq\") pod \"8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967fk785\" (UID: \"b828da70-b43b-421e-a2b2-d1a2bd562ae4\") " pod="openstack-operators/8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967fk785" Feb 18 09:14:44 crc kubenswrapper[4556]: I0218 09:14:44.115652 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967fk785" Feb 18 09:14:44 crc kubenswrapper[4556]: I0218 09:14:44.518845 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967fk785"] Feb 18 09:14:45 crc kubenswrapper[4556]: I0218 09:14:45.453089 4556 generic.go:334] "Generic (PLEG): container finished" podID="b828da70-b43b-421e-a2b2-d1a2bd562ae4" containerID="294d1daa93d97957a49caba311bd6e2720d4835ea44d0ecc87fab9a9f13e74ea" exitCode=0 Feb 18 09:14:45 crc kubenswrapper[4556]: I0218 09:14:45.453199 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967fk785" event={"ID":"b828da70-b43b-421e-a2b2-d1a2bd562ae4","Type":"ContainerDied","Data":"294d1daa93d97957a49caba311bd6e2720d4835ea44d0ecc87fab9a9f13e74ea"} Feb 18 09:14:45 crc kubenswrapper[4556]: I0218 09:14:45.453393 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967fk785" event={"ID":"b828da70-b43b-421e-a2b2-d1a2bd562ae4","Type":"ContainerStarted","Data":"f5bacfdf6e2bcc39e04c7efc397f140fe511b3c45b1d263ffbfa0b42f76f14ef"} Feb 18 09:14:46 crc kubenswrapper[4556]: I0218 09:14:46.460023 4556 generic.go:334] "Generic (PLEG): container finished" podID="b828da70-b43b-421e-a2b2-d1a2bd562ae4" containerID="0c9f2f38d89aea5a86dec9985403da7495775a7d6664670dd8b84b3b36ab581f" exitCode=0 Feb 18 09:14:46 crc kubenswrapper[4556]: I0218 09:14:46.460105 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967fk785" event={"ID":"b828da70-b43b-421e-a2b2-d1a2bd562ae4","Type":"ContainerDied","Data":"0c9f2f38d89aea5a86dec9985403da7495775a7d6664670dd8b84b3b36ab581f"} Feb 18 09:14:47 crc kubenswrapper[4556]: I0218 09:14:47.469383 4556 generic.go:334] "Generic (PLEG): container finished" podID="b828da70-b43b-421e-a2b2-d1a2bd562ae4" containerID="0a04bea34093d135585e120a86b2899b21160bff07010c6ea9f169b20e2adea4" exitCode=0 Feb 18 09:14:47 crc kubenswrapper[4556]: I0218 09:14:47.469433 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967fk785" event={"ID":"b828da70-b43b-421e-a2b2-d1a2bd562ae4","Type":"ContainerDied","Data":"0a04bea34093d135585e120a86b2899b21160bff07010c6ea9f169b20e2adea4"} Feb 18 09:14:48 crc kubenswrapper[4556]: I0218 09:14:48.690736 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967fk785" Feb 18 09:14:48 crc kubenswrapper[4556]: I0218 09:14:48.838696 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8v9lq\" (UniqueName: \"kubernetes.io/projected/b828da70-b43b-421e-a2b2-d1a2bd562ae4-kube-api-access-8v9lq\") pod \"b828da70-b43b-421e-a2b2-d1a2bd562ae4\" (UID: \"b828da70-b43b-421e-a2b2-d1a2bd562ae4\") " Feb 18 09:14:48 crc kubenswrapper[4556]: I0218 09:14:48.839033 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b828da70-b43b-421e-a2b2-d1a2bd562ae4-util\") pod \"b828da70-b43b-421e-a2b2-d1a2bd562ae4\" (UID: \"b828da70-b43b-421e-a2b2-d1a2bd562ae4\") " Feb 18 09:14:48 crc kubenswrapper[4556]: I0218 09:14:48.839115 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b828da70-b43b-421e-a2b2-d1a2bd562ae4-bundle\") pod \"b828da70-b43b-421e-a2b2-d1a2bd562ae4\" (UID: \"b828da70-b43b-421e-a2b2-d1a2bd562ae4\") " Feb 18 09:14:48 crc kubenswrapper[4556]: I0218 09:14:48.840087 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b828da70-b43b-421e-a2b2-d1a2bd562ae4-bundle" (OuterVolumeSpecName: "bundle") pod "b828da70-b43b-421e-a2b2-d1a2bd562ae4" (UID: "b828da70-b43b-421e-a2b2-d1a2bd562ae4"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:14:48 crc kubenswrapper[4556]: I0218 09:14:48.851974 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b828da70-b43b-421e-a2b2-d1a2bd562ae4-kube-api-access-8v9lq" (OuterVolumeSpecName: "kube-api-access-8v9lq") pod "b828da70-b43b-421e-a2b2-d1a2bd562ae4" (UID: "b828da70-b43b-421e-a2b2-d1a2bd562ae4"). InnerVolumeSpecName "kube-api-access-8v9lq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:14:48 crc kubenswrapper[4556]: I0218 09:14:48.855018 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b828da70-b43b-421e-a2b2-d1a2bd562ae4-util" (OuterVolumeSpecName: "util") pod "b828da70-b43b-421e-a2b2-d1a2bd562ae4" (UID: "b828da70-b43b-421e-a2b2-d1a2bd562ae4"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:14:48 crc kubenswrapper[4556]: I0218 09:14:48.941672 4556 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b828da70-b43b-421e-a2b2-d1a2bd562ae4-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:14:48 crc kubenswrapper[4556]: I0218 09:14:48.941710 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8v9lq\" (UniqueName: \"kubernetes.io/projected/b828da70-b43b-421e-a2b2-d1a2bd562ae4-kube-api-access-8v9lq\") on node \"crc\" DevicePath \"\"" Feb 18 09:14:48 crc kubenswrapper[4556]: I0218 09:14:48.941724 4556 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b828da70-b43b-421e-a2b2-d1a2bd562ae4-util\") on node \"crc\" DevicePath \"\"" Feb 18 09:14:49 crc kubenswrapper[4556]: I0218 09:14:49.481297 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967fk785" event={"ID":"b828da70-b43b-421e-a2b2-d1a2bd562ae4","Type":"ContainerDied","Data":"f5bacfdf6e2bcc39e04c7efc397f140fe511b3c45b1d263ffbfa0b42f76f14ef"} Feb 18 09:14:49 crc kubenswrapper[4556]: I0218 09:14:49.481331 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967fk785" Feb 18 09:14:49 crc kubenswrapper[4556]: I0218 09:14:49.481341 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f5bacfdf6e2bcc39e04c7efc397f140fe511b3c45b1d263ffbfa0b42f76f14ef" Feb 18 09:14:51 crc kubenswrapper[4556]: I0218 09:14:51.040240 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-init-6679bf9b57-r8zc2"] Feb 18 09:14:51 crc kubenswrapper[4556]: E0218 09:14:51.040835 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b828da70-b43b-421e-a2b2-d1a2bd562ae4" containerName="util" Feb 18 09:14:51 crc kubenswrapper[4556]: I0218 09:14:51.040848 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="b828da70-b43b-421e-a2b2-d1a2bd562ae4" containerName="util" Feb 18 09:14:51 crc kubenswrapper[4556]: E0218 09:14:51.040862 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b828da70-b43b-421e-a2b2-d1a2bd562ae4" containerName="extract" Feb 18 09:14:51 crc kubenswrapper[4556]: I0218 09:14:51.040868 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="b828da70-b43b-421e-a2b2-d1a2bd562ae4" containerName="extract" Feb 18 09:14:51 crc kubenswrapper[4556]: E0218 09:14:51.040879 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b828da70-b43b-421e-a2b2-d1a2bd562ae4" containerName="pull" Feb 18 09:14:51 crc kubenswrapper[4556]: I0218 09:14:51.040885 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="b828da70-b43b-421e-a2b2-d1a2bd562ae4" containerName="pull" Feb 18 09:14:51 crc kubenswrapper[4556]: I0218 09:14:51.041001 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="b828da70-b43b-421e-a2b2-d1a2bd562ae4" containerName="extract" Feb 18 09:14:51 crc kubenswrapper[4556]: I0218 09:14:51.041386 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-6679bf9b57-r8zc2" Feb 18 09:14:51 crc kubenswrapper[4556]: I0218 09:14:51.043248 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-init-dockercfg-l9rrv" Feb 18 09:14:51 crc kubenswrapper[4556]: I0218 09:14:51.064302 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-6679bf9b57-r8zc2"] Feb 18 09:14:51 crc kubenswrapper[4556]: I0218 09:14:51.066381 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rcrqs\" (UniqueName: \"kubernetes.io/projected/acd47be1-151c-4fd9-9cc6-f49e1cc3f403-kube-api-access-rcrqs\") pod \"openstack-operator-controller-init-6679bf9b57-r8zc2\" (UID: \"acd47be1-151c-4fd9-9cc6-f49e1cc3f403\") " pod="openstack-operators/openstack-operator-controller-init-6679bf9b57-r8zc2" Feb 18 09:14:51 crc kubenswrapper[4556]: I0218 09:14:51.167324 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rcrqs\" (UniqueName: \"kubernetes.io/projected/acd47be1-151c-4fd9-9cc6-f49e1cc3f403-kube-api-access-rcrqs\") pod \"openstack-operator-controller-init-6679bf9b57-r8zc2\" (UID: \"acd47be1-151c-4fd9-9cc6-f49e1cc3f403\") " pod="openstack-operators/openstack-operator-controller-init-6679bf9b57-r8zc2" Feb 18 09:14:51 crc kubenswrapper[4556]: I0218 09:14:51.185631 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rcrqs\" (UniqueName: \"kubernetes.io/projected/acd47be1-151c-4fd9-9cc6-f49e1cc3f403-kube-api-access-rcrqs\") pod \"openstack-operator-controller-init-6679bf9b57-r8zc2\" (UID: \"acd47be1-151c-4fd9-9cc6-f49e1cc3f403\") " pod="openstack-operators/openstack-operator-controller-init-6679bf9b57-r8zc2" Feb 18 09:14:51 crc kubenswrapper[4556]: I0218 09:14:51.356062 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-6679bf9b57-r8zc2" Feb 18 09:14:51 crc kubenswrapper[4556]: I0218 09:14:51.727186 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-6679bf9b57-r8zc2"] Feb 18 09:14:52 crc kubenswrapper[4556]: I0218 09:14:52.497850 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-6679bf9b57-r8zc2" event={"ID":"acd47be1-151c-4fd9-9cc6-f49e1cc3f403","Type":"ContainerStarted","Data":"e9e3415e4a264b335807904c19623255f6e51c4106ad25866134793b8b15d3cf"} Feb 18 09:14:55 crc kubenswrapper[4556]: I0218 09:14:55.519747 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-6679bf9b57-r8zc2" event={"ID":"acd47be1-151c-4fd9-9cc6-f49e1cc3f403","Type":"ContainerStarted","Data":"89c55bd8002f3c2099ddc638bd18404c964620147a0bbf20d28d2dfff47c8641"} Feb 18 09:14:55 crc kubenswrapper[4556]: I0218 09:14:55.520107 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-init-6679bf9b57-r8zc2" Feb 18 09:14:55 crc kubenswrapper[4556]: I0218 09:14:55.541755 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-init-6679bf9b57-r8zc2" podStartSLOduration=1.345628311 podStartE2EDuration="4.541737799s" podCreationTimestamp="2026-02-18 09:14:51 +0000 UTC" firstStartedPulling="2026-02-18 09:14:51.733284565 +0000 UTC m=+648.750245534" lastFinishedPulling="2026-02-18 09:14:54.929394043 +0000 UTC m=+651.946355022" observedRunningTime="2026-02-18 09:14:55.539707166 +0000 UTC m=+652.556668146" watchObservedRunningTime="2026-02-18 09:14:55.541737799 +0000 UTC m=+652.558698779" Feb 18 09:15:00 crc kubenswrapper[4556]: I0218 09:15:00.136966 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29523435-9ztbm"] Feb 18 09:15:00 crc kubenswrapper[4556]: I0218 09:15:00.138930 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29523435-9ztbm" Feb 18 09:15:00 crc kubenswrapper[4556]: I0218 09:15:00.140618 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 18 09:15:00 crc kubenswrapper[4556]: I0218 09:15:00.140622 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 18 09:15:00 crc kubenswrapper[4556]: I0218 09:15:00.145432 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29523435-9ztbm"] Feb 18 09:15:00 crc kubenswrapper[4556]: I0218 09:15:00.188037 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/62835cbc-94a5-4b92-9b0c-eb44d39a5af5-config-volume\") pod \"collect-profiles-29523435-9ztbm\" (UID: \"62835cbc-94a5-4b92-9b0c-eb44d39a5af5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523435-9ztbm" Feb 18 09:15:00 crc kubenswrapper[4556]: I0218 09:15:00.188100 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8hwrs\" (UniqueName: \"kubernetes.io/projected/62835cbc-94a5-4b92-9b0c-eb44d39a5af5-kube-api-access-8hwrs\") pod \"collect-profiles-29523435-9ztbm\" (UID: \"62835cbc-94a5-4b92-9b0c-eb44d39a5af5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523435-9ztbm" Feb 18 09:15:00 crc kubenswrapper[4556]: I0218 09:15:00.188148 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/62835cbc-94a5-4b92-9b0c-eb44d39a5af5-secret-volume\") pod \"collect-profiles-29523435-9ztbm\" (UID: \"62835cbc-94a5-4b92-9b0c-eb44d39a5af5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523435-9ztbm" Feb 18 09:15:00 crc kubenswrapper[4556]: I0218 09:15:00.289612 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8hwrs\" (UniqueName: \"kubernetes.io/projected/62835cbc-94a5-4b92-9b0c-eb44d39a5af5-kube-api-access-8hwrs\") pod \"collect-profiles-29523435-9ztbm\" (UID: \"62835cbc-94a5-4b92-9b0c-eb44d39a5af5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523435-9ztbm" Feb 18 09:15:00 crc kubenswrapper[4556]: I0218 09:15:00.289666 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/62835cbc-94a5-4b92-9b0c-eb44d39a5af5-secret-volume\") pod \"collect-profiles-29523435-9ztbm\" (UID: \"62835cbc-94a5-4b92-9b0c-eb44d39a5af5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523435-9ztbm" Feb 18 09:15:00 crc kubenswrapper[4556]: I0218 09:15:00.289725 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/62835cbc-94a5-4b92-9b0c-eb44d39a5af5-config-volume\") pod \"collect-profiles-29523435-9ztbm\" (UID: \"62835cbc-94a5-4b92-9b0c-eb44d39a5af5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523435-9ztbm" Feb 18 09:15:00 crc kubenswrapper[4556]: I0218 09:15:00.290473 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/62835cbc-94a5-4b92-9b0c-eb44d39a5af5-config-volume\") pod \"collect-profiles-29523435-9ztbm\" (UID: \"62835cbc-94a5-4b92-9b0c-eb44d39a5af5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523435-9ztbm" Feb 18 09:15:00 crc kubenswrapper[4556]: I0218 09:15:00.295079 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/62835cbc-94a5-4b92-9b0c-eb44d39a5af5-secret-volume\") pod \"collect-profiles-29523435-9ztbm\" (UID: \"62835cbc-94a5-4b92-9b0c-eb44d39a5af5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523435-9ztbm" Feb 18 09:15:00 crc kubenswrapper[4556]: I0218 09:15:00.304050 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8hwrs\" (UniqueName: \"kubernetes.io/projected/62835cbc-94a5-4b92-9b0c-eb44d39a5af5-kube-api-access-8hwrs\") pod \"collect-profiles-29523435-9ztbm\" (UID: \"62835cbc-94a5-4b92-9b0c-eb44d39a5af5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523435-9ztbm" Feb 18 09:15:00 crc kubenswrapper[4556]: I0218 09:15:00.456681 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29523435-9ztbm" Feb 18 09:15:00 crc kubenswrapper[4556]: I0218 09:15:00.800043 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29523435-9ztbm"] Feb 18 09:15:01 crc kubenswrapper[4556]: I0218 09:15:01.358977 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-init-6679bf9b57-r8zc2" Feb 18 09:15:01 crc kubenswrapper[4556]: I0218 09:15:01.557239 4556 generic.go:334] "Generic (PLEG): container finished" podID="62835cbc-94a5-4b92-9b0c-eb44d39a5af5" containerID="7e284035a6bf3b5a640f304fc8621b89e3a15039dd83d408395f86d674659470" exitCode=0 Feb 18 09:15:01 crc kubenswrapper[4556]: I0218 09:15:01.557498 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29523435-9ztbm" event={"ID":"62835cbc-94a5-4b92-9b0c-eb44d39a5af5","Type":"ContainerDied","Data":"7e284035a6bf3b5a640f304fc8621b89e3a15039dd83d408395f86d674659470"} Feb 18 09:15:01 crc kubenswrapper[4556]: I0218 09:15:01.557610 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29523435-9ztbm" event={"ID":"62835cbc-94a5-4b92-9b0c-eb44d39a5af5","Type":"ContainerStarted","Data":"64368645e2d2b5d8d1a60633843eb7370c66c291de650d9964541b1b430079c2"} Feb 18 09:15:02 crc kubenswrapper[4556]: I0218 09:15:02.800980 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29523435-9ztbm" Feb 18 09:15:02 crc kubenswrapper[4556]: I0218 09:15:02.824582 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/62835cbc-94a5-4b92-9b0c-eb44d39a5af5-config-volume\") pod \"62835cbc-94a5-4b92-9b0c-eb44d39a5af5\" (UID: \"62835cbc-94a5-4b92-9b0c-eb44d39a5af5\") " Feb 18 09:15:02 crc kubenswrapper[4556]: I0218 09:15:02.824677 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/62835cbc-94a5-4b92-9b0c-eb44d39a5af5-secret-volume\") pod \"62835cbc-94a5-4b92-9b0c-eb44d39a5af5\" (UID: \"62835cbc-94a5-4b92-9b0c-eb44d39a5af5\") " Feb 18 09:15:02 crc kubenswrapper[4556]: I0218 09:15:02.824723 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8hwrs\" (UniqueName: \"kubernetes.io/projected/62835cbc-94a5-4b92-9b0c-eb44d39a5af5-kube-api-access-8hwrs\") pod \"62835cbc-94a5-4b92-9b0c-eb44d39a5af5\" (UID: \"62835cbc-94a5-4b92-9b0c-eb44d39a5af5\") " Feb 18 09:15:02 crc kubenswrapper[4556]: I0218 09:15:02.825253 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62835cbc-94a5-4b92-9b0c-eb44d39a5af5-config-volume" (OuterVolumeSpecName: "config-volume") pod "62835cbc-94a5-4b92-9b0c-eb44d39a5af5" (UID: "62835cbc-94a5-4b92-9b0c-eb44d39a5af5"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:15:02 crc kubenswrapper[4556]: I0218 09:15:02.829533 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62835cbc-94a5-4b92-9b0c-eb44d39a5af5-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "62835cbc-94a5-4b92-9b0c-eb44d39a5af5" (UID: "62835cbc-94a5-4b92-9b0c-eb44d39a5af5"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:15:02 crc kubenswrapper[4556]: I0218 09:15:02.829571 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62835cbc-94a5-4b92-9b0c-eb44d39a5af5-kube-api-access-8hwrs" (OuterVolumeSpecName: "kube-api-access-8hwrs") pod "62835cbc-94a5-4b92-9b0c-eb44d39a5af5" (UID: "62835cbc-94a5-4b92-9b0c-eb44d39a5af5"). InnerVolumeSpecName "kube-api-access-8hwrs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:15:02 crc kubenswrapper[4556]: I0218 09:15:02.926645 4556 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/62835cbc-94a5-4b92-9b0c-eb44d39a5af5-config-volume\") on node \"crc\" DevicePath \"\"" Feb 18 09:15:02 crc kubenswrapper[4556]: I0218 09:15:02.926676 4556 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/62835cbc-94a5-4b92-9b0c-eb44d39a5af5-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 18 09:15:02 crc kubenswrapper[4556]: I0218 09:15:02.926687 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8hwrs\" (UniqueName: \"kubernetes.io/projected/62835cbc-94a5-4b92-9b0c-eb44d39a5af5-kube-api-access-8hwrs\") on node \"crc\" DevicePath \"\"" Feb 18 09:15:03 crc kubenswrapper[4556]: I0218 09:15:03.571410 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29523435-9ztbm" event={"ID":"62835cbc-94a5-4b92-9b0c-eb44d39a5af5","Type":"ContainerDied","Data":"64368645e2d2b5d8d1a60633843eb7370c66c291de650d9964541b1b430079c2"} Feb 18 09:15:03 crc kubenswrapper[4556]: I0218 09:15:03.571764 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="64368645e2d2b5d8d1a60633843eb7370c66c291de650d9964541b1b430079c2" Feb 18 09:15:03 crc kubenswrapper[4556]: I0218 09:15:03.571459 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29523435-9ztbm" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.169498 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-868647ff47-x74c8"] Feb 18 09:15:20 crc kubenswrapper[4556]: E0218 09:15:20.170181 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62835cbc-94a5-4b92-9b0c-eb44d39a5af5" containerName="collect-profiles" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.170197 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="62835cbc-94a5-4b92-9b0c-eb44d39a5af5" containerName="collect-profiles" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.170424 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="62835cbc-94a5-4b92-9b0c-eb44d39a5af5" containerName="collect-profiles" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.170914 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-x74c8" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.175351 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-vkjch" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.184408 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-868647ff47-x74c8"] Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.188072 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-5d946d989d-n4x7d"] Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.188827 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-n4x7d" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.190349 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-qwtv4" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.197220 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-6d8bf5c495-5pv48"] Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.197792 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-5pv48" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.199683 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-zvqr4" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.200808 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-6d8bf5c495-5pv48"] Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.215235 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987464f4-mfhjc"] Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.215942 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987464f4-mfhjc" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.218787 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-m6mpf" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.228198 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-5d946d989d-n4x7d"] Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.231460 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-69f49c598c-6sck2"] Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.235038 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987464f4-mfhjc"] Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.235125 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-6sck2" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.237227 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-64gtm" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.242858 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8z9fq\" (UniqueName: \"kubernetes.io/projected/463e8b44-a344-401e-bd53-22738560ca32-kube-api-access-8z9fq\") pod \"cinder-operator-controller-manager-5d946d989d-n4x7d\" (UID: \"463e8b44-a344-401e-bd53-22738560ca32\") " pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-n4x7d" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.242983 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mdpg\" (UniqueName: \"kubernetes.io/projected/fff15cfd-9701-4efa-81a6-9a482e09ca0f-kube-api-access-8mdpg\") pod \"designate-operator-controller-manager-6d8bf5c495-5pv48\" (UID: \"fff15cfd-9701-4efa-81a6-9a482e09ca0f\") " pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-5pv48" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.243075 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6n7hw\" (UniqueName: \"kubernetes.io/projected/a980ea11-bf5f-4967-b22b-3e63454bb9ae-kube-api-access-6n7hw\") pod \"heat-operator-controller-manager-69f49c598c-6sck2\" (UID: \"a980ea11-bf5f-4967-b22b-3e63454bb9ae\") " pod="openstack-operators/heat-operator-controller-manager-69f49c598c-6sck2" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.243174 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4lmg\" (UniqueName: \"kubernetes.io/projected/f194b6f2-c143-4a57-8d3e-378a08147713-kube-api-access-f4lmg\") pod \"glance-operator-controller-manager-77987464f4-mfhjc\" (UID: \"f194b6f2-c143-4a57-8d3e-378a08147713\") " pod="openstack-operators/glance-operator-controller-manager-77987464f4-mfhjc" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.243288 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n66pr\" (UniqueName: \"kubernetes.io/projected/9cf84ffa-7a16-4e6a-aa7a-c90d6c1635fe-kube-api-access-n66pr\") pod \"barbican-operator-controller-manager-868647ff47-x74c8\" (UID: \"9cf84ffa-7a16-4e6a-aa7a-c90d6c1635fe\") " pod="openstack-operators/barbican-operator-controller-manager-868647ff47-x74c8" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.243216 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5b9b8895d5-g6hvq"] Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.244084 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-g6hvq" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.245560 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-br7zh" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.257577 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-69f49c598c-6sck2"] Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.263511 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-79d975b745-gp4rt"] Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.264168 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-79d975b745-gp4rt" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.265689 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-kl7fq" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.266006 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.266725 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5b9b8895d5-g6hvq"] Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.276776 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-79d975b745-gp4rt"] Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.279880 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-554564d7fc-lktrc"] Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.280621 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-lktrc" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.282412 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-qv2gz" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.291127 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-554564d7fc-lktrc"] Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.294231 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-b4d948c87-9rkqc"] Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.295007 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-9rkqc" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.297422 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-6c5dc" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.304667 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-54f6768c69-r5kw5"] Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.305446 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-b4d948c87-9rkqc"] Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.305536 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-r5kw5" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.309452 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-sbzf9" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.310387 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6994f66f48-t2ng7"] Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.311010 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-t2ng7" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.312006 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-7nkl6" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.313649 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-54f6768c69-r5kw5"] Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.320374 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6994f66f48-t2ng7"] Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.325748 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-64ddbf8bb-wdp2t"] Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.326518 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-wdp2t" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.328521 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-gblpq" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.329418 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-64ddbf8bb-wdp2t"] Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.337798 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-567668f5cf-t9gtc"] Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.339144 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-t9gtc" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.340320 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-k2x6s" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.341588 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-567668f5cf-t9gtc"] Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.345624 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n66pr\" (UniqueName: \"kubernetes.io/projected/9cf84ffa-7a16-4e6a-aa7a-c90d6c1635fe-kube-api-access-n66pr\") pod \"barbican-operator-controller-manager-868647ff47-x74c8\" (UID: \"9cf84ffa-7a16-4e6a-aa7a-c90d6c1635fe\") " pod="openstack-operators/barbican-operator-controller-manager-868647ff47-x74c8" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.345708 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8z9fq\" (UniqueName: \"kubernetes.io/projected/463e8b44-a344-401e-bd53-22738560ca32-kube-api-access-8z9fq\") pod \"cinder-operator-controller-manager-5d946d989d-n4x7d\" (UID: \"463e8b44-a344-401e-bd53-22738560ca32\") " pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-n4x7d" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.345734 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mdpg\" (UniqueName: \"kubernetes.io/projected/fff15cfd-9701-4efa-81a6-9a482e09ca0f-kube-api-access-8mdpg\") pod \"designate-operator-controller-manager-6d8bf5c495-5pv48\" (UID: \"fff15cfd-9701-4efa-81a6-9a482e09ca0f\") " pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-5pv48" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.345763 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6n7hw\" (UniqueName: \"kubernetes.io/projected/a980ea11-bf5f-4967-b22b-3e63454bb9ae-kube-api-access-6n7hw\") pod \"heat-operator-controller-manager-69f49c598c-6sck2\" (UID: \"a980ea11-bf5f-4967-b22b-3e63454bb9ae\") " pod="openstack-operators/heat-operator-controller-manager-69f49c598c-6sck2" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.345786 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4lmg\" (UniqueName: \"kubernetes.io/projected/f194b6f2-c143-4a57-8d3e-378a08147713-kube-api-access-f4lmg\") pod \"glance-operator-controller-manager-77987464f4-mfhjc\" (UID: \"f194b6f2-c143-4a57-8d3e-378a08147713\") " pod="openstack-operators/glance-operator-controller-manager-77987464f4-mfhjc" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.358212 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69f8888797-rpk9w"] Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.359000 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-rpk9w" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.362124 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-mjhjz" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.367467 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-d44cf6b75-qr5np"] Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.368108 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-qr5np" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.370641 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69f8888797-rpk9w"] Feb 18 09:15:20 crc kubenswrapper[4556]: W0218 09:15:20.373043 4556 reflector.go:561] object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-dxsw4": failed to list *v1.Secret: secrets "ovn-operator-controller-manager-dockercfg-dxsw4" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openstack-operators": no relationship found between node 'crc' and this object Feb 18 09:15:20 crc kubenswrapper[4556]: E0218 09:15:20.373128 4556 reflector.go:158] "Unhandled Error" err="object-\"openstack-operators\"/\"ovn-operator-controller-manager-dockercfg-dxsw4\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"ovn-operator-controller-manager-dockercfg-dxsw4\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openstack-operators\": no relationship found between node 'crc' and this object" logger="UnhandledError" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.374026 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n66pr\" (UniqueName: \"kubernetes.io/projected/9cf84ffa-7a16-4e6a-aa7a-c90d6c1635fe-kube-api-access-n66pr\") pod \"barbican-operator-controller-manager-868647ff47-x74c8\" (UID: \"9cf84ffa-7a16-4e6a-aa7a-c90d6c1635fe\") " pod="openstack-operators/barbican-operator-controller-manager-868647ff47-x74c8" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.396034 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4lmg\" (UniqueName: \"kubernetes.io/projected/f194b6f2-c143-4a57-8d3e-378a08147713-kube-api-access-f4lmg\") pod \"glance-operator-controller-manager-77987464f4-mfhjc\" (UID: \"f194b6f2-c143-4a57-8d3e-378a08147713\") " pod="openstack-operators/glance-operator-controller-manager-77987464f4-mfhjc" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.396369 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8z9fq\" (UniqueName: \"kubernetes.io/projected/463e8b44-a344-401e-bd53-22738560ca32-kube-api-access-8z9fq\") pod \"cinder-operator-controller-manager-5d946d989d-n4x7d\" (UID: \"463e8b44-a344-401e-bd53-22738560ca32\") " pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-n4x7d" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.412388 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-fb5fcc5b8-9mlxh"] Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.415122 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mdpg\" (UniqueName: \"kubernetes.io/projected/fff15cfd-9701-4efa-81a6-9a482e09ca0f-kube-api-access-8mdpg\") pod \"designate-operator-controller-manager-6d8bf5c495-5pv48\" (UID: \"fff15cfd-9701-4efa-81a6-9a482e09ca0f\") " pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-5pv48" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.416617 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6n7hw\" (UniqueName: \"kubernetes.io/projected/a980ea11-bf5f-4967-b22b-3e63454bb9ae-kube-api-access-6n7hw\") pod \"heat-operator-controller-manager-69f49c598c-6sck2\" (UID: \"a980ea11-bf5f-4967-b22b-3e63454bb9ae\") " pod="openstack-operators/heat-operator-controller-manager-69f49c598c-6sck2" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.454637 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-d44cf6b75-qr5np"] Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.454753 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-fb5fcc5b8-9mlxh" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.462529 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzskf\" (UniqueName: \"kubernetes.io/projected/9d670c80-f1d0-4b08-b83b-086977d49380-kube-api-access-hzskf\") pod \"neutron-operator-controller-manager-64ddbf8bb-wdp2t\" (UID: \"9d670c80-f1d0-4b08-b83b-086977d49380\") " pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-wdp2t" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.462570 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7rtx\" (UniqueName: \"kubernetes.io/projected/cd593415-1b63-4965-9865-21a1ae2e4742-kube-api-access-q7rtx\") pod \"keystone-operator-controller-manager-b4d948c87-9rkqc\" (UID: \"cd593415-1b63-4965-9865-21a1ae2e4742\") " pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-9rkqc" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.462596 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2p42z\" (UniqueName: \"kubernetes.io/projected/6e7470ce-ab97-4534-84e8-e951071eb6ee-kube-api-access-2p42z\") pod \"manila-operator-controller-manager-54f6768c69-r5kw5\" (UID: \"6e7470ce-ab97-4534-84e8-e951071eb6ee\") " pod="openstack-operators/manila-operator-controller-manager-54f6768c69-r5kw5" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.462613 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5s4kd\" (UniqueName: \"kubernetes.io/projected/fdd8c829-9082-40ae-8b65-46a8bd293c4d-kube-api-access-5s4kd\") pod \"mariadb-operator-controller-manager-6994f66f48-t2ng7\" (UID: \"fdd8c829-9082-40ae-8b65-46a8bd293c4d\") " pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-t2ng7" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.462632 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/853bc90c-e14b-46ea-912d-5aaefa5f908a-cert\") pod \"infra-operator-controller-manager-79d975b745-gp4rt\" (UID: \"853bc90c-e14b-46ea-912d-5aaefa5f908a\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-gp4rt" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.462653 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddvzc\" (UniqueName: \"kubernetes.io/projected/4fdde554-179d-4128-a260-38c3de6e9d64-kube-api-access-ddvzc\") pod \"ironic-operator-controller-manager-554564d7fc-lktrc\" (UID: \"4fdde554-179d-4128-a260-38c3de6e9d64\") " pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-lktrc" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.462681 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdzj6\" (UniqueName: \"kubernetes.io/projected/853bc90c-e14b-46ea-912d-5aaefa5f908a-kube-api-access-sdzj6\") pod \"infra-operator-controller-manager-79d975b745-gp4rt\" (UID: \"853bc90c-e14b-46ea-912d-5aaefa5f908a\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-gp4rt" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.462697 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzkd4\" (UniqueName: \"kubernetes.io/projected/d672d582-25f7-4654-bc52-5f07631f5e9e-kube-api-access-bzkd4\") pod \"nova-operator-controller-manager-567668f5cf-t9gtc\" (UID: \"d672d582-25f7-4654-bc52-5f07631f5e9e\") " pod="openstack-operators/nova-operator-controller-manager-567668f5cf-t9gtc" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.462742 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgpvw\" (UniqueName: \"kubernetes.io/projected/6a219296-13f5-4c55-99ec-35e34d43a341-kube-api-access-kgpvw\") pod \"horizon-operator-controller-manager-5b9b8895d5-g6hvq\" (UID: \"6a219296-13f5-4c55-99ec-35e34d43a341\") " pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-g6hvq" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.463329 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-x9k29" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.463464 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.480075 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-fb5fcc5b8-9mlxh"] Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.493740 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-8497b45c89-5vgnl"] Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.494431 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-5vgnl" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.495606 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-x74c8" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.496275 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-kvddt" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.498918 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-8497b45c89-5vgnl"] Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.505369 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-n4x7d" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.516504 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-5pv48" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.525552 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-68f46476f-jz577"] Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.526380 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-68f46476f-jz577" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.531346 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-9zgpj" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.531790 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987464f4-mfhjc" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.545227 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-68f46476f-jz577"] Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.553906 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-6sck2" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.563451 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcw8t\" (UniqueName: \"kubernetes.io/projected/2c577857-1c7f-4d16-9c13-d3d83eddb3de-kube-api-access-hcw8t\") pod \"ovn-operator-controller-manager-d44cf6b75-qr5np\" (UID: \"2c577857-1c7f-4d16-9c13-d3d83eddb3de\") " pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-qr5np" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.563496 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgpvw\" (UniqueName: \"kubernetes.io/projected/6a219296-13f5-4c55-99ec-35e34d43a341-kube-api-access-kgpvw\") pod \"horizon-operator-controller-manager-5b9b8895d5-g6hvq\" (UID: \"6a219296-13f5-4c55-99ec-35e34d43a341\") " pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-g6hvq" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.563520 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cn82x\" (UniqueName: \"kubernetes.io/projected/f2bd3ad9-919a-40ec-b111-f9655a6650a6-kube-api-access-cn82x\") pod \"placement-operator-controller-manager-8497b45c89-5vgnl\" (UID: \"f2bd3ad9-919a-40ec-b111-f9655a6650a6\") " pod="openstack-operators/placement-operator-controller-manager-8497b45c89-5vgnl" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.563542 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n44nr\" (UniqueName: \"kubernetes.io/projected/26bd53bd-69fd-4b14-afd5-5ffb7e557346-kube-api-access-n44nr\") pod \"swift-operator-controller-manager-68f46476f-jz577\" (UID: \"26bd53bd-69fd-4b14-afd5-5ffb7e557346\") " pod="openstack-operators/swift-operator-controller-manager-68f46476f-jz577" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.563558 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/48ca8d44-b660-43ee-8a9e-6489da3cee3d-cert\") pod \"openstack-baremetal-operator-controller-manager-fb5fcc5b8-9mlxh\" (UID: \"48ca8d44-b660-43ee-8a9e-6489da3cee3d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-fb5fcc5b8-9mlxh" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.563575 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzskf\" (UniqueName: \"kubernetes.io/projected/9d670c80-f1d0-4b08-b83b-086977d49380-kube-api-access-hzskf\") pod \"neutron-operator-controller-manager-64ddbf8bb-wdp2t\" (UID: \"9d670c80-f1d0-4b08-b83b-086977d49380\") " pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-wdp2t" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.563595 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccnqp\" (UniqueName: \"kubernetes.io/projected/96be73f9-04a8-4ec6-9b75-ae0dd4ae84c8-kube-api-access-ccnqp\") pod \"octavia-operator-controller-manager-69f8888797-rpk9w\" (UID: \"96be73f9-04a8-4ec6-9b75-ae0dd4ae84c8\") " pod="openstack-operators/octavia-operator-controller-manager-69f8888797-rpk9w" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.563610 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7rtx\" (UniqueName: \"kubernetes.io/projected/cd593415-1b63-4965-9865-21a1ae2e4742-kube-api-access-q7rtx\") pod \"keystone-operator-controller-manager-b4d948c87-9rkqc\" (UID: \"cd593415-1b63-4965-9865-21a1ae2e4742\") " pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-9rkqc" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.563630 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2p42z\" (UniqueName: \"kubernetes.io/projected/6e7470ce-ab97-4534-84e8-e951071eb6ee-kube-api-access-2p42z\") pod \"manila-operator-controller-manager-54f6768c69-r5kw5\" (UID: \"6e7470ce-ab97-4534-84e8-e951071eb6ee\") " pod="openstack-operators/manila-operator-controller-manager-54f6768c69-r5kw5" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.563645 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5s4kd\" (UniqueName: \"kubernetes.io/projected/fdd8c829-9082-40ae-8b65-46a8bd293c4d-kube-api-access-5s4kd\") pod \"mariadb-operator-controller-manager-6994f66f48-t2ng7\" (UID: \"fdd8c829-9082-40ae-8b65-46a8bd293c4d\") " pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-t2ng7" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.563663 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mb9qm\" (UniqueName: \"kubernetes.io/projected/48ca8d44-b660-43ee-8a9e-6489da3cee3d-kube-api-access-mb9qm\") pod \"openstack-baremetal-operator-controller-manager-fb5fcc5b8-9mlxh\" (UID: \"48ca8d44-b660-43ee-8a9e-6489da3cee3d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-fb5fcc5b8-9mlxh" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.563679 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/853bc90c-e14b-46ea-912d-5aaefa5f908a-cert\") pod \"infra-operator-controller-manager-79d975b745-gp4rt\" (UID: \"853bc90c-e14b-46ea-912d-5aaefa5f908a\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-gp4rt" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.563697 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ddvzc\" (UniqueName: \"kubernetes.io/projected/4fdde554-179d-4128-a260-38c3de6e9d64-kube-api-access-ddvzc\") pod \"ironic-operator-controller-manager-554564d7fc-lktrc\" (UID: \"4fdde554-179d-4128-a260-38c3de6e9d64\") " pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-lktrc" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.563720 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdzj6\" (UniqueName: \"kubernetes.io/projected/853bc90c-e14b-46ea-912d-5aaefa5f908a-kube-api-access-sdzj6\") pod \"infra-operator-controller-manager-79d975b745-gp4rt\" (UID: \"853bc90c-e14b-46ea-912d-5aaefa5f908a\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-gp4rt" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.563735 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bzkd4\" (UniqueName: \"kubernetes.io/projected/d672d582-25f7-4654-bc52-5f07631f5e9e-kube-api-access-bzkd4\") pod \"nova-operator-controller-manager-567668f5cf-t9gtc\" (UID: \"d672d582-25f7-4654-bc52-5f07631f5e9e\") " pod="openstack-operators/nova-operator-controller-manager-567668f5cf-t9gtc" Feb 18 09:15:20 crc kubenswrapper[4556]: E0218 09:15:20.564336 4556 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 18 09:15:20 crc kubenswrapper[4556]: E0218 09:15:20.564379 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/853bc90c-e14b-46ea-912d-5aaefa5f908a-cert podName:853bc90c-e14b-46ea-912d-5aaefa5f908a nodeName:}" failed. No retries permitted until 2026-02-18 09:15:21.064364628 +0000 UTC m=+678.081325607 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/853bc90c-e14b-46ea-912d-5aaefa5f908a-cert") pod "infra-operator-controller-manager-79d975b745-gp4rt" (UID: "853bc90c-e14b-46ea-912d-5aaefa5f908a") : secret "infra-operator-webhook-server-cert" not found Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.586485 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdzj6\" (UniqueName: \"kubernetes.io/projected/853bc90c-e14b-46ea-912d-5aaefa5f908a-kube-api-access-sdzj6\") pod \"infra-operator-controller-manager-79d975b745-gp4rt\" (UID: \"853bc90c-e14b-46ea-912d-5aaefa5f908a\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-gp4rt" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.586486 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzkd4\" (UniqueName: \"kubernetes.io/projected/d672d582-25f7-4654-bc52-5f07631f5e9e-kube-api-access-bzkd4\") pod \"nova-operator-controller-manager-567668f5cf-t9gtc\" (UID: \"d672d582-25f7-4654-bc52-5f07631f5e9e\") " pod="openstack-operators/nova-operator-controller-manager-567668f5cf-t9gtc" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.587819 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2p42z\" (UniqueName: \"kubernetes.io/projected/6e7470ce-ab97-4534-84e8-e951071eb6ee-kube-api-access-2p42z\") pod \"manila-operator-controller-manager-54f6768c69-r5kw5\" (UID: \"6e7470ce-ab97-4534-84e8-e951071eb6ee\") " pod="openstack-operators/manila-operator-controller-manager-54f6768c69-r5kw5" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.590097 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5s4kd\" (UniqueName: \"kubernetes.io/projected/fdd8c829-9082-40ae-8b65-46a8bd293c4d-kube-api-access-5s4kd\") pod \"mariadb-operator-controller-manager-6994f66f48-t2ng7\" (UID: \"fdd8c829-9082-40ae-8b65-46a8bd293c4d\") " pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-t2ng7" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.590842 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7rtx\" (UniqueName: \"kubernetes.io/projected/cd593415-1b63-4965-9865-21a1ae2e4742-kube-api-access-q7rtx\") pod \"keystone-operator-controller-manager-b4d948c87-9rkqc\" (UID: \"cd593415-1b63-4965-9865-21a1ae2e4742\") " pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-9rkqc" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.590857 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgpvw\" (UniqueName: \"kubernetes.io/projected/6a219296-13f5-4c55-99ec-35e34d43a341-kube-api-access-kgpvw\") pod \"horizon-operator-controller-manager-5b9b8895d5-g6hvq\" (UID: \"6a219296-13f5-4c55-99ec-35e34d43a341\") " pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-g6hvq" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.597600 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzskf\" (UniqueName: \"kubernetes.io/projected/9d670c80-f1d0-4b08-b83b-086977d49380-kube-api-access-hzskf\") pod \"neutron-operator-controller-manager-64ddbf8bb-wdp2t\" (UID: \"9d670c80-f1d0-4b08-b83b-086977d49380\") " pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-wdp2t" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.598283 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-4f5nl"] Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.599114 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-4f5nl" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.601661 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-r8hjj" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.617024 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ddvzc\" (UniqueName: \"kubernetes.io/projected/4fdde554-179d-4128-a260-38c3de6e9d64-kube-api-access-ddvzc\") pod \"ironic-operator-controller-manager-554564d7fc-lktrc\" (UID: \"4fdde554-179d-4128-a260-38c3de6e9d64\") " pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-lktrc" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.617799 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-9rkqc" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.619188 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-4f5nl"] Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.629493 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-r5kw5" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.637087 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-t2ng7" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.651601 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-wdp2t" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.660595 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-t9gtc" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.665073 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcw8t\" (UniqueName: \"kubernetes.io/projected/2c577857-1c7f-4d16-9c13-d3d83eddb3de-kube-api-access-hcw8t\") pod \"ovn-operator-controller-manager-d44cf6b75-qr5np\" (UID: \"2c577857-1c7f-4d16-9c13-d3d83eddb3de\") " pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-qr5np" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.665104 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cn82x\" (UniqueName: \"kubernetes.io/projected/f2bd3ad9-919a-40ec-b111-f9655a6650a6-kube-api-access-cn82x\") pod \"placement-operator-controller-manager-8497b45c89-5vgnl\" (UID: \"f2bd3ad9-919a-40ec-b111-f9655a6650a6\") " pod="openstack-operators/placement-operator-controller-manager-8497b45c89-5vgnl" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.665133 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n44nr\" (UniqueName: \"kubernetes.io/projected/26bd53bd-69fd-4b14-afd5-5ffb7e557346-kube-api-access-n44nr\") pod \"swift-operator-controller-manager-68f46476f-jz577\" (UID: \"26bd53bd-69fd-4b14-afd5-5ffb7e557346\") " pod="openstack-operators/swift-operator-controller-manager-68f46476f-jz577" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.665167 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/48ca8d44-b660-43ee-8a9e-6489da3cee3d-cert\") pod \"openstack-baremetal-operator-controller-manager-fb5fcc5b8-9mlxh\" (UID: \"48ca8d44-b660-43ee-8a9e-6489da3cee3d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-fb5fcc5b8-9mlxh" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.665196 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccnqp\" (UniqueName: \"kubernetes.io/projected/96be73f9-04a8-4ec6-9b75-ae0dd4ae84c8-kube-api-access-ccnqp\") pod \"octavia-operator-controller-manager-69f8888797-rpk9w\" (UID: \"96be73f9-04a8-4ec6-9b75-ae0dd4ae84c8\") " pod="openstack-operators/octavia-operator-controller-manager-69f8888797-rpk9w" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.665273 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mb9qm\" (UniqueName: \"kubernetes.io/projected/48ca8d44-b660-43ee-8a9e-6489da3cee3d-kube-api-access-mb9qm\") pod \"openstack-baremetal-operator-controller-manager-fb5fcc5b8-9mlxh\" (UID: \"48ca8d44-b660-43ee-8a9e-6489da3cee3d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-fb5fcc5b8-9mlxh" Feb 18 09:15:20 crc kubenswrapper[4556]: E0218 09:15:20.667433 4556 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 18 09:15:20 crc kubenswrapper[4556]: E0218 09:15:20.667500 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/48ca8d44-b660-43ee-8a9e-6489da3cee3d-cert podName:48ca8d44-b660-43ee-8a9e-6489da3cee3d nodeName:}" failed. No retries permitted until 2026-02-18 09:15:21.167462746 +0000 UTC m=+678.184423726 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/48ca8d44-b660-43ee-8a9e-6489da3cee3d-cert") pod "openstack-baremetal-operator-controller-manager-fb5fcc5b8-9mlxh" (UID: "48ca8d44-b660-43ee-8a9e-6489da3cee3d") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.687805 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcw8t\" (UniqueName: \"kubernetes.io/projected/2c577857-1c7f-4d16-9c13-d3d83eddb3de-kube-api-access-hcw8t\") pod \"ovn-operator-controller-manager-d44cf6b75-qr5np\" (UID: \"2c577857-1c7f-4d16-9c13-d3d83eddb3de\") " pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-qr5np" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.687854 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-7866795846-9bsld"] Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.689202 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-7866795846-9bsld" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.692953 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-kx6f4" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.694601 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n44nr\" (UniqueName: \"kubernetes.io/projected/26bd53bd-69fd-4b14-afd5-5ffb7e557346-kube-api-access-n44nr\") pod \"swift-operator-controller-manager-68f46476f-jz577\" (UID: \"26bd53bd-69fd-4b14-afd5-5ffb7e557346\") " pod="openstack-operators/swift-operator-controller-manager-68f46476f-jz577" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.694773 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-7866795846-9bsld"] Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.694820 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mb9qm\" (UniqueName: \"kubernetes.io/projected/48ca8d44-b660-43ee-8a9e-6489da3cee3d-kube-api-access-mb9qm\") pod \"openstack-baremetal-operator-controller-manager-fb5fcc5b8-9mlxh\" (UID: \"48ca8d44-b660-43ee-8a9e-6489da3cee3d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-fb5fcc5b8-9mlxh" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.696392 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cn82x\" (UniqueName: \"kubernetes.io/projected/f2bd3ad9-919a-40ec-b111-f9655a6650a6-kube-api-access-cn82x\") pod \"placement-operator-controller-manager-8497b45c89-5vgnl\" (UID: \"f2bd3ad9-919a-40ec-b111-f9655a6650a6\") " pod="openstack-operators/placement-operator-controller-manager-8497b45c89-5vgnl" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.696587 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccnqp\" (UniqueName: \"kubernetes.io/projected/96be73f9-04a8-4ec6-9b75-ae0dd4ae84c8-kube-api-access-ccnqp\") pod \"octavia-operator-controller-manager-69f8888797-rpk9w\" (UID: \"96be73f9-04a8-4ec6-9b75-ae0dd4ae84c8\") " pod="openstack-operators/octavia-operator-controller-manager-69f8888797-rpk9w" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.768369 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vc9p6\" (UniqueName: \"kubernetes.io/projected/acad90fd-fc19-4b74-bb00-192d0e6061a6-kube-api-access-vc9p6\") pod \"telemetry-operator-controller-manager-7f45b4ff68-4f5nl\" (UID: \"acad90fd-fc19-4b74-bb00-192d0e6061a6\") " pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-4f5nl" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.785087 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-5db88f68c-98pqn"] Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.788737 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-rpk9w" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.791858 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-98pqn" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.793911 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-t2rjg" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.800922 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-5db88f68c-98pqn"] Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.830774 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-5vgnl" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.843328 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-68f46476f-jz577" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.861123 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-g6hvq" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.869526 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vc9p6\" (UniqueName: \"kubernetes.io/projected/acad90fd-fc19-4b74-bb00-192d0e6061a6-kube-api-access-vc9p6\") pod \"telemetry-operator-controller-manager-7f45b4ff68-4f5nl\" (UID: \"acad90fd-fc19-4b74-bb00-192d0e6061a6\") " pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-4f5nl" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.869593 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j77tn\" (UniqueName: \"kubernetes.io/projected/bc370f09-39b1-4e31-bbc6-7756c63f0c30-kube-api-access-j77tn\") pod \"test-operator-controller-manager-7866795846-9bsld\" (UID: \"bc370f09-39b1-4e31-bbc6-7756c63f0c30\") " pod="openstack-operators/test-operator-controller-manager-7866795846-9bsld" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.880825 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-69ff7bc449-jjnjd"] Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.881933 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-69ff7bc449-jjnjd" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.884677 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-877nh" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.884695 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.884810 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.888788 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vc9p6\" (UniqueName: \"kubernetes.io/projected/acad90fd-fc19-4b74-bb00-192d0e6061a6-kube-api-access-vc9p6\") pod \"telemetry-operator-controller-manager-7f45b4ff68-4f5nl\" (UID: \"acad90fd-fc19-4b74-bb00-192d0e6061a6\") " pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-4f5nl" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.898331 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-lktrc" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.903544 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-69ff7bc449-jjnjd"] Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.925947 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-4f5nl" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.944558 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-868647ff47-x74c8"] Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.970257 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hw6mn\" (UniqueName: \"kubernetes.io/projected/8e02c422-7a17-4d4f-9ac1-7fe86c5dd472-kube-api-access-hw6mn\") pod \"watcher-operator-controller-manager-5db88f68c-98pqn\" (UID: \"8e02c422-7a17-4d4f-9ac1-7fe86c5dd472\") " pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-98pqn" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.971406 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j77tn\" (UniqueName: \"kubernetes.io/projected/bc370f09-39b1-4e31-bbc6-7756c63f0c30-kube-api-access-j77tn\") pod \"test-operator-controller-manager-7866795846-9bsld\" (UID: \"bc370f09-39b1-4e31-bbc6-7756c63f0c30\") " pod="openstack-operators/test-operator-controller-manager-7866795846-9bsld" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.972911 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-vt6jv"] Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.975031 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-vt6jv" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.980053 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-6dl8n" Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.983403 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-vt6jv"] Feb 18 09:15:20 crc kubenswrapper[4556]: I0218 09:15:20.990729 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j77tn\" (UniqueName: \"kubernetes.io/projected/bc370f09-39b1-4e31-bbc6-7756c63f0c30-kube-api-access-j77tn\") pod \"test-operator-controller-manager-7866795846-9bsld\" (UID: \"bc370f09-39b1-4e31-bbc6-7756c63f0c30\") " pod="openstack-operators/test-operator-controller-manager-7866795846-9bsld" Feb 18 09:15:21 crc kubenswrapper[4556]: I0218 09:15:21.013396 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-7866795846-9bsld" Feb 18 09:15:21 crc kubenswrapper[4556]: I0218 09:15:21.072493 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4fb1fa58-065b-4412-99e1-e9d77cdf4b41-metrics-certs\") pod \"openstack-operator-controller-manager-69ff7bc449-jjnjd\" (UID: \"4fb1fa58-065b-4412-99e1-e9d77cdf4b41\") " pod="openstack-operators/openstack-operator-controller-manager-69ff7bc449-jjnjd" Feb 18 09:15:21 crc kubenswrapper[4556]: I0218 09:15:21.072570 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4fb1fa58-065b-4412-99e1-e9d77cdf4b41-webhook-certs\") pod \"openstack-operator-controller-manager-69ff7bc449-jjnjd\" (UID: \"4fb1fa58-065b-4412-99e1-e9d77cdf4b41\") " pod="openstack-operators/openstack-operator-controller-manager-69ff7bc449-jjnjd" Feb 18 09:15:21 crc kubenswrapper[4556]: I0218 09:15:21.072596 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hw6mn\" (UniqueName: \"kubernetes.io/projected/8e02c422-7a17-4d4f-9ac1-7fe86c5dd472-kube-api-access-hw6mn\") pod \"watcher-operator-controller-manager-5db88f68c-98pqn\" (UID: \"8e02c422-7a17-4d4f-9ac1-7fe86c5dd472\") " pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-98pqn" Feb 18 09:15:21 crc kubenswrapper[4556]: I0218 09:15:21.072700 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/853bc90c-e14b-46ea-912d-5aaefa5f908a-cert\") pod \"infra-operator-controller-manager-79d975b745-gp4rt\" (UID: \"853bc90c-e14b-46ea-912d-5aaefa5f908a\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-gp4rt" Feb 18 09:15:21 crc kubenswrapper[4556]: I0218 09:15:21.072720 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xnsp\" (UniqueName: \"kubernetes.io/projected/4fb1fa58-065b-4412-99e1-e9d77cdf4b41-kube-api-access-6xnsp\") pod \"openstack-operator-controller-manager-69ff7bc449-jjnjd\" (UID: \"4fb1fa58-065b-4412-99e1-e9d77cdf4b41\") " pod="openstack-operators/openstack-operator-controller-manager-69ff7bc449-jjnjd" Feb 18 09:15:21 crc kubenswrapper[4556]: E0218 09:15:21.073516 4556 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 18 09:15:21 crc kubenswrapper[4556]: E0218 09:15:21.073575 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/853bc90c-e14b-46ea-912d-5aaefa5f908a-cert podName:853bc90c-e14b-46ea-912d-5aaefa5f908a nodeName:}" failed. No retries permitted until 2026-02-18 09:15:22.073563277 +0000 UTC m=+679.090524257 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/853bc90c-e14b-46ea-912d-5aaefa5f908a-cert") pod "infra-operator-controller-manager-79d975b745-gp4rt" (UID: "853bc90c-e14b-46ea-912d-5aaefa5f908a") : secret "infra-operator-webhook-server-cert" not found Feb 18 09:15:21 crc kubenswrapper[4556]: I0218 09:15:21.076643 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987464f4-mfhjc"] Feb 18 09:15:21 crc kubenswrapper[4556]: I0218 09:15:21.093328 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hw6mn\" (UniqueName: \"kubernetes.io/projected/8e02c422-7a17-4d4f-9ac1-7fe86c5dd472-kube-api-access-hw6mn\") pod \"watcher-operator-controller-manager-5db88f68c-98pqn\" (UID: \"8e02c422-7a17-4d4f-9ac1-7fe86c5dd472\") " pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-98pqn" Feb 18 09:15:21 crc kubenswrapper[4556]: I0218 09:15:21.094441 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-6d8bf5c495-5pv48"] Feb 18 09:15:21 crc kubenswrapper[4556]: I0218 09:15:21.102794 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-5d946d989d-n4x7d"] Feb 18 09:15:21 crc kubenswrapper[4556]: I0218 09:15:21.138594 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-98pqn" Feb 18 09:15:21 crc kubenswrapper[4556]: I0218 09:15:21.173835 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xnsp\" (UniqueName: \"kubernetes.io/projected/4fb1fa58-065b-4412-99e1-e9d77cdf4b41-kube-api-access-6xnsp\") pod \"openstack-operator-controller-manager-69ff7bc449-jjnjd\" (UID: \"4fb1fa58-065b-4412-99e1-e9d77cdf4b41\") " pod="openstack-operators/openstack-operator-controller-manager-69ff7bc449-jjnjd" Feb 18 09:15:21 crc kubenswrapper[4556]: I0218 09:15:21.173896 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghrkj\" (UniqueName: \"kubernetes.io/projected/28452a75-3b84-4add-9952-c5f0193e0954-kube-api-access-ghrkj\") pod \"rabbitmq-cluster-operator-manager-668c99d594-vt6jv\" (UID: \"28452a75-3b84-4add-9952-c5f0193e0954\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-vt6jv" Feb 18 09:15:21 crc kubenswrapper[4556]: I0218 09:15:21.173927 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4fb1fa58-065b-4412-99e1-e9d77cdf4b41-metrics-certs\") pod \"openstack-operator-controller-manager-69ff7bc449-jjnjd\" (UID: \"4fb1fa58-065b-4412-99e1-e9d77cdf4b41\") " pod="openstack-operators/openstack-operator-controller-manager-69ff7bc449-jjnjd" Feb 18 09:15:21 crc kubenswrapper[4556]: I0218 09:15:21.173962 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4fb1fa58-065b-4412-99e1-e9d77cdf4b41-webhook-certs\") pod \"openstack-operator-controller-manager-69ff7bc449-jjnjd\" (UID: \"4fb1fa58-065b-4412-99e1-e9d77cdf4b41\") " pod="openstack-operators/openstack-operator-controller-manager-69ff7bc449-jjnjd" Feb 18 09:15:21 crc kubenswrapper[4556]: I0218 09:15:21.173996 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/48ca8d44-b660-43ee-8a9e-6489da3cee3d-cert\") pod \"openstack-baremetal-operator-controller-manager-fb5fcc5b8-9mlxh\" (UID: \"48ca8d44-b660-43ee-8a9e-6489da3cee3d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-fb5fcc5b8-9mlxh" Feb 18 09:15:21 crc kubenswrapper[4556]: E0218 09:15:21.174121 4556 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 18 09:15:21 crc kubenswrapper[4556]: E0218 09:15:21.174173 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/48ca8d44-b660-43ee-8a9e-6489da3cee3d-cert podName:48ca8d44-b660-43ee-8a9e-6489da3cee3d nodeName:}" failed. No retries permitted until 2026-02-18 09:15:22.174146209 +0000 UTC m=+679.191107189 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/48ca8d44-b660-43ee-8a9e-6489da3cee3d-cert") pod "openstack-baremetal-operator-controller-manager-fb5fcc5b8-9mlxh" (UID: "48ca8d44-b660-43ee-8a9e-6489da3cee3d") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 18 09:15:21 crc kubenswrapper[4556]: E0218 09:15:21.174231 4556 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 18 09:15:21 crc kubenswrapper[4556]: E0218 09:15:21.174278 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4fb1fa58-065b-4412-99e1-e9d77cdf4b41-metrics-certs podName:4fb1fa58-065b-4412-99e1-e9d77cdf4b41 nodeName:}" failed. No retries permitted until 2026-02-18 09:15:21.67426364 +0000 UTC m=+678.691224620 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4fb1fa58-065b-4412-99e1-e9d77cdf4b41-metrics-certs") pod "openstack-operator-controller-manager-69ff7bc449-jjnjd" (UID: "4fb1fa58-065b-4412-99e1-e9d77cdf4b41") : secret "metrics-server-cert" not found Feb 18 09:15:21 crc kubenswrapper[4556]: E0218 09:15:21.174328 4556 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 18 09:15:21 crc kubenswrapper[4556]: E0218 09:15:21.174346 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4fb1fa58-065b-4412-99e1-e9d77cdf4b41-webhook-certs podName:4fb1fa58-065b-4412-99e1-e9d77cdf4b41 nodeName:}" failed. No retries permitted until 2026-02-18 09:15:21.674340485 +0000 UTC m=+678.691301465 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/4fb1fa58-065b-4412-99e1-e9d77cdf4b41-webhook-certs") pod "openstack-operator-controller-manager-69ff7bc449-jjnjd" (UID: "4fb1fa58-065b-4412-99e1-e9d77cdf4b41") : secret "webhook-server-cert" not found Feb 18 09:15:21 crc kubenswrapper[4556]: I0218 09:15:21.190953 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xnsp\" (UniqueName: \"kubernetes.io/projected/4fb1fa58-065b-4412-99e1-e9d77cdf4b41-kube-api-access-6xnsp\") pod \"openstack-operator-controller-manager-69ff7bc449-jjnjd\" (UID: \"4fb1fa58-065b-4412-99e1-e9d77cdf4b41\") " pod="openstack-operators/openstack-operator-controller-manager-69ff7bc449-jjnjd" Feb 18 09:15:21 crc kubenswrapper[4556]: I0218 09:15:21.235662 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-54f6768c69-r5kw5"] Feb 18 09:15:21 crc kubenswrapper[4556]: W0218 09:15:21.243703 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6e7470ce_ab97_4534_84e8_e951071eb6ee.slice/crio-b77873692fec9aea7ea1f465d00e7239519ed0953fe19909e3e3270737521533 WatchSource:0}: Error finding container b77873692fec9aea7ea1f465d00e7239519ed0953fe19909e3e3270737521533: Status 404 returned error can't find the container with id b77873692fec9aea7ea1f465d00e7239519ed0953fe19909e3e3270737521533 Feb 18 09:15:21 crc kubenswrapper[4556]: I0218 09:15:21.275394 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghrkj\" (UniqueName: \"kubernetes.io/projected/28452a75-3b84-4add-9952-c5f0193e0954-kube-api-access-ghrkj\") pod \"rabbitmq-cluster-operator-manager-668c99d594-vt6jv\" (UID: \"28452a75-3b84-4add-9952-c5f0193e0954\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-vt6jv" Feb 18 09:15:21 crc kubenswrapper[4556]: I0218 09:15:21.288276 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghrkj\" (UniqueName: \"kubernetes.io/projected/28452a75-3b84-4add-9952-c5f0193e0954-kube-api-access-ghrkj\") pod \"rabbitmq-cluster-operator-manager-668c99d594-vt6jv\" (UID: \"28452a75-3b84-4add-9952-c5f0193e0954\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-vt6jv" Feb 18 09:15:21 crc kubenswrapper[4556]: I0218 09:15:21.304401 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-vt6jv" Feb 18 09:15:21 crc kubenswrapper[4556]: I0218 09:15:21.515597 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69f8888797-rpk9w"] Feb 18 09:15:21 crc kubenswrapper[4556]: I0218 09:15:21.519796 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-b4d948c87-9rkqc"] Feb 18 09:15:21 crc kubenswrapper[4556]: I0218 09:15:21.530265 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-69f49c598c-6sck2"] Feb 18 09:15:21 crc kubenswrapper[4556]: I0218 09:15:21.540750 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-64ddbf8bb-wdp2t"] Feb 18 09:15:21 crc kubenswrapper[4556]: W0218 09:15:21.547508 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d670c80_f1d0_4b08_b83b_086977d49380.slice/crio-37bd3ddec1376fbec7f075c48b3f010b50fede386a8454c871752cf80a64f92b WatchSource:0}: Error finding container 37bd3ddec1376fbec7f075c48b3f010b50fede386a8454c871752cf80a64f92b: Status 404 returned error can't find the container with id 37bd3ddec1376fbec7f075c48b3f010b50fede386a8454c871752cf80a64f92b Feb 18 09:15:21 crc kubenswrapper[4556]: I0218 09:15:21.549367 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6994f66f48-t2ng7"] Feb 18 09:15:21 crc kubenswrapper[4556]: I0218 09:15:21.554973 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-567668f5cf-t9gtc"] Feb 18 09:15:21 crc kubenswrapper[4556]: W0218 09:15:21.559415 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd672d582_25f7_4654_bc52_5f07631f5e9e.slice/crio-e4105bbd90f149c034808712a56c76849be2807213c15c695609dd59adf67011 WatchSource:0}: Error finding container e4105bbd90f149c034808712a56c76849be2807213c15c695609dd59adf67011: Status 404 returned error can't find the container with id e4105bbd90f149c034808712a56c76849be2807213c15c695609dd59adf67011 Feb 18 09:15:21 crc kubenswrapper[4556]: W0218 09:15:21.559765 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfdd8c829_9082_40ae_8b65_46a8bd293c4d.slice/crio-5c6a95a5446bea28eabb79c721116167551c9a01cfe69a94dcf0d5f8e7ccfc24 WatchSource:0}: Error finding container 5c6a95a5446bea28eabb79c721116167551c9a01cfe69a94dcf0d5f8e7ccfc24: Status 404 returned error can't find the container with id 5c6a95a5446bea28eabb79c721116167551c9a01cfe69a94dcf0d5f8e7ccfc24 Feb 18 09:15:21 crc kubenswrapper[4556]: I0218 09:15:21.569871 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-4f5nl"] Feb 18 09:15:21 crc kubenswrapper[4556]: I0218 09:15:21.573949 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-vt6jv"] Feb 18 09:15:21 crc kubenswrapper[4556]: W0218 09:15:21.576523 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podacad90fd_fc19_4b74_bb00_192d0e6061a6.slice/crio-49a4c24aad97e20c607246ce8dc9cf410d10ddd64e7b38f5575670251d23c291 WatchSource:0}: Error finding container 49a4c24aad97e20c607246ce8dc9cf410d10ddd64e7b38f5575670251d23c291: Status 404 returned error can't find the container with id 49a4c24aad97e20c607246ce8dc9cf410d10ddd64e7b38f5575670251d23c291 Feb 18 09:15:21 crc kubenswrapper[4556]: I0218 09:15:21.577986 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-8497b45c89-5vgnl"] Feb 18 09:15:21 crc kubenswrapper[4556]: W0218 09:15:21.579208 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod28452a75_3b84_4add_9952_c5f0193e0954.slice/crio-46fff9d0868aad7b8eba71d7f79903f857cd4e36474981d5a8822a7edf7228fb WatchSource:0}: Error finding container 46fff9d0868aad7b8eba71d7f79903f857cd4e36474981d5a8822a7edf7228fb: Status 404 returned error can't find the container with id 46fff9d0868aad7b8eba71d7f79903f857cd4e36474981d5a8822a7edf7228fb Feb 18 09:15:21 crc kubenswrapper[4556]: W0218 09:15:21.582307 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf2bd3ad9_919a_40ec_b111_f9655a6650a6.slice/crio-aaf0d4f16e94cfcb2f1cbb2ad41d116ef31468082e2ec0e852d0f63bfc93dfd9 WatchSource:0}: Error finding container aaf0d4f16e94cfcb2f1cbb2ad41d116ef31468082e2ec0e852d0f63bfc93dfd9: Status 404 returned error can't find the container with id aaf0d4f16e94cfcb2f1cbb2ad41d116ef31468082e2ec0e852d0f63bfc93dfd9 Feb 18 09:15:21 crc kubenswrapper[4556]: E0218 09:15:21.588222 4556 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ghrkj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-vt6jv_openstack-operators(28452a75-3b84-4add-9952-c5f0193e0954): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 18 09:15:21 crc kubenswrapper[4556]: E0218 09:15:21.590001 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-vt6jv" podUID="28452a75-3b84-4add-9952-c5f0193e0954" Feb 18 09:15:21 crc kubenswrapper[4556]: I0218 09:15:21.621104 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-7866795846-9bsld"] Feb 18 09:15:21 crc kubenswrapper[4556]: I0218 09:15:21.627247 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-68f46476f-jz577"] Feb 18 09:15:21 crc kubenswrapper[4556]: E0218 09:15:21.628282 4556 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:3d676f1281e24ef07de617570d2f7fbf625032e41866d1551a856c052248bb04,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-n44nr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-68f46476f-jz577_openstack-operators(26bd53bd-69fd-4b14-afd5-5ffb7e557346): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 18 09:15:21 crc kubenswrapper[4556]: E0218 09:15:21.629771 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-68f46476f-jz577" podUID="26bd53bd-69fd-4b14-afd5-5ffb7e557346" Feb 18 09:15:21 crc kubenswrapper[4556]: I0218 09:15:21.634264 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-5db88f68c-98pqn"] Feb 18 09:15:21 crc kubenswrapper[4556]: E0218 09:15:21.636275 4556 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:f0fabdf79095def0f8b1c0442925548a94ca94bed4de2d3b171277129f8079e6,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-j77tn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-7866795846-9bsld_openstack-operators(bc370f09-39b1-4e31-bbc6-7756c63f0c30): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 18 09:15:21 crc kubenswrapper[4556]: E0218 09:15:21.637866 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-7866795846-9bsld" podUID="bc370f09-39b1-4e31-bbc6-7756c63f0c30" Feb 18 09:15:21 crc kubenswrapper[4556]: I0218 09:15:21.640271 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5b9b8895d5-g6hvq"] Feb 18 09:15:21 crc kubenswrapper[4556]: E0218 09:15:21.641662 4556 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:7e1b0b7b172ad0d707ab80dd72d609e1d0f5bbd38a22c24a28ed0f17a960c867,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ddvzc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-554564d7fc-lktrc_openstack-operators(4fdde554-179d-4128-a260-38c3de6e9d64): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 18 09:15:21 crc kubenswrapper[4556]: E0218 09:15:21.643262 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-lktrc" podUID="4fdde554-179d-4128-a260-38c3de6e9d64" Feb 18 09:15:21 crc kubenswrapper[4556]: I0218 09:15:21.643619 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-554564d7fc-lktrc"] Feb 18 09:15:21 crc kubenswrapper[4556]: E0218 09:15:21.647978 4556 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:9f2e1299d908411457e53b49e1062265d2b9d76f6719db24d1be9347c388e4da,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kgpvw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-5b9b8895d5-g6hvq_openstack-operators(6a219296-13f5-4c55-99ec-35e34d43a341): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 18 09:15:21 crc kubenswrapper[4556]: E0218 09:15:21.648095 4556 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:d01ae848290e880c09127d5297418dea40fc7f090fdab9bf2c578c7e7f53aec0,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hw6mn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-5db88f68c-98pqn_openstack-operators(8e02c422-7a17-4d4f-9ac1-7fe86c5dd472): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 18 09:15:21 crc kubenswrapper[4556]: E0218 09:15:21.649826 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-98pqn" podUID="8e02c422-7a17-4d4f-9ac1-7fe86c5dd472" Feb 18 09:15:21 crc kubenswrapper[4556]: E0218 09:15:21.649860 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-g6hvq" podUID="6a219296-13f5-4c55-99ec-35e34d43a341" Feb 18 09:15:21 crc kubenswrapper[4556]: I0218 09:15:21.672310 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-g6hvq" event={"ID":"6a219296-13f5-4c55-99ec-35e34d43a341","Type":"ContainerStarted","Data":"8bf7262c05ca8d0debfa6213c8bb4f74dc810c69dea911afa8ce5048790fe1a4"} Feb 18 09:15:21 crc kubenswrapper[4556]: E0218 09:15:21.673863 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/horizon-operator@sha256:9f2e1299d908411457e53b49e1062265d2b9d76f6719db24d1be9347c388e4da\\\"\"" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-g6hvq" podUID="6a219296-13f5-4c55-99ec-35e34d43a341" Feb 18 09:15:21 crc kubenswrapper[4556]: I0218 09:15:21.674135 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-6sck2" event={"ID":"a980ea11-bf5f-4967-b22b-3e63454bb9ae","Type":"ContainerStarted","Data":"3821e6cb4b12b5a9603ce98f64fd649da7c9f3804a8cdcccca72513ef757db55"} Feb 18 09:15:21 crc kubenswrapper[4556]: I0218 09:15:21.675666 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-68f46476f-jz577" event={"ID":"26bd53bd-69fd-4b14-afd5-5ffb7e557346","Type":"ContainerStarted","Data":"20e74a950f3d43c67cd2eeb6bb75cfcfbbec4faa5260e8368aa8b474ff31ca95"} Feb 18 09:15:21 crc kubenswrapper[4556]: E0218 09:15:21.676446 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:3d676f1281e24ef07de617570d2f7fbf625032e41866d1551a856c052248bb04\\\"\"" pod="openstack-operators/swift-operator-controller-manager-68f46476f-jz577" podUID="26bd53bd-69fd-4b14-afd5-5ffb7e557346" Feb 18 09:15:21 crc kubenswrapper[4556]: I0218 09:15:21.676684 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-7866795846-9bsld" event={"ID":"bc370f09-39b1-4e31-bbc6-7756c63f0c30","Type":"ContainerStarted","Data":"41d545d4dc4ea2ce44e567d4b1ef56d54a5d6a42013e30ce37eb9713ffe6b158"} Feb 18 09:15:21 crc kubenswrapper[4556]: E0218 09:15:21.677557 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:f0fabdf79095def0f8b1c0442925548a94ca94bed4de2d3b171277129f8079e6\\\"\"" pod="openstack-operators/test-operator-controller-manager-7866795846-9bsld" podUID="bc370f09-39b1-4e31-bbc6-7756c63f0c30" Feb 18 09:15:21 crc kubenswrapper[4556]: I0218 09:15:21.677731 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-4f5nl" event={"ID":"acad90fd-fc19-4b74-bb00-192d0e6061a6","Type":"ContainerStarted","Data":"49a4c24aad97e20c607246ce8dc9cf410d10ddd64e7b38f5575670251d23c291"} Feb 18 09:15:21 crc kubenswrapper[4556]: I0218 09:15:21.680343 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4fb1fa58-065b-4412-99e1-e9d77cdf4b41-metrics-certs\") pod \"openstack-operator-controller-manager-69ff7bc449-jjnjd\" (UID: \"4fb1fa58-065b-4412-99e1-e9d77cdf4b41\") " pod="openstack-operators/openstack-operator-controller-manager-69ff7bc449-jjnjd" Feb 18 09:15:21 crc kubenswrapper[4556]: I0218 09:15:21.680402 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4fb1fa58-065b-4412-99e1-e9d77cdf4b41-webhook-certs\") pod \"openstack-operator-controller-manager-69ff7bc449-jjnjd\" (UID: \"4fb1fa58-065b-4412-99e1-e9d77cdf4b41\") " pod="openstack-operators/openstack-operator-controller-manager-69ff7bc449-jjnjd" Feb 18 09:15:21 crc kubenswrapper[4556]: E0218 09:15:21.681091 4556 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 18 09:15:21 crc kubenswrapper[4556]: E0218 09:15:21.681115 4556 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 18 09:15:21 crc kubenswrapper[4556]: E0218 09:15:21.681144 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4fb1fa58-065b-4412-99e1-e9d77cdf4b41-metrics-certs podName:4fb1fa58-065b-4412-99e1-e9d77cdf4b41 nodeName:}" failed. No retries permitted until 2026-02-18 09:15:22.681131601 +0000 UTC m=+679.698092581 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4fb1fa58-065b-4412-99e1-e9d77cdf4b41-metrics-certs") pod "openstack-operator-controller-manager-69ff7bc449-jjnjd" (UID: "4fb1fa58-065b-4412-99e1-e9d77cdf4b41") : secret "metrics-server-cert" not found Feb 18 09:15:21 crc kubenswrapper[4556]: E0218 09:15:21.681179 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4fb1fa58-065b-4412-99e1-e9d77cdf4b41-webhook-certs podName:4fb1fa58-065b-4412-99e1-e9d77cdf4b41 nodeName:}" failed. No retries permitted until 2026-02-18 09:15:22.681165514 +0000 UTC m=+679.698126485 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/4fb1fa58-065b-4412-99e1-e9d77cdf4b41-webhook-certs") pod "openstack-operator-controller-manager-69ff7bc449-jjnjd" (UID: "4fb1fa58-065b-4412-99e1-e9d77cdf4b41") : secret "webhook-server-cert" not found Feb 18 09:15:21 crc kubenswrapper[4556]: I0218 09:15:21.681251 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987464f4-mfhjc" event={"ID":"f194b6f2-c143-4a57-8d3e-378a08147713","Type":"ContainerStarted","Data":"8e00bf70de61b0eeda690912dfbf8db07221860bf4afe9cf866355dbe181154b"} Feb 18 09:15:21 crc kubenswrapper[4556]: I0218 09:15:21.685007 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-wdp2t" event={"ID":"9d670c80-f1d0-4b08-b83b-086977d49380","Type":"ContainerStarted","Data":"37bd3ddec1376fbec7f075c48b3f010b50fede386a8454c871752cf80a64f92b"} Feb 18 09:15:21 crc kubenswrapper[4556]: I0218 09:15:21.685881 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-5pv48" event={"ID":"fff15cfd-9701-4efa-81a6-9a482e09ca0f","Type":"ContainerStarted","Data":"7c41ec22bab5f0d8b5b91d61531a18f0d9aeda7768eb33077421710ec5ae56b5"} Feb 18 09:15:21 crc kubenswrapper[4556]: I0218 09:15:21.687239 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-vt6jv" event={"ID":"28452a75-3b84-4add-9952-c5f0193e0954","Type":"ContainerStarted","Data":"46fff9d0868aad7b8eba71d7f79903f857cd4e36474981d5a8822a7edf7228fb"} Feb 18 09:15:21 crc kubenswrapper[4556]: E0218 09:15:21.688392 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-vt6jv" podUID="28452a75-3b84-4add-9952-c5f0193e0954" Feb 18 09:15:21 crc kubenswrapper[4556]: I0218 09:15:21.691640 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-98pqn" event={"ID":"8e02c422-7a17-4d4f-9ac1-7fe86c5dd472","Type":"ContainerStarted","Data":"8c6bf5715655af8365a1acbb8e76df86122207fef559c6b944290d4870b6f1f7"} Feb 18 09:15:21 crc kubenswrapper[4556]: E0218 09:15:21.692658 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:d01ae848290e880c09127d5297418dea40fc7f090fdab9bf2c578c7e7f53aec0\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-98pqn" podUID="8e02c422-7a17-4d4f-9ac1-7fe86c5dd472" Feb 18 09:15:21 crc kubenswrapper[4556]: I0218 09:15:21.695832 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-5vgnl" event={"ID":"f2bd3ad9-919a-40ec-b111-f9655a6650a6","Type":"ContainerStarted","Data":"aaf0d4f16e94cfcb2f1cbb2ad41d116ef31468082e2ec0e852d0f63bfc93dfd9"} Feb 18 09:15:21 crc kubenswrapper[4556]: I0218 09:15:21.697146 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-lktrc" event={"ID":"4fdde554-179d-4128-a260-38c3de6e9d64","Type":"ContainerStarted","Data":"45c40eb069e050762576a468279d675b2bc1b3d416c6d82754af2d579ceb08c4"} Feb 18 09:15:21 crc kubenswrapper[4556]: E0218 09:15:21.698080 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ironic-operator@sha256:7e1b0b7b172ad0d707ab80dd72d609e1d0f5bbd38a22c24a28ed0f17a960c867\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-lktrc" podUID="4fdde554-179d-4128-a260-38c3de6e9d64" Feb 18 09:15:21 crc kubenswrapper[4556]: I0218 09:15:21.698408 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-9rkqc" event={"ID":"cd593415-1b63-4965-9865-21a1ae2e4742","Type":"ContainerStarted","Data":"45ba4cf39cf43157a069fa39a7e4c76693a7c701869dba285201e9cdc1b41f27"} Feb 18 09:15:21 crc kubenswrapper[4556]: I0218 09:15:21.699238 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-x74c8" event={"ID":"9cf84ffa-7a16-4e6a-aa7a-c90d6c1635fe","Type":"ContainerStarted","Data":"b3671279e7c531eff861b5530560ba1fbd3555c8e97a48ca4004f44359deed48"} Feb 18 09:15:21 crc kubenswrapper[4556]: I0218 09:15:21.701345 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-t2ng7" event={"ID":"fdd8c829-9082-40ae-8b65-46a8bd293c4d","Type":"ContainerStarted","Data":"5c6a95a5446bea28eabb79c721116167551c9a01cfe69a94dcf0d5f8e7ccfc24"} Feb 18 09:15:21 crc kubenswrapper[4556]: I0218 09:15:21.702193 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-t9gtc" event={"ID":"d672d582-25f7-4654-bc52-5f07631f5e9e","Type":"ContainerStarted","Data":"e4105bbd90f149c034808712a56c76849be2807213c15c695609dd59adf67011"} Feb 18 09:15:21 crc kubenswrapper[4556]: I0218 09:15:21.703110 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-n4x7d" event={"ID":"463e8b44-a344-401e-bd53-22738560ca32","Type":"ContainerStarted","Data":"13ad9135288ba1120323f83c9906520ca693a702c2b36a8d3d8680f2fefc0496"} Feb 18 09:15:21 crc kubenswrapper[4556]: I0218 09:15:21.705403 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-rpk9w" event={"ID":"96be73f9-04a8-4ec6-9b75-ae0dd4ae84c8","Type":"ContainerStarted","Data":"8025b17a52dad3794fd15b5447ec22f890608352da0add83ff6dff256fce5899"} Feb 18 09:15:21 crc kubenswrapper[4556]: I0218 09:15:21.706574 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-r5kw5" event={"ID":"6e7470ce-ab97-4534-84e8-e951071eb6ee","Type":"ContainerStarted","Data":"b77873692fec9aea7ea1f465d00e7239519ed0953fe19909e3e3270737521533"} Feb 18 09:15:21 crc kubenswrapper[4556]: I0218 09:15:21.709851 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-dxsw4" Feb 18 09:15:21 crc kubenswrapper[4556]: I0218 09:15:21.717502 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-qr5np" Feb 18 09:15:21 crc kubenswrapper[4556]: I0218 09:15:21.860613 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-d44cf6b75-qr5np"] Feb 18 09:15:21 crc kubenswrapper[4556]: W0218 09:15:21.863060 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2c577857_1c7f_4d16_9c13_d3d83eddb3de.slice/crio-a89bafd56371d68fbd7b13cc659db1fb7256c101fa31f048c5f9affe0f84df5b WatchSource:0}: Error finding container a89bafd56371d68fbd7b13cc659db1fb7256c101fa31f048c5f9affe0f84df5b: Status 404 returned error can't find the container with id a89bafd56371d68fbd7b13cc659db1fb7256c101fa31f048c5f9affe0f84df5b Feb 18 09:15:22 crc kubenswrapper[4556]: I0218 09:15:22.086810 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/853bc90c-e14b-46ea-912d-5aaefa5f908a-cert\") pod \"infra-operator-controller-manager-79d975b745-gp4rt\" (UID: \"853bc90c-e14b-46ea-912d-5aaefa5f908a\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-gp4rt" Feb 18 09:15:22 crc kubenswrapper[4556]: E0218 09:15:22.087028 4556 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 18 09:15:22 crc kubenswrapper[4556]: E0218 09:15:22.087100 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/853bc90c-e14b-46ea-912d-5aaefa5f908a-cert podName:853bc90c-e14b-46ea-912d-5aaefa5f908a nodeName:}" failed. No retries permitted until 2026-02-18 09:15:24.087080565 +0000 UTC m=+681.104041545 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/853bc90c-e14b-46ea-912d-5aaefa5f908a-cert") pod "infra-operator-controller-manager-79d975b745-gp4rt" (UID: "853bc90c-e14b-46ea-912d-5aaefa5f908a") : secret "infra-operator-webhook-server-cert" not found Feb 18 09:15:22 crc kubenswrapper[4556]: I0218 09:15:22.187715 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/48ca8d44-b660-43ee-8a9e-6489da3cee3d-cert\") pod \"openstack-baremetal-operator-controller-manager-fb5fcc5b8-9mlxh\" (UID: \"48ca8d44-b660-43ee-8a9e-6489da3cee3d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-fb5fcc5b8-9mlxh" Feb 18 09:15:22 crc kubenswrapper[4556]: E0218 09:15:22.187968 4556 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 18 09:15:22 crc kubenswrapper[4556]: E0218 09:15:22.188007 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/48ca8d44-b660-43ee-8a9e-6489da3cee3d-cert podName:48ca8d44-b660-43ee-8a9e-6489da3cee3d nodeName:}" failed. No retries permitted until 2026-02-18 09:15:24.187995785 +0000 UTC m=+681.204956765 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/48ca8d44-b660-43ee-8a9e-6489da3cee3d-cert") pod "openstack-baremetal-operator-controller-manager-fb5fcc5b8-9mlxh" (UID: "48ca8d44-b660-43ee-8a9e-6489da3cee3d") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 18 09:15:22 crc kubenswrapper[4556]: I0218 09:15:22.700058 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4fb1fa58-065b-4412-99e1-e9d77cdf4b41-metrics-certs\") pod \"openstack-operator-controller-manager-69ff7bc449-jjnjd\" (UID: \"4fb1fa58-065b-4412-99e1-e9d77cdf4b41\") " pod="openstack-operators/openstack-operator-controller-manager-69ff7bc449-jjnjd" Feb 18 09:15:22 crc kubenswrapper[4556]: I0218 09:15:22.700464 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4fb1fa58-065b-4412-99e1-e9d77cdf4b41-webhook-certs\") pod \"openstack-operator-controller-manager-69ff7bc449-jjnjd\" (UID: \"4fb1fa58-065b-4412-99e1-e9d77cdf4b41\") " pod="openstack-operators/openstack-operator-controller-manager-69ff7bc449-jjnjd" Feb 18 09:15:22 crc kubenswrapper[4556]: E0218 09:15:22.700718 4556 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 18 09:15:22 crc kubenswrapper[4556]: E0218 09:15:22.700767 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4fb1fa58-065b-4412-99e1-e9d77cdf4b41-webhook-certs podName:4fb1fa58-065b-4412-99e1-e9d77cdf4b41 nodeName:}" failed. No retries permitted until 2026-02-18 09:15:24.700751246 +0000 UTC m=+681.717712225 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/4fb1fa58-065b-4412-99e1-e9d77cdf4b41-webhook-certs") pod "openstack-operator-controller-manager-69ff7bc449-jjnjd" (UID: "4fb1fa58-065b-4412-99e1-e9d77cdf4b41") : secret "webhook-server-cert" not found Feb 18 09:15:22 crc kubenswrapper[4556]: E0218 09:15:22.701050 4556 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 18 09:15:22 crc kubenswrapper[4556]: E0218 09:15:22.701087 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4fb1fa58-065b-4412-99e1-e9d77cdf4b41-metrics-certs podName:4fb1fa58-065b-4412-99e1-e9d77cdf4b41 nodeName:}" failed. No retries permitted until 2026-02-18 09:15:24.701078203 +0000 UTC m=+681.718039193 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4fb1fa58-065b-4412-99e1-e9d77cdf4b41-metrics-certs") pod "openstack-operator-controller-manager-69ff7bc449-jjnjd" (UID: "4fb1fa58-065b-4412-99e1-e9d77cdf4b41") : secret "metrics-server-cert" not found Feb 18 09:15:22 crc kubenswrapper[4556]: I0218 09:15:22.776527 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-qr5np" event={"ID":"2c577857-1c7f-4d16-9c13-d3d83eddb3de","Type":"ContainerStarted","Data":"a89bafd56371d68fbd7b13cc659db1fb7256c101fa31f048c5f9affe0f84df5b"} Feb 18 09:15:22 crc kubenswrapper[4556]: E0218 09:15:22.796121 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ironic-operator@sha256:7e1b0b7b172ad0d707ab80dd72d609e1d0f5bbd38a22c24a28ed0f17a960c867\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-lktrc" podUID="4fdde554-179d-4128-a260-38c3de6e9d64" Feb 18 09:15:22 crc kubenswrapper[4556]: E0218 09:15:22.796449 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:3d676f1281e24ef07de617570d2f7fbf625032e41866d1551a856c052248bb04\\\"\"" pod="openstack-operators/swift-operator-controller-manager-68f46476f-jz577" podUID="26bd53bd-69fd-4b14-afd5-5ffb7e557346" Feb 18 09:15:22 crc kubenswrapper[4556]: E0218 09:15:22.796554 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/horizon-operator@sha256:9f2e1299d908411457e53b49e1062265d2b9d76f6719db24d1be9347c388e4da\\\"\"" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-g6hvq" podUID="6a219296-13f5-4c55-99ec-35e34d43a341" Feb 18 09:15:22 crc kubenswrapper[4556]: E0218 09:15:22.797435 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:f0fabdf79095def0f8b1c0442925548a94ca94bed4de2d3b171277129f8079e6\\\"\"" pod="openstack-operators/test-operator-controller-manager-7866795846-9bsld" podUID="bc370f09-39b1-4e31-bbc6-7756c63f0c30" Feb 18 09:15:22 crc kubenswrapper[4556]: E0218 09:15:22.803753 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-vt6jv" podUID="28452a75-3b84-4add-9952-c5f0193e0954" Feb 18 09:15:22 crc kubenswrapper[4556]: E0218 09:15:22.803806 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:d01ae848290e880c09127d5297418dea40fc7f090fdab9bf2c578c7e7f53aec0\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-98pqn" podUID="8e02c422-7a17-4d4f-9ac1-7fe86c5dd472" Feb 18 09:15:24 crc kubenswrapper[4556]: I0218 09:15:24.125950 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/853bc90c-e14b-46ea-912d-5aaefa5f908a-cert\") pod \"infra-operator-controller-manager-79d975b745-gp4rt\" (UID: \"853bc90c-e14b-46ea-912d-5aaefa5f908a\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-gp4rt" Feb 18 09:15:24 crc kubenswrapper[4556]: E0218 09:15:24.126218 4556 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 18 09:15:24 crc kubenswrapper[4556]: E0218 09:15:24.126296 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/853bc90c-e14b-46ea-912d-5aaefa5f908a-cert podName:853bc90c-e14b-46ea-912d-5aaefa5f908a nodeName:}" failed. No retries permitted until 2026-02-18 09:15:28.126271523 +0000 UTC m=+685.143232503 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/853bc90c-e14b-46ea-912d-5aaefa5f908a-cert") pod "infra-operator-controller-manager-79d975b745-gp4rt" (UID: "853bc90c-e14b-46ea-912d-5aaefa5f908a") : secret "infra-operator-webhook-server-cert" not found Feb 18 09:15:24 crc kubenswrapper[4556]: I0218 09:15:24.228673 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/48ca8d44-b660-43ee-8a9e-6489da3cee3d-cert\") pod \"openstack-baremetal-operator-controller-manager-fb5fcc5b8-9mlxh\" (UID: \"48ca8d44-b660-43ee-8a9e-6489da3cee3d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-fb5fcc5b8-9mlxh" Feb 18 09:15:24 crc kubenswrapper[4556]: E0218 09:15:24.229144 4556 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 18 09:15:24 crc kubenswrapper[4556]: E0218 09:15:24.229273 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/48ca8d44-b660-43ee-8a9e-6489da3cee3d-cert podName:48ca8d44-b660-43ee-8a9e-6489da3cee3d nodeName:}" failed. No retries permitted until 2026-02-18 09:15:28.22921998 +0000 UTC m=+685.246180960 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/48ca8d44-b660-43ee-8a9e-6489da3cee3d-cert") pod "openstack-baremetal-operator-controller-manager-fb5fcc5b8-9mlxh" (UID: "48ca8d44-b660-43ee-8a9e-6489da3cee3d") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 18 09:15:24 crc kubenswrapper[4556]: I0218 09:15:24.742962 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4fb1fa58-065b-4412-99e1-e9d77cdf4b41-metrics-certs\") pod \"openstack-operator-controller-manager-69ff7bc449-jjnjd\" (UID: \"4fb1fa58-065b-4412-99e1-e9d77cdf4b41\") " pod="openstack-operators/openstack-operator-controller-manager-69ff7bc449-jjnjd" Feb 18 09:15:24 crc kubenswrapper[4556]: I0218 09:15:24.743109 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4fb1fa58-065b-4412-99e1-e9d77cdf4b41-webhook-certs\") pod \"openstack-operator-controller-manager-69ff7bc449-jjnjd\" (UID: \"4fb1fa58-065b-4412-99e1-e9d77cdf4b41\") " pod="openstack-operators/openstack-operator-controller-manager-69ff7bc449-jjnjd" Feb 18 09:15:24 crc kubenswrapper[4556]: E0218 09:15:24.743139 4556 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 18 09:15:24 crc kubenswrapper[4556]: E0218 09:15:24.743245 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4fb1fa58-065b-4412-99e1-e9d77cdf4b41-metrics-certs podName:4fb1fa58-065b-4412-99e1-e9d77cdf4b41 nodeName:}" failed. No retries permitted until 2026-02-18 09:15:28.743227323 +0000 UTC m=+685.760188302 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4fb1fa58-065b-4412-99e1-e9d77cdf4b41-metrics-certs") pod "openstack-operator-controller-manager-69ff7bc449-jjnjd" (UID: "4fb1fa58-065b-4412-99e1-e9d77cdf4b41") : secret "metrics-server-cert" not found Feb 18 09:15:24 crc kubenswrapper[4556]: E0218 09:15:24.743282 4556 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 18 09:15:24 crc kubenswrapper[4556]: E0218 09:15:24.743360 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4fb1fa58-065b-4412-99e1-e9d77cdf4b41-webhook-certs podName:4fb1fa58-065b-4412-99e1-e9d77cdf4b41 nodeName:}" failed. No retries permitted until 2026-02-18 09:15:28.743343022 +0000 UTC m=+685.760304001 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/4fb1fa58-065b-4412-99e1-e9d77cdf4b41-webhook-certs") pod "openstack-operator-controller-manager-69ff7bc449-jjnjd" (UID: "4fb1fa58-065b-4412-99e1-e9d77cdf4b41") : secret "webhook-server-cert" not found Feb 18 09:15:28 crc kubenswrapper[4556]: I0218 09:15:28.196142 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/853bc90c-e14b-46ea-912d-5aaefa5f908a-cert\") pod \"infra-operator-controller-manager-79d975b745-gp4rt\" (UID: \"853bc90c-e14b-46ea-912d-5aaefa5f908a\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-gp4rt" Feb 18 09:15:28 crc kubenswrapper[4556]: E0218 09:15:28.196661 4556 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 18 09:15:28 crc kubenswrapper[4556]: E0218 09:15:28.196719 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/853bc90c-e14b-46ea-912d-5aaefa5f908a-cert podName:853bc90c-e14b-46ea-912d-5aaefa5f908a nodeName:}" failed. No retries permitted until 2026-02-18 09:15:36.196705871 +0000 UTC m=+693.213666842 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/853bc90c-e14b-46ea-912d-5aaefa5f908a-cert") pod "infra-operator-controller-manager-79d975b745-gp4rt" (UID: "853bc90c-e14b-46ea-912d-5aaefa5f908a") : secret "infra-operator-webhook-server-cert" not found Feb 18 09:15:28 crc kubenswrapper[4556]: I0218 09:15:28.298619 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/48ca8d44-b660-43ee-8a9e-6489da3cee3d-cert\") pod \"openstack-baremetal-operator-controller-manager-fb5fcc5b8-9mlxh\" (UID: \"48ca8d44-b660-43ee-8a9e-6489da3cee3d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-fb5fcc5b8-9mlxh" Feb 18 09:15:28 crc kubenswrapper[4556]: E0218 09:15:28.298817 4556 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 18 09:15:28 crc kubenswrapper[4556]: E0218 09:15:28.298882 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/48ca8d44-b660-43ee-8a9e-6489da3cee3d-cert podName:48ca8d44-b660-43ee-8a9e-6489da3cee3d nodeName:}" failed. No retries permitted until 2026-02-18 09:15:36.29886585 +0000 UTC m=+693.315826840 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/48ca8d44-b660-43ee-8a9e-6489da3cee3d-cert") pod "openstack-baremetal-operator-controller-manager-fb5fcc5b8-9mlxh" (UID: "48ca8d44-b660-43ee-8a9e-6489da3cee3d") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 18 09:15:28 crc kubenswrapper[4556]: I0218 09:15:28.805680 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4fb1fa58-065b-4412-99e1-e9d77cdf4b41-metrics-certs\") pod \"openstack-operator-controller-manager-69ff7bc449-jjnjd\" (UID: \"4fb1fa58-065b-4412-99e1-e9d77cdf4b41\") " pod="openstack-operators/openstack-operator-controller-manager-69ff7bc449-jjnjd" Feb 18 09:15:28 crc kubenswrapper[4556]: I0218 09:15:28.805752 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4fb1fa58-065b-4412-99e1-e9d77cdf4b41-webhook-certs\") pod \"openstack-operator-controller-manager-69ff7bc449-jjnjd\" (UID: \"4fb1fa58-065b-4412-99e1-e9d77cdf4b41\") " pod="openstack-operators/openstack-operator-controller-manager-69ff7bc449-jjnjd" Feb 18 09:15:28 crc kubenswrapper[4556]: E0218 09:15:28.805970 4556 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 18 09:15:28 crc kubenswrapper[4556]: E0218 09:15:28.806035 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4fb1fa58-065b-4412-99e1-e9d77cdf4b41-webhook-certs podName:4fb1fa58-065b-4412-99e1-e9d77cdf4b41 nodeName:}" failed. No retries permitted until 2026-02-18 09:15:36.806021785 +0000 UTC m=+693.822982764 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/4fb1fa58-065b-4412-99e1-e9d77cdf4b41-webhook-certs") pod "openstack-operator-controller-manager-69ff7bc449-jjnjd" (UID: "4fb1fa58-065b-4412-99e1-e9d77cdf4b41") : secret "webhook-server-cert" not found Feb 18 09:15:28 crc kubenswrapper[4556]: E0218 09:15:28.806084 4556 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 18 09:15:28 crc kubenswrapper[4556]: E0218 09:15:28.806185 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4fb1fa58-065b-4412-99e1-e9d77cdf4b41-metrics-certs podName:4fb1fa58-065b-4412-99e1-e9d77cdf4b41 nodeName:}" failed. No retries permitted until 2026-02-18 09:15:36.806135569 +0000 UTC m=+693.823096549 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4fb1fa58-065b-4412-99e1-e9d77cdf4b41-metrics-certs") pod "openstack-operator-controller-manager-69ff7bc449-jjnjd" (UID: "4fb1fa58-065b-4412-99e1-e9d77cdf4b41") : secret "metrics-server-cert" not found Feb 18 09:15:31 crc kubenswrapper[4556]: I0218 09:15:31.727967 4556 patch_prober.go:28] interesting pod/machine-config-daemon-f76hs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 09:15:31 crc kubenswrapper[4556]: I0218 09:15:31.728383 4556 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 09:15:32 crc kubenswrapper[4556]: I0218 09:15:32.864076 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-t9gtc" event={"ID":"d672d582-25f7-4654-bc52-5f07631f5e9e","Type":"ContainerStarted","Data":"81671d6794e2902b27365dd0e80f499def9ac8b6442fb8d7126072f02ca2ebab"} Feb 18 09:15:32 crc kubenswrapper[4556]: I0218 09:15:32.864416 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-t9gtc" Feb 18 09:15:32 crc kubenswrapper[4556]: I0218 09:15:32.866224 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-n4x7d" event={"ID":"463e8b44-a344-401e-bd53-22738560ca32","Type":"ContainerStarted","Data":"cd5c1e01a7c62b95aeeb5506ed4bc0db4e7990f070ad6b0dc91fcd48c1d601a7"} Feb 18 09:15:32 crc kubenswrapper[4556]: I0218 09:15:32.866374 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-n4x7d" Feb 18 09:15:32 crc kubenswrapper[4556]: I0218 09:15:32.869414 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-4f5nl" event={"ID":"acad90fd-fc19-4b74-bb00-192d0e6061a6","Type":"ContainerStarted","Data":"c81c41bcfd0089233937f66bcf1c421b31e77242c42d95b3cbf86e9012386a7c"} Feb 18 09:15:32 crc kubenswrapper[4556]: I0218 09:15:32.869524 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-4f5nl" Feb 18 09:15:32 crc kubenswrapper[4556]: I0218 09:15:32.873013 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-r5kw5" event={"ID":"6e7470ce-ab97-4534-84e8-e951071eb6ee","Type":"ContainerStarted","Data":"301dffab3d4c507ff7451b84feace3bf45332d5b521b8780d468f229a431bd81"} Feb 18 09:15:32 crc kubenswrapper[4556]: I0218 09:15:32.873100 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-r5kw5" Feb 18 09:15:32 crc kubenswrapper[4556]: I0218 09:15:32.874628 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-6sck2" event={"ID":"a980ea11-bf5f-4967-b22b-3e63454bb9ae","Type":"ContainerStarted","Data":"2238237d9f18b55b8e5082de0ac483f467944021b786ebca6f2edbde691ac678"} Feb 18 09:15:32 crc kubenswrapper[4556]: I0218 09:15:32.874746 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-6sck2" Feb 18 09:15:32 crc kubenswrapper[4556]: I0218 09:15:32.877142 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-qr5np" event={"ID":"2c577857-1c7f-4d16-9c13-d3d83eddb3de","Type":"ContainerStarted","Data":"4fd9a102a8e9dc8a46dced9616f683c85485c2dd7b87df402c607e324b8dcfff"} Feb 18 09:15:32 crc kubenswrapper[4556]: I0218 09:15:32.877218 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-qr5np" Feb 18 09:15:32 crc kubenswrapper[4556]: I0218 09:15:32.882475 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-5pv48" event={"ID":"fff15cfd-9701-4efa-81a6-9a482e09ca0f","Type":"ContainerStarted","Data":"854bc3779a84b4cc826ecbe4c6fbe6f5c601165bedd8632fb2833d7dfb9339da"} Feb 18 09:15:32 crc kubenswrapper[4556]: I0218 09:15:32.882543 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-5pv48" Feb 18 09:15:32 crc kubenswrapper[4556]: I0218 09:15:32.884433 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-t2ng7" event={"ID":"fdd8c829-9082-40ae-8b65-46a8bd293c4d","Type":"ContainerStarted","Data":"b8532a6ccaf08d8ccfa0667199775276f4de674c4062a3395eff3153440d3ed1"} Feb 18 09:15:32 crc kubenswrapper[4556]: I0218 09:15:32.884545 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-t2ng7" Feb 18 09:15:32 crc kubenswrapper[4556]: I0218 09:15:32.885961 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-5vgnl" event={"ID":"f2bd3ad9-919a-40ec-b111-f9655a6650a6","Type":"ContainerStarted","Data":"d56e9282de3267ab2b4e70479d7f4dbafbf5f21f3538fc0fa6a705e310bf9dc8"} Feb 18 09:15:32 crc kubenswrapper[4556]: I0218 09:15:32.886061 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-5vgnl" Feb 18 09:15:32 crc kubenswrapper[4556]: I0218 09:15:32.890415 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-rpk9w" event={"ID":"96be73f9-04a8-4ec6-9b75-ae0dd4ae84c8","Type":"ContainerStarted","Data":"906671b31cd67c481418059d5af4264d633d9e48cba80cf4b03cd6cdd175a8d9"} Feb 18 09:15:32 crc kubenswrapper[4556]: I0218 09:15:32.890524 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-rpk9w" Feb 18 09:15:32 crc kubenswrapper[4556]: I0218 09:15:32.893144 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987464f4-mfhjc" event={"ID":"f194b6f2-c143-4a57-8d3e-378a08147713","Type":"ContainerStarted","Data":"0bdfca84ca57678481ab9158c63e3cee33423fa565439616f980461ffd2ef31d"} Feb 18 09:15:32 crc kubenswrapper[4556]: I0218 09:15:32.893279 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-77987464f4-mfhjc" Feb 18 09:15:32 crc kubenswrapper[4556]: I0218 09:15:32.894510 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-x74c8" event={"ID":"9cf84ffa-7a16-4e6a-aa7a-c90d6c1635fe","Type":"ContainerStarted","Data":"4257f2c940a2c2246c47e9f7299abddc95fd637ab8e5a1a6b353f5c8a52145c1"} Feb 18 09:15:32 crc kubenswrapper[4556]: I0218 09:15:32.894904 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-x74c8" Feb 18 09:15:32 crc kubenswrapper[4556]: I0218 09:15:32.896305 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-wdp2t" event={"ID":"9d670c80-f1d0-4b08-b83b-086977d49380","Type":"ContainerStarted","Data":"f752cadfc357eb760afed53cda1f0842fd074a1d82720d567608662f011d6175"} Feb 18 09:15:32 crc kubenswrapper[4556]: I0218 09:15:32.896440 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-wdp2t" Feb 18 09:15:32 crc kubenswrapper[4556]: I0218 09:15:32.911484 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-9rkqc" event={"ID":"cd593415-1b63-4965-9865-21a1ae2e4742","Type":"ContainerStarted","Data":"a6c3dba17a6baa94f0237386a2933cbeda8ee2a129e1a657b4ae22b5892fd2f3"} Feb 18 09:15:32 crc kubenswrapper[4556]: I0218 09:15:32.912097 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-9rkqc" Feb 18 09:15:32 crc kubenswrapper[4556]: I0218 09:15:32.932666 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-t9gtc" podStartSLOduration=2.3616234990000002 podStartE2EDuration="12.932644396s" podCreationTimestamp="2026-02-18 09:15:20 +0000 UTC" firstStartedPulling="2026-02-18 09:15:21.566574217 +0000 UTC m=+678.583535197" lastFinishedPulling="2026-02-18 09:15:32.137595113 +0000 UTC m=+689.154556094" observedRunningTime="2026-02-18 09:15:32.90584348 +0000 UTC m=+689.922804460" watchObservedRunningTime="2026-02-18 09:15:32.932644396 +0000 UTC m=+689.949605376" Feb 18 09:15:32 crc kubenswrapper[4556]: I0218 09:15:32.959930 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-5vgnl" podStartSLOduration=2.440696752 podStartE2EDuration="12.959906463s" podCreationTimestamp="2026-02-18 09:15:20 +0000 UTC" firstStartedPulling="2026-02-18 09:15:21.587129488 +0000 UTC m=+678.604090468" lastFinishedPulling="2026-02-18 09:15:32.106339199 +0000 UTC m=+689.123300179" observedRunningTime="2026-02-18 09:15:32.921093438 +0000 UTC m=+689.938054418" watchObservedRunningTime="2026-02-18 09:15:32.959906463 +0000 UTC m=+689.976867442" Feb 18 09:15:32 crc kubenswrapper[4556]: I0218 09:15:32.976113 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-r5kw5" podStartSLOduration=2.116440611 podStartE2EDuration="12.976095945s" podCreationTimestamp="2026-02-18 09:15:20 +0000 UTC" firstStartedPulling="2026-02-18 09:15:21.246418335 +0000 UTC m=+678.263379315" lastFinishedPulling="2026-02-18 09:15:32.106073668 +0000 UTC m=+689.123034649" observedRunningTime="2026-02-18 09:15:32.973716013 +0000 UTC m=+689.990676994" watchObservedRunningTime="2026-02-18 09:15:32.976095945 +0000 UTC m=+689.993056925" Feb 18 09:15:33 crc kubenswrapper[4556]: I0218 09:15:33.004585 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-5pv48" podStartSLOduration=2.010797289 podStartE2EDuration="13.004562955s" podCreationTimestamp="2026-02-18 09:15:20 +0000 UTC" firstStartedPulling="2026-02-18 09:15:21.112319054 +0000 UTC m=+678.129280023" lastFinishedPulling="2026-02-18 09:15:32.106084709 +0000 UTC m=+689.123045689" observedRunningTime="2026-02-18 09:15:33.002840595 +0000 UTC m=+690.019801575" watchObservedRunningTime="2026-02-18 09:15:33.004562955 +0000 UTC m=+690.021523926" Feb 18 09:15:33 crc kubenswrapper[4556]: I0218 09:15:33.038473 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-4f5nl" podStartSLOduration=2.511795363 podStartE2EDuration="13.038449463s" podCreationTimestamp="2026-02-18 09:15:20 +0000 UTC" firstStartedPulling="2026-02-18 09:15:21.579177622 +0000 UTC m=+678.596138601" lastFinishedPulling="2026-02-18 09:15:32.105831722 +0000 UTC m=+689.122792701" observedRunningTime="2026-02-18 09:15:33.034185736 +0000 UTC m=+690.051146717" watchObservedRunningTime="2026-02-18 09:15:33.038449463 +0000 UTC m=+690.055410443" Feb 18 09:15:33 crc kubenswrapper[4556]: I0218 09:15:33.064242 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-77987464f4-mfhjc" podStartSLOduration=2.06927283 podStartE2EDuration="13.064225336s" podCreationTimestamp="2026-02-18 09:15:20 +0000 UTC" firstStartedPulling="2026-02-18 09:15:21.111406591 +0000 UTC m=+678.128367571" lastFinishedPulling="2026-02-18 09:15:32.106359096 +0000 UTC m=+689.123320077" observedRunningTime="2026-02-18 09:15:33.061614759 +0000 UTC m=+690.078575738" watchObservedRunningTime="2026-02-18 09:15:33.064225336 +0000 UTC m=+690.081186316" Feb 18 09:15:33 crc kubenswrapper[4556]: I0218 09:15:33.079279 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-n4x7d" podStartSLOduration=2.105556021 podStartE2EDuration="13.079241853s" podCreationTimestamp="2026-02-18 09:15:20 +0000 UTC" firstStartedPulling="2026-02-18 09:15:21.131654382 +0000 UTC m=+678.148615362" lastFinishedPulling="2026-02-18 09:15:32.105340213 +0000 UTC m=+689.122301194" observedRunningTime="2026-02-18 09:15:33.077456965 +0000 UTC m=+690.094417945" watchObservedRunningTime="2026-02-18 09:15:33.079241853 +0000 UTC m=+690.096202834" Feb 18 09:15:33 crc kubenswrapper[4556]: I0218 09:15:33.105498 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-wdp2t" podStartSLOduration=2.547892403 podStartE2EDuration="13.105475339s" podCreationTimestamp="2026-02-18 09:15:20 +0000 UTC" firstStartedPulling="2026-02-18 09:15:21.551448183 +0000 UTC m=+678.568409153" lastFinishedPulling="2026-02-18 09:15:32.109031108 +0000 UTC m=+689.125992089" observedRunningTime="2026-02-18 09:15:33.103322307 +0000 UTC m=+690.120283286" watchObservedRunningTime="2026-02-18 09:15:33.105475339 +0000 UTC m=+690.122436320" Feb 18 09:15:33 crc kubenswrapper[4556]: I0218 09:15:33.147679 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-6sck2" podStartSLOduration=2.60051211 podStartE2EDuration="13.14766228s" podCreationTimestamp="2026-02-18 09:15:20 +0000 UTC" firstStartedPulling="2026-02-18 09:15:21.558299291 +0000 UTC m=+678.575260271" lastFinishedPulling="2026-02-18 09:15:32.10544946 +0000 UTC m=+689.122410441" observedRunningTime="2026-02-18 09:15:33.129848664 +0000 UTC m=+690.146809644" watchObservedRunningTime="2026-02-18 09:15:33.14766228 +0000 UTC m=+690.164623261" Feb 18 09:15:33 crc kubenswrapper[4556]: I0218 09:15:33.176564 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-qr5np" podStartSLOduration=2.935669865 podStartE2EDuration="13.176548132s" podCreationTimestamp="2026-02-18 09:15:20 +0000 UTC" firstStartedPulling="2026-02-18 09:15:21.864884334 +0000 UTC m=+678.881845314" lastFinishedPulling="2026-02-18 09:15:32.105762601 +0000 UTC m=+689.122723581" observedRunningTime="2026-02-18 09:15:33.150240296 +0000 UTC m=+690.167201276" watchObservedRunningTime="2026-02-18 09:15:33.176548132 +0000 UTC m=+690.193509111" Feb 18 09:15:33 crc kubenswrapper[4556]: I0218 09:15:33.180240 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-x74c8" podStartSLOduration=2.054508016 podStartE2EDuration="13.180235861s" podCreationTimestamp="2026-02-18 09:15:20 +0000 UTC" firstStartedPulling="2026-02-18 09:15:20.980575307 +0000 UTC m=+677.997536287" lastFinishedPulling="2026-02-18 09:15:32.106303151 +0000 UTC m=+689.123264132" observedRunningTime="2026-02-18 09:15:33.173319579 +0000 UTC m=+690.190280560" watchObservedRunningTime="2026-02-18 09:15:33.180235861 +0000 UTC m=+690.197196841" Feb 18 09:15:33 crc kubenswrapper[4556]: I0218 09:15:33.205606 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-rpk9w" podStartSLOduration=2.6311928289999997 podStartE2EDuration="13.205594796s" podCreationTimestamp="2026-02-18 09:15:20 +0000 UTC" firstStartedPulling="2026-02-18 09:15:21.53105612 +0000 UTC m=+678.548017100" lastFinishedPulling="2026-02-18 09:15:32.105458086 +0000 UTC m=+689.122419067" observedRunningTime="2026-02-18 09:15:33.200124634 +0000 UTC m=+690.217085614" watchObservedRunningTime="2026-02-18 09:15:33.205594796 +0000 UTC m=+690.222555777" Feb 18 09:15:33 crc kubenswrapper[4556]: I0218 09:15:33.241780 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-t2ng7" podStartSLOduration=2.702858421 podStartE2EDuration="13.241765505s" podCreationTimestamp="2026-02-18 09:15:20 +0000 UTC" firstStartedPulling="2026-02-18 09:15:21.566543419 +0000 UTC m=+678.583504399" lastFinishedPulling="2026-02-18 09:15:32.105450502 +0000 UTC m=+689.122411483" observedRunningTime="2026-02-18 09:15:33.240176395 +0000 UTC m=+690.257137376" watchObservedRunningTime="2026-02-18 09:15:33.241765505 +0000 UTC m=+690.258726485" Feb 18 09:15:33 crc kubenswrapper[4556]: I0218 09:15:33.263017 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-9rkqc" podStartSLOduration=2.655925966 podStartE2EDuration="13.263003133s" podCreationTimestamp="2026-02-18 09:15:20 +0000 UTC" firstStartedPulling="2026-02-18 09:15:21.530767957 +0000 UTC m=+678.547728937" lastFinishedPulling="2026-02-18 09:15:32.137845124 +0000 UTC m=+689.154806104" observedRunningTime="2026-02-18 09:15:33.258935487 +0000 UTC m=+690.275896467" watchObservedRunningTime="2026-02-18 09:15:33.263003133 +0000 UTC m=+690.279964113" Feb 18 09:15:35 crc kubenswrapper[4556]: I0218 09:15:35.941188 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-98pqn" event={"ID":"8e02c422-7a17-4d4f-9ac1-7fe86c5dd472","Type":"ContainerStarted","Data":"d239079da9157a6a044533e96fe97245b72b6ae806b54cb5cf907f7599906afb"} Feb 18 09:15:35 crc kubenswrapper[4556]: I0218 09:15:35.941939 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-98pqn" Feb 18 09:15:35 crc kubenswrapper[4556]: I0218 09:15:35.954962 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-98pqn" podStartSLOduration=2.495039262 podStartE2EDuration="15.954946023s" podCreationTimestamp="2026-02-18 09:15:20 +0000 UTC" firstStartedPulling="2026-02-18 09:15:21.648015557 +0000 UTC m=+678.664976537" lastFinishedPulling="2026-02-18 09:15:35.107922327 +0000 UTC m=+692.124883298" observedRunningTime="2026-02-18 09:15:35.951370164 +0000 UTC m=+692.968331144" watchObservedRunningTime="2026-02-18 09:15:35.954946023 +0000 UTC m=+692.971906992" Feb 18 09:15:36 crc kubenswrapper[4556]: I0218 09:15:36.239665 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/853bc90c-e14b-46ea-912d-5aaefa5f908a-cert\") pod \"infra-operator-controller-manager-79d975b745-gp4rt\" (UID: \"853bc90c-e14b-46ea-912d-5aaefa5f908a\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-gp4rt" Feb 18 09:15:36 crc kubenswrapper[4556]: I0218 09:15:36.244727 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/853bc90c-e14b-46ea-912d-5aaefa5f908a-cert\") pod \"infra-operator-controller-manager-79d975b745-gp4rt\" (UID: \"853bc90c-e14b-46ea-912d-5aaefa5f908a\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-gp4rt" Feb 18 09:15:36 crc kubenswrapper[4556]: I0218 09:15:36.342217 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/48ca8d44-b660-43ee-8a9e-6489da3cee3d-cert\") pod \"openstack-baremetal-operator-controller-manager-fb5fcc5b8-9mlxh\" (UID: \"48ca8d44-b660-43ee-8a9e-6489da3cee3d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-fb5fcc5b8-9mlxh" Feb 18 09:15:36 crc kubenswrapper[4556]: I0218 09:15:36.346955 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/48ca8d44-b660-43ee-8a9e-6489da3cee3d-cert\") pod \"openstack-baremetal-operator-controller-manager-fb5fcc5b8-9mlxh\" (UID: \"48ca8d44-b660-43ee-8a9e-6489da3cee3d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-fb5fcc5b8-9mlxh" Feb 18 09:15:36 crc kubenswrapper[4556]: I0218 09:15:36.384049 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-fb5fcc5b8-9mlxh" Feb 18 09:15:36 crc kubenswrapper[4556]: I0218 09:15:36.478104 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-79d975b745-gp4rt" Feb 18 09:15:36 crc kubenswrapper[4556]: I0218 09:15:36.846955 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4fb1fa58-065b-4412-99e1-e9d77cdf4b41-metrics-certs\") pod \"openstack-operator-controller-manager-69ff7bc449-jjnjd\" (UID: \"4fb1fa58-065b-4412-99e1-e9d77cdf4b41\") " pod="openstack-operators/openstack-operator-controller-manager-69ff7bc449-jjnjd" Feb 18 09:15:36 crc kubenswrapper[4556]: E0218 09:15:36.847324 4556 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 18 09:15:36 crc kubenswrapper[4556]: I0218 09:15:36.847349 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4fb1fa58-065b-4412-99e1-e9d77cdf4b41-webhook-certs\") pod \"openstack-operator-controller-manager-69ff7bc449-jjnjd\" (UID: \"4fb1fa58-065b-4412-99e1-e9d77cdf4b41\") " pod="openstack-operators/openstack-operator-controller-manager-69ff7bc449-jjnjd" Feb 18 09:15:36 crc kubenswrapper[4556]: E0218 09:15:36.847394 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4fb1fa58-065b-4412-99e1-e9d77cdf4b41-metrics-certs podName:4fb1fa58-065b-4412-99e1-e9d77cdf4b41 nodeName:}" failed. No retries permitted until 2026-02-18 09:15:52.847379151 +0000 UTC m=+709.864340130 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4fb1fa58-065b-4412-99e1-e9d77cdf4b41-metrics-certs") pod "openstack-operator-controller-manager-69ff7bc449-jjnjd" (UID: "4fb1fa58-065b-4412-99e1-e9d77cdf4b41") : secret "metrics-server-cert" not found Feb 18 09:15:36 crc kubenswrapper[4556]: E0218 09:15:36.847506 4556 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 18 09:15:36 crc kubenswrapper[4556]: E0218 09:15:36.847552 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4fb1fa58-065b-4412-99e1-e9d77cdf4b41-webhook-certs podName:4fb1fa58-065b-4412-99e1-e9d77cdf4b41 nodeName:}" failed. No retries permitted until 2026-02-18 09:15:52.847540314 +0000 UTC m=+709.864501285 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/4fb1fa58-065b-4412-99e1-e9d77cdf4b41-webhook-certs") pod "openstack-operator-controller-manager-69ff7bc449-jjnjd" (UID: "4fb1fa58-065b-4412-99e1-e9d77cdf4b41") : secret "webhook-server-cert" not found Feb 18 09:15:37 crc kubenswrapper[4556]: I0218 09:15:37.161647 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-fb5fcc5b8-9mlxh"] Feb 18 09:15:37 crc kubenswrapper[4556]: W0218 09:15:37.164675 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod48ca8d44_b660_43ee_8a9e_6489da3cee3d.slice/crio-b1c8e0b56fbc68a7ddd74a9a88e63a09cc6af2398b3624e05f1b0c95f2f53675 WatchSource:0}: Error finding container b1c8e0b56fbc68a7ddd74a9a88e63a09cc6af2398b3624e05f1b0c95f2f53675: Status 404 returned error can't find the container with id b1c8e0b56fbc68a7ddd74a9a88e63a09cc6af2398b3624e05f1b0c95f2f53675 Feb 18 09:15:37 crc kubenswrapper[4556]: I0218 09:15:37.309945 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-79d975b745-gp4rt"] Feb 18 09:15:37 crc kubenswrapper[4556]: I0218 09:15:37.963473 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-g6hvq" event={"ID":"6a219296-13f5-4c55-99ec-35e34d43a341","Type":"ContainerStarted","Data":"f45369a7a4bf952366e39210ec91657edcb197a15b61aa4a6ff137b5c4828175"} Feb 18 09:15:37 crc kubenswrapper[4556]: I0218 09:15:37.963969 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-g6hvq" Feb 18 09:15:37 crc kubenswrapper[4556]: I0218 09:15:37.965018 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-fb5fcc5b8-9mlxh" event={"ID":"48ca8d44-b660-43ee-8a9e-6489da3cee3d","Type":"ContainerStarted","Data":"b1c8e0b56fbc68a7ddd74a9a88e63a09cc6af2398b3624e05f1b0c95f2f53675"} Feb 18 09:15:37 crc kubenswrapper[4556]: I0218 09:15:37.967404 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-68f46476f-jz577" event={"ID":"26bd53bd-69fd-4b14-afd5-5ffb7e557346","Type":"ContainerStarted","Data":"13f8d51c43b63e2b627ccae5c8d176891e4c34233fdc146dc257e71f67370183"} Feb 18 09:15:37 crc kubenswrapper[4556]: I0218 09:15:37.967865 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-68f46476f-jz577" Feb 18 09:15:37 crc kubenswrapper[4556]: I0218 09:15:37.969378 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-79d975b745-gp4rt" event={"ID":"853bc90c-e14b-46ea-912d-5aaefa5f908a","Type":"ContainerStarted","Data":"93f31bd50fcc1548a58c9da677c1a0534a6aaedba79b812238a1ec790fc166fc"} Feb 18 09:15:37 crc kubenswrapper[4556]: I0218 09:15:37.987000 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-g6hvq" podStartSLOduration=2.717363114 podStartE2EDuration="17.986971198s" podCreationTimestamp="2026-02-18 09:15:20 +0000 UTC" firstStartedPulling="2026-02-18 09:15:21.64787949 +0000 UTC m=+678.664840470" lastFinishedPulling="2026-02-18 09:15:36.917487574 +0000 UTC m=+693.934448554" observedRunningTime="2026-02-18 09:15:37.97731224 +0000 UTC m=+694.994273220" watchObservedRunningTime="2026-02-18 09:15:37.986971198 +0000 UTC m=+695.003932178" Feb 18 09:15:37 crc kubenswrapper[4556]: I0218 09:15:37.992015 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-68f46476f-jz577" podStartSLOduration=2.689846597 podStartE2EDuration="17.992000799s" podCreationTimestamp="2026-02-18 09:15:20 +0000 UTC" firstStartedPulling="2026-02-18 09:15:21.628180706 +0000 UTC m=+678.645141676" lastFinishedPulling="2026-02-18 09:15:36.930334898 +0000 UTC m=+693.947295878" observedRunningTime="2026-02-18 09:15:37.991942509 +0000 UTC m=+695.008903489" watchObservedRunningTime="2026-02-18 09:15:37.992000799 +0000 UTC m=+695.008961779" Feb 18 09:15:38 crc kubenswrapper[4556]: I0218 09:15:38.978314 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-7866795846-9bsld" event={"ID":"bc370f09-39b1-4e31-bbc6-7756c63f0c30","Type":"ContainerStarted","Data":"0fafa3d4a5d556f47e6f36454d6ee4e24bf8a3dd41f249eed50da405c9d743d6"} Feb 18 09:15:38 crc kubenswrapper[4556]: I0218 09:15:38.978866 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-7866795846-9bsld" Feb 18 09:15:38 crc kubenswrapper[4556]: I0218 09:15:38.997517 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-7866795846-9bsld" podStartSLOduration=2.495322658 podStartE2EDuration="18.997495649s" podCreationTimestamp="2026-02-18 09:15:20 +0000 UTC" firstStartedPulling="2026-02-18 09:15:21.636076817 +0000 UTC m=+678.653037797" lastFinishedPulling="2026-02-18 09:15:38.138249808 +0000 UTC m=+695.155210788" observedRunningTime="2026-02-18 09:15:38.992539245 +0000 UTC m=+696.009500225" watchObservedRunningTime="2026-02-18 09:15:38.997495649 +0000 UTC m=+696.014456629" Feb 18 09:15:40 crc kubenswrapper[4556]: I0218 09:15:40.499805 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-x74c8" Feb 18 09:15:40 crc kubenswrapper[4556]: I0218 09:15:40.511408 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-n4x7d" Feb 18 09:15:40 crc kubenswrapper[4556]: I0218 09:15:40.520007 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-5pv48" Feb 18 09:15:40 crc kubenswrapper[4556]: I0218 09:15:40.539100 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-77987464f4-mfhjc" Feb 18 09:15:40 crc kubenswrapper[4556]: I0218 09:15:40.558058 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-6sck2" Feb 18 09:15:40 crc kubenswrapper[4556]: I0218 09:15:40.620299 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-9rkqc" Feb 18 09:15:40 crc kubenswrapper[4556]: I0218 09:15:40.639628 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-r5kw5" Feb 18 09:15:40 crc kubenswrapper[4556]: I0218 09:15:40.640770 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-t2ng7" Feb 18 09:15:40 crc kubenswrapper[4556]: I0218 09:15:40.655820 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-wdp2t" Feb 18 09:15:40 crc kubenswrapper[4556]: I0218 09:15:40.664426 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-t9gtc" Feb 18 09:15:40 crc kubenswrapper[4556]: I0218 09:15:40.791756 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-rpk9w" Feb 18 09:15:40 crc kubenswrapper[4556]: I0218 09:15:40.833880 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-5vgnl" Feb 18 09:15:40 crc kubenswrapper[4556]: I0218 09:15:40.929988 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-4f5nl" Feb 18 09:15:41 crc kubenswrapper[4556]: I0218 09:15:41.142188 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-98pqn" Feb 18 09:15:41 crc kubenswrapper[4556]: I0218 09:15:41.721063 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-qr5np" Feb 18 09:15:44 crc kubenswrapper[4556]: I0218 09:15:44.009630 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-vt6jv" event={"ID":"28452a75-3b84-4add-9952-c5f0193e0954","Type":"ContainerStarted","Data":"23970b9cc36f7ce6c219f10f26bcebf1027d98d201d0965d6eaeb7813bd35c1a"} Feb 18 09:15:44 crc kubenswrapper[4556]: I0218 09:15:44.011299 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-fb5fcc5b8-9mlxh" event={"ID":"48ca8d44-b660-43ee-8a9e-6489da3cee3d","Type":"ContainerStarted","Data":"a57f361529e39226378779d86903bae794132b2b2ebb6d506f3f797fb3b7a8c8"} Feb 18 09:15:44 crc kubenswrapper[4556]: I0218 09:15:44.011813 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-fb5fcc5b8-9mlxh" Feb 18 09:15:44 crc kubenswrapper[4556]: I0218 09:15:44.012759 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-79d975b745-gp4rt" event={"ID":"853bc90c-e14b-46ea-912d-5aaefa5f908a","Type":"ContainerStarted","Data":"570f77cd40fd5814dc9afb00ceb93444a560ac4f414063f774375b69eec22e76"} Feb 18 09:15:44 crc kubenswrapper[4556]: I0218 09:15:44.012895 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-79d975b745-gp4rt" Feb 18 09:15:44 crc kubenswrapper[4556]: I0218 09:15:44.014346 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-lktrc" event={"ID":"4fdde554-179d-4128-a260-38c3de6e9d64","Type":"ContainerStarted","Data":"9bdd6e18b3fb71e34d9a808428b6590ea0e957dbaa82440cff005a166df3f7b4"} Feb 18 09:15:44 crc kubenswrapper[4556]: I0218 09:15:44.014520 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-lktrc" Feb 18 09:15:44 crc kubenswrapper[4556]: I0218 09:15:44.027575 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-vt6jv" podStartSLOduration=2.160426738 podStartE2EDuration="24.027563851s" podCreationTimestamp="2026-02-18 09:15:20 +0000 UTC" firstStartedPulling="2026-02-18 09:15:21.58810582 +0000 UTC m=+678.605066800" lastFinishedPulling="2026-02-18 09:15:43.455242932 +0000 UTC m=+700.472203913" observedRunningTime="2026-02-18 09:15:44.022327469 +0000 UTC m=+701.039288449" watchObservedRunningTime="2026-02-18 09:15:44.027563851 +0000 UTC m=+701.044524831" Feb 18 09:15:44 crc kubenswrapper[4556]: I0218 09:15:44.042081 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-79d975b745-gp4rt" podStartSLOduration=17.910707219 podStartE2EDuration="24.042067121s" podCreationTimestamp="2026-02-18 09:15:20 +0000 UTC" firstStartedPulling="2026-02-18 09:15:37.309625134 +0000 UTC m=+694.326586114" lastFinishedPulling="2026-02-18 09:15:43.440985036 +0000 UTC m=+700.457946016" observedRunningTime="2026-02-18 09:15:44.040848431 +0000 UTC m=+701.057809411" watchObservedRunningTime="2026-02-18 09:15:44.042067121 +0000 UTC m=+701.059028101" Feb 18 09:15:44 crc kubenswrapper[4556]: I0218 09:15:44.054683 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-lktrc" podStartSLOduration=2.251376432 podStartE2EDuration="24.054664643s" podCreationTimestamp="2026-02-18 09:15:20 +0000 UTC" firstStartedPulling="2026-02-18 09:15:21.641558873 +0000 UTC m=+678.658519853" lastFinishedPulling="2026-02-18 09:15:43.444847085 +0000 UTC m=+700.461808064" observedRunningTime="2026-02-18 09:15:44.049710024 +0000 UTC m=+701.066671004" watchObservedRunningTime="2026-02-18 09:15:44.054664643 +0000 UTC m=+701.071625623" Feb 18 09:15:44 crc kubenswrapper[4556]: I0218 09:15:44.074455 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-fb5fcc5b8-9mlxh" podStartSLOduration=17.803983777 podStartE2EDuration="24.074440352s" podCreationTimestamp="2026-02-18 09:15:20 +0000 UTC" firstStartedPulling="2026-02-18 09:15:37.167529252 +0000 UTC m=+694.184490232" lastFinishedPulling="2026-02-18 09:15:43.437985826 +0000 UTC m=+700.454946807" observedRunningTime="2026-02-18 09:15:44.069632931 +0000 UTC m=+701.086593911" watchObservedRunningTime="2026-02-18 09:15:44.074440352 +0000 UTC m=+701.091401332" Feb 18 09:15:50 crc kubenswrapper[4556]: I0218 09:15:50.846486 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-68f46476f-jz577" Feb 18 09:15:50 crc kubenswrapper[4556]: I0218 09:15:50.865452 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-g6hvq" Feb 18 09:15:50 crc kubenswrapper[4556]: I0218 09:15:50.900341 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-lktrc" Feb 18 09:15:51 crc kubenswrapper[4556]: I0218 09:15:51.016231 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-7866795846-9bsld" Feb 18 09:15:52 crc kubenswrapper[4556]: I0218 09:15:52.886270 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4fb1fa58-065b-4412-99e1-e9d77cdf4b41-metrics-certs\") pod \"openstack-operator-controller-manager-69ff7bc449-jjnjd\" (UID: \"4fb1fa58-065b-4412-99e1-e9d77cdf4b41\") " pod="openstack-operators/openstack-operator-controller-manager-69ff7bc449-jjnjd" Feb 18 09:15:52 crc kubenswrapper[4556]: I0218 09:15:52.886329 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4fb1fa58-065b-4412-99e1-e9d77cdf4b41-webhook-certs\") pod \"openstack-operator-controller-manager-69ff7bc449-jjnjd\" (UID: \"4fb1fa58-065b-4412-99e1-e9d77cdf4b41\") " pod="openstack-operators/openstack-operator-controller-manager-69ff7bc449-jjnjd" Feb 18 09:15:52 crc kubenswrapper[4556]: I0218 09:15:52.891502 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4fb1fa58-065b-4412-99e1-e9d77cdf4b41-webhook-certs\") pod \"openstack-operator-controller-manager-69ff7bc449-jjnjd\" (UID: \"4fb1fa58-065b-4412-99e1-e9d77cdf4b41\") " pod="openstack-operators/openstack-operator-controller-manager-69ff7bc449-jjnjd" Feb 18 09:15:52 crc kubenswrapper[4556]: I0218 09:15:52.891974 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4fb1fa58-065b-4412-99e1-e9d77cdf4b41-metrics-certs\") pod \"openstack-operator-controller-manager-69ff7bc449-jjnjd\" (UID: \"4fb1fa58-065b-4412-99e1-e9d77cdf4b41\") " pod="openstack-operators/openstack-operator-controller-manager-69ff7bc449-jjnjd" Feb 18 09:15:53 crc kubenswrapper[4556]: I0218 09:15:53.008978 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-69ff7bc449-jjnjd" Feb 18 09:15:53 crc kubenswrapper[4556]: I0218 09:15:53.381831 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-69ff7bc449-jjnjd"] Feb 18 09:15:54 crc kubenswrapper[4556]: I0218 09:15:54.081186 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-69ff7bc449-jjnjd" event={"ID":"4fb1fa58-065b-4412-99e1-e9d77cdf4b41","Type":"ContainerStarted","Data":"d60f614c188d7629dc8f84f5659c475657ee8e4d0448b4e57a9844ef49272bd0"} Feb 18 09:15:54 crc kubenswrapper[4556]: I0218 09:15:54.081470 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-69ff7bc449-jjnjd" event={"ID":"4fb1fa58-065b-4412-99e1-e9d77cdf4b41","Type":"ContainerStarted","Data":"4f2745acb0e4045651f5d4bf6e0f59b017f8707b446298fa255272920cf2736e"} Feb 18 09:15:54 crc kubenswrapper[4556]: I0218 09:15:54.081488 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-69ff7bc449-jjnjd" Feb 18 09:15:54 crc kubenswrapper[4556]: I0218 09:15:54.100496 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-69ff7bc449-jjnjd" podStartSLOduration=34.100480995 podStartE2EDuration="34.100480995s" podCreationTimestamp="2026-02-18 09:15:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:15:54.098598232 +0000 UTC m=+711.115559211" watchObservedRunningTime="2026-02-18 09:15:54.100480995 +0000 UTC m=+711.117441975" Feb 18 09:15:56 crc kubenswrapper[4556]: I0218 09:15:56.390447 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-fb5fcc5b8-9mlxh" Feb 18 09:15:56 crc kubenswrapper[4556]: I0218 09:15:56.482648 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-79d975b745-gp4rt" Feb 18 09:16:01 crc kubenswrapper[4556]: I0218 09:16:01.728058 4556 patch_prober.go:28] interesting pod/machine-config-daemon-f76hs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 09:16:01 crc kubenswrapper[4556]: I0218 09:16:01.728514 4556 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 09:16:03 crc kubenswrapper[4556]: I0218 09:16:03.013503 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-69ff7bc449-jjnjd" Feb 18 09:16:11 crc kubenswrapper[4556]: I0218 09:16:11.219030 4556 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 18 09:16:17 crc kubenswrapper[4556]: I0218 09:16:17.917068 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-855cbc58c5-dh8dd"] Feb 18 09:16:17 crc kubenswrapper[4556]: I0218 09:16:17.918494 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-855cbc58c5-dh8dd" Feb 18 09:16:17 crc kubenswrapper[4556]: I0218 09:16:17.929794 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Feb 18 09:16:17 crc kubenswrapper[4556]: I0218 09:16:17.930013 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Feb 18 09:16:17 crc kubenswrapper[4556]: I0218 09:16:17.930110 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Feb 18 09:16:17 crc kubenswrapper[4556]: I0218 09:16:17.930584 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-zljxm" Feb 18 09:16:17 crc kubenswrapper[4556]: I0218 09:16:17.933990 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-855cbc58c5-dh8dd"] Feb 18 09:16:17 crc kubenswrapper[4556]: I0218 09:16:17.965207 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6fcf94d689-vsztz"] Feb 18 09:16:17 crc kubenswrapper[4556]: I0218 09:16:17.966224 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6fcf94d689-vsztz" Feb 18 09:16:17 crc kubenswrapper[4556]: I0218 09:16:17.968285 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Feb 18 09:16:17 crc kubenswrapper[4556]: I0218 09:16:17.973173 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6fcf94d689-vsztz"] Feb 18 09:16:18 crc kubenswrapper[4556]: I0218 09:16:18.080824 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fd56f2a-ea59-4b71-8468-56d18eda194a-config\") pod \"dnsmasq-dns-855cbc58c5-dh8dd\" (UID: \"2fd56f2a-ea59-4b71-8468-56d18eda194a\") " pod="openstack/dnsmasq-dns-855cbc58c5-dh8dd" Feb 18 09:16:18 crc kubenswrapper[4556]: I0218 09:16:18.080870 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2cpxc\" (UniqueName: \"kubernetes.io/projected/2fd56f2a-ea59-4b71-8468-56d18eda194a-kube-api-access-2cpxc\") pod \"dnsmasq-dns-855cbc58c5-dh8dd\" (UID: \"2fd56f2a-ea59-4b71-8468-56d18eda194a\") " pod="openstack/dnsmasq-dns-855cbc58c5-dh8dd" Feb 18 09:16:18 crc kubenswrapper[4556]: I0218 09:16:18.080917 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jz862\" (UniqueName: \"kubernetes.io/projected/ab4c456b-9c1a-4bd8-a35a-fe73617002c9-kube-api-access-jz862\") pod \"dnsmasq-dns-6fcf94d689-vsztz\" (UID: \"ab4c456b-9c1a-4bd8-a35a-fe73617002c9\") " pod="openstack/dnsmasq-dns-6fcf94d689-vsztz" Feb 18 09:16:18 crc kubenswrapper[4556]: I0218 09:16:18.080946 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ab4c456b-9c1a-4bd8-a35a-fe73617002c9-dns-svc\") pod \"dnsmasq-dns-6fcf94d689-vsztz\" (UID: \"ab4c456b-9c1a-4bd8-a35a-fe73617002c9\") " pod="openstack/dnsmasq-dns-6fcf94d689-vsztz" Feb 18 09:16:18 crc kubenswrapper[4556]: I0218 09:16:18.080962 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab4c456b-9c1a-4bd8-a35a-fe73617002c9-config\") pod \"dnsmasq-dns-6fcf94d689-vsztz\" (UID: \"ab4c456b-9c1a-4bd8-a35a-fe73617002c9\") " pod="openstack/dnsmasq-dns-6fcf94d689-vsztz" Feb 18 09:16:18 crc kubenswrapper[4556]: I0218 09:16:18.182034 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ab4c456b-9c1a-4bd8-a35a-fe73617002c9-dns-svc\") pod \"dnsmasq-dns-6fcf94d689-vsztz\" (UID: \"ab4c456b-9c1a-4bd8-a35a-fe73617002c9\") " pod="openstack/dnsmasq-dns-6fcf94d689-vsztz" Feb 18 09:16:18 crc kubenswrapper[4556]: I0218 09:16:18.182069 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab4c456b-9c1a-4bd8-a35a-fe73617002c9-config\") pod \"dnsmasq-dns-6fcf94d689-vsztz\" (UID: \"ab4c456b-9c1a-4bd8-a35a-fe73617002c9\") " pod="openstack/dnsmasq-dns-6fcf94d689-vsztz" Feb 18 09:16:18 crc kubenswrapper[4556]: I0218 09:16:18.182113 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fd56f2a-ea59-4b71-8468-56d18eda194a-config\") pod \"dnsmasq-dns-855cbc58c5-dh8dd\" (UID: \"2fd56f2a-ea59-4b71-8468-56d18eda194a\") " pod="openstack/dnsmasq-dns-855cbc58c5-dh8dd" Feb 18 09:16:18 crc kubenswrapper[4556]: I0218 09:16:18.182141 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2cpxc\" (UniqueName: \"kubernetes.io/projected/2fd56f2a-ea59-4b71-8468-56d18eda194a-kube-api-access-2cpxc\") pod \"dnsmasq-dns-855cbc58c5-dh8dd\" (UID: \"2fd56f2a-ea59-4b71-8468-56d18eda194a\") " pod="openstack/dnsmasq-dns-855cbc58c5-dh8dd" Feb 18 09:16:18 crc kubenswrapper[4556]: I0218 09:16:18.182202 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jz862\" (UniqueName: \"kubernetes.io/projected/ab4c456b-9c1a-4bd8-a35a-fe73617002c9-kube-api-access-jz862\") pod \"dnsmasq-dns-6fcf94d689-vsztz\" (UID: \"ab4c456b-9c1a-4bd8-a35a-fe73617002c9\") " pod="openstack/dnsmasq-dns-6fcf94d689-vsztz" Feb 18 09:16:18 crc kubenswrapper[4556]: I0218 09:16:18.183102 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ab4c456b-9c1a-4bd8-a35a-fe73617002c9-dns-svc\") pod \"dnsmasq-dns-6fcf94d689-vsztz\" (UID: \"ab4c456b-9c1a-4bd8-a35a-fe73617002c9\") " pod="openstack/dnsmasq-dns-6fcf94d689-vsztz" Feb 18 09:16:18 crc kubenswrapper[4556]: I0218 09:16:18.183106 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab4c456b-9c1a-4bd8-a35a-fe73617002c9-config\") pod \"dnsmasq-dns-6fcf94d689-vsztz\" (UID: \"ab4c456b-9c1a-4bd8-a35a-fe73617002c9\") " pod="openstack/dnsmasq-dns-6fcf94d689-vsztz" Feb 18 09:16:18 crc kubenswrapper[4556]: I0218 09:16:18.183237 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fd56f2a-ea59-4b71-8468-56d18eda194a-config\") pod \"dnsmasq-dns-855cbc58c5-dh8dd\" (UID: \"2fd56f2a-ea59-4b71-8468-56d18eda194a\") " pod="openstack/dnsmasq-dns-855cbc58c5-dh8dd" Feb 18 09:16:18 crc kubenswrapper[4556]: I0218 09:16:18.197829 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jz862\" (UniqueName: \"kubernetes.io/projected/ab4c456b-9c1a-4bd8-a35a-fe73617002c9-kube-api-access-jz862\") pod \"dnsmasq-dns-6fcf94d689-vsztz\" (UID: \"ab4c456b-9c1a-4bd8-a35a-fe73617002c9\") " pod="openstack/dnsmasq-dns-6fcf94d689-vsztz" Feb 18 09:16:18 crc kubenswrapper[4556]: I0218 09:16:18.197842 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2cpxc\" (UniqueName: \"kubernetes.io/projected/2fd56f2a-ea59-4b71-8468-56d18eda194a-kube-api-access-2cpxc\") pod \"dnsmasq-dns-855cbc58c5-dh8dd\" (UID: \"2fd56f2a-ea59-4b71-8468-56d18eda194a\") " pod="openstack/dnsmasq-dns-855cbc58c5-dh8dd" Feb 18 09:16:18 crc kubenswrapper[4556]: I0218 09:16:18.232485 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-855cbc58c5-dh8dd" Feb 18 09:16:18 crc kubenswrapper[4556]: I0218 09:16:18.279318 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6fcf94d689-vsztz" Feb 18 09:16:18 crc kubenswrapper[4556]: I0218 09:16:18.605475 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-855cbc58c5-dh8dd"] Feb 18 09:16:18 crc kubenswrapper[4556]: I0218 09:16:18.665343 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6fcf94d689-vsztz"] Feb 18 09:16:18 crc kubenswrapper[4556]: W0218 09:16:18.668492 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podab4c456b_9c1a_4bd8_a35a_fe73617002c9.slice/crio-99b51f22c992596c809d36264d75223901c85c5053091409733a31c4adad6e49 WatchSource:0}: Error finding container 99b51f22c992596c809d36264d75223901c85c5053091409733a31c4adad6e49: Status 404 returned error can't find the container with id 99b51f22c992596c809d36264d75223901c85c5053091409733a31c4adad6e49 Feb 18 09:16:19 crc kubenswrapper[4556]: I0218 09:16:19.204609 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6fcf94d689-vsztz" event={"ID":"ab4c456b-9c1a-4bd8-a35a-fe73617002c9","Type":"ContainerStarted","Data":"99b51f22c992596c809d36264d75223901c85c5053091409733a31c4adad6e49"} Feb 18 09:16:19 crc kubenswrapper[4556]: I0218 09:16:19.205664 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-855cbc58c5-dh8dd" event={"ID":"2fd56f2a-ea59-4b71-8468-56d18eda194a","Type":"ContainerStarted","Data":"f8ff36375c709a7773baceded667406673606a77aff7bd08a32b83de0b697e6c"} Feb 18 09:16:20 crc kubenswrapper[4556]: I0218 09:16:20.722188 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6fcf94d689-vsztz"] Feb 18 09:16:20 crc kubenswrapper[4556]: I0218 09:16:20.740045 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-f54874ffc-rb6x9"] Feb 18 09:16:20 crc kubenswrapper[4556]: I0218 09:16:20.741065 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f54874ffc-rb6x9" Feb 18 09:16:20 crc kubenswrapper[4556]: I0218 09:16:20.749421 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f54874ffc-rb6x9"] Feb 18 09:16:20 crc kubenswrapper[4556]: I0218 09:16:20.919638 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/78221461-c2ed-45d9-bd36-aa078d36d757-dns-svc\") pod \"dnsmasq-dns-f54874ffc-rb6x9\" (UID: \"78221461-c2ed-45d9-bd36-aa078d36d757\") " pod="openstack/dnsmasq-dns-f54874ffc-rb6x9" Feb 18 09:16:20 crc kubenswrapper[4556]: I0218 09:16:20.919963 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pph7d\" (UniqueName: \"kubernetes.io/projected/78221461-c2ed-45d9-bd36-aa078d36d757-kube-api-access-pph7d\") pod \"dnsmasq-dns-f54874ffc-rb6x9\" (UID: \"78221461-c2ed-45d9-bd36-aa078d36d757\") " pod="openstack/dnsmasq-dns-f54874ffc-rb6x9" Feb 18 09:16:20 crc kubenswrapper[4556]: I0218 09:16:20.920010 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78221461-c2ed-45d9-bd36-aa078d36d757-config\") pod \"dnsmasq-dns-f54874ffc-rb6x9\" (UID: \"78221461-c2ed-45d9-bd36-aa078d36d757\") " pod="openstack/dnsmasq-dns-f54874ffc-rb6x9" Feb 18 09:16:21 crc kubenswrapper[4556]: I0218 09:16:21.021659 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pph7d\" (UniqueName: \"kubernetes.io/projected/78221461-c2ed-45d9-bd36-aa078d36d757-kube-api-access-pph7d\") pod \"dnsmasq-dns-f54874ffc-rb6x9\" (UID: \"78221461-c2ed-45d9-bd36-aa078d36d757\") " pod="openstack/dnsmasq-dns-f54874ffc-rb6x9" Feb 18 09:16:21 crc kubenswrapper[4556]: I0218 09:16:21.021713 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78221461-c2ed-45d9-bd36-aa078d36d757-config\") pod \"dnsmasq-dns-f54874ffc-rb6x9\" (UID: \"78221461-c2ed-45d9-bd36-aa078d36d757\") " pod="openstack/dnsmasq-dns-f54874ffc-rb6x9" Feb 18 09:16:21 crc kubenswrapper[4556]: I0218 09:16:21.021757 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/78221461-c2ed-45d9-bd36-aa078d36d757-dns-svc\") pod \"dnsmasq-dns-f54874ffc-rb6x9\" (UID: \"78221461-c2ed-45d9-bd36-aa078d36d757\") " pod="openstack/dnsmasq-dns-f54874ffc-rb6x9" Feb 18 09:16:21 crc kubenswrapper[4556]: I0218 09:16:21.022545 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/78221461-c2ed-45d9-bd36-aa078d36d757-dns-svc\") pod \"dnsmasq-dns-f54874ffc-rb6x9\" (UID: \"78221461-c2ed-45d9-bd36-aa078d36d757\") " pod="openstack/dnsmasq-dns-f54874ffc-rb6x9" Feb 18 09:16:21 crc kubenswrapper[4556]: I0218 09:16:21.022550 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78221461-c2ed-45d9-bd36-aa078d36d757-config\") pod \"dnsmasq-dns-f54874ffc-rb6x9\" (UID: \"78221461-c2ed-45d9-bd36-aa078d36d757\") " pod="openstack/dnsmasq-dns-f54874ffc-rb6x9" Feb 18 09:16:21 crc kubenswrapper[4556]: I0218 09:16:21.024927 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-855cbc58c5-dh8dd"] Feb 18 09:16:21 crc kubenswrapper[4556]: I0218 09:16:21.041063 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pph7d\" (UniqueName: \"kubernetes.io/projected/78221461-c2ed-45d9-bd36-aa078d36d757-kube-api-access-pph7d\") pod \"dnsmasq-dns-f54874ffc-rb6x9\" (UID: \"78221461-c2ed-45d9-bd36-aa078d36d757\") " pod="openstack/dnsmasq-dns-f54874ffc-rb6x9" Feb 18 09:16:21 crc kubenswrapper[4556]: I0218 09:16:21.052890 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-67ff45466c-wdkxp"] Feb 18 09:16:21 crc kubenswrapper[4556]: I0218 09:16:21.053906 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67ff45466c-wdkxp" Feb 18 09:16:21 crc kubenswrapper[4556]: I0218 09:16:21.061791 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67ff45466c-wdkxp"] Feb 18 09:16:21 crc kubenswrapper[4556]: I0218 09:16:21.070144 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f54874ffc-rb6x9" Feb 18 09:16:21 crc kubenswrapper[4556]: I0218 09:16:21.123384 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmrpp\" (UniqueName: \"kubernetes.io/projected/fa586b98-2a59-4673-a0bf-f7175f30d356-kube-api-access-wmrpp\") pod \"dnsmasq-dns-67ff45466c-wdkxp\" (UID: \"fa586b98-2a59-4673-a0bf-f7175f30d356\") " pod="openstack/dnsmasq-dns-67ff45466c-wdkxp" Feb 18 09:16:21 crc kubenswrapper[4556]: I0218 09:16:21.123609 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fa586b98-2a59-4673-a0bf-f7175f30d356-dns-svc\") pod \"dnsmasq-dns-67ff45466c-wdkxp\" (UID: \"fa586b98-2a59-4673-a0bf-f7175f30d356\") " pod="openstack/dnsmasq-dns-67ff45466c-wdkxp" Feb 18 09:16:21 crc kubenswrapper[4556]: I0218 09:16:21.123689 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa586b98-2a59-4673-a0bf-f7175f30d356-config\") pod \"dnsmasq-dns-67ff45466c-wdkxp\" (UID: \"fa586b98-2a59-4673-a0bf-f7175f30d356\") " pod="openstack/dnsmasq-dns-67ff45466c-wdkxp" Feb 18 09:16:21 crc kubenswrapper[4556]: I0218 09:16:21.226238 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa586b98-2a59-4673-a0bf-f7175f30d356-config\") pod \"dnsmasq-dns-67ff45466c-wdkxp\" (UID: \"fa586b98-2a59-4673-a0bf-f7175f30d356\") " pod="openstack/dnsmasq-dns-67ff45466c-wdkxp" Feb 18 09:16:21 crc kubenswrapper[4556]: I0218 09:16:21.226364 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmrpp\" (UniqueName: \"kubernetes.io/projected/fa586b98-2a59-4673-a0bf-f7175f30d356-kube-api-access-wmrpp\") pod \"dnsmasq-dns-67ff45466c-wdkxp\" (UID: \"fa586b98-2a59-4673-a0bf-f7175f30d356\") " pod="openstack/dnsmasq-dns-67ff45466c-wdkxp" Feb 18 09:16:21 crc kubenswrapper[4556]: I0218 09:16:21.226385 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fa586b98-2a59-4673-a0bf-f7175f30d356-dns-svc\") pod \"dnsmasq-dns-67ff45466c-wdkxp\" (UID: \"fa586b98-2a59-4673-a0bf-f7175f30d356\") " pod="openstack/dnsmasq-dns-67ff45466c-wdkxp" Feb 18 09:16:21 crc kubenswrapper[4556]: I0218 09:16:21.227278 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fa586b98-2a59-4673-a0bf-f7175f30d356-dns-svc\") pod \"dnsmasq-dns-67ff45466c-wdkxp\" (UID: \"fa586b98-2a59-4673-a0bf-f7175f30d356\") " pod="openstack/dnsmasq-dns-67ff45466c-wdkxp" Feb 18 09:16:21 crc kubenswrapper[4556]: I0218 09:16:21.227658 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa586b98-2a59-4673-a0bf-f7175f30d356-config\") pod \"dnsmasq-dns-67ff45466c-wdkxp\" (UID: \"fa586b98-2a59-4673-a0bf-f7175f30d356\") " pod="openstack/dnsmasq-dns-67ff45466c-wdkxp" Feb 18 09:16:21 crc kubenswrapper[4556]: I0218 09:16:21.246352 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmrpp\" (UniqueName: \"kubernetes.io/projected/fa586b98-2a59-4673-a0bf-f7175f30d356-kube-api-access-wmrpp\") pod \"dnsmasq-dns-67ff45466c-wdkxp\" (UID: \"fa586b98-2a59-4673-a0bf-f7175f30d356\") " pod="openstack/dnsmasq-dns-67ff45466c-wdkxp" Feb 18 09:16:21 crc kubenswrapper[4556]: I0218 09:16:21.380513 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67ff45466c-wdkxp" Feb 18 09:16:21 crc kubenswrapper[4556]: I0218 09:16:21.502013 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f54874ffc-rb6x9"] Feb 18 09:16:21 crc kubenswrapper[4556]: W0218 09:16:21.512199 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod78221461_c2ed_45d9_bd36_aa078d36d757.slice/crio-5d38656375670fe555a05d1c74743fce293b4370622cd92c9cce18bd0ace0a44 WatchSource:0}: Error finding container 5d38656375670fe555a05d1c74743fce293b4370622cd92c9cce18bd0ace0a44: Status 404 returned error can't find the container with id 5d38656375670fe555a05d1c74743fce293b4370622cd92c9cce18bd0ace0a44 Feb 18 09:16:21 crc kubenswrapper[4556]: I0218 09:16:21.769030 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67ff45466c-wdkxp"] Feb 18 09:16:21 crc kubenswrapper[4556]: W0218 09:16:21.775066 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfa586b98_2a59_4673_a0bf_f7175f30d356.slice/crio-031b6d98624d5f4f1ccbf8ef80d5c2c14a3af7a2b1d6eaa11caf0fb3954e993c WatchSource:0}: Error finding container 031b6d98624d5f4f1ccbf8ef80d5c2c14a3af7a2b1d6eaa11caf0fb3954e993c: Status 404 returned error can't find the container with id 031b6d98624d5f4f1ccbf8ef80d5c2c14a3af7a2b1d6eaa11caf0fb3954e993c Feb 18 09:16:21 crc kubenswrapper[4556]: I0218 09:16:21.866231 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Feb 18 09:16:21 crc kubenswrapper[4556]: I0218 09:16:21.867201 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 18 09:16:21 crc kubenswrapper[4556]: I0218 09:16:21.871240 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Feb 18 09:16:21 crc kubenswrapper[4556]: I0218 09:16:21.871294 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Feb 18 09:16:21 crc kubenswrapper[4556]: I0218 09:16:21.876423 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Feb 18 09:16:21 crc kubenswrapper[4556]: I0218 09:16:21.876491 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Feb 18 09:16:21 crc kubenswrapper[4556]: I0218 09:16:21.876497 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Feb 18 09:16:21 crc kubenswrapper[4556]: I0218 09:16:21.876561 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Feb 18 09:16:21 crc kubenswrapper[4556]: I0218 09:16:21.876635 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-rmfvs" Feb 18 09:16:21 crc kubenswrapper[4556]: I0218 09:16:21.882611 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.040669 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1d7961e1-6f96-4707-907d-1fa0c30641e2-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"1d7961e1-6f96-4707-907d-1fa0c30641e2\") " pod="openstack/rabbitmq-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.040722 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1d7961e1-6f96-4707-907d-1fa0c30641e2-server-conf\") pod \"rabbitmq-server-0\" (UID: \"1d7961e1-6f96-4707-907d-1fa0c30641e2\") " pod="openstack/rabbitmq-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.040773 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"1d7961e1-6f96-4707-907d-1fa0c30641e2\") " pod="openstack/rabbitmq-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.040821 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1d7961e1-6f96-4707-907d-1fa0c30641e2-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"1d7961e1-6f96-4707-907d-1fa0c30641e2\") " pod="openstack/rabbitmq-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.040911 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1d7961e1-6f96-4707-907d-1fa0c30641e2-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"1d7961e1-6f96-4707-907d-1fa0c30641e2\") " pod="openstack/rabbitmq-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.040950 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1d7961e1-6f96-4707-907d-1fa0c30641e2-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"1d7961e1-6f96-4707-907d-1fa0c30641e2\") " pod="openstack/rabbitmq-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.041012 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1d7961e1-6f96-4707-907d-1fa0c30641e2-pod-info\") pod \"rabbitmq-server-0\" (UID: \"1d7961e1-6f96-4707-907d-1fa0c30641e2\") " pod="openstack/rabbitmq-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.041054 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6n8rf\" (UniqueName: \"kubernetes.io/projected/1d7961e1-6f96-4707-907d-1fa0c30641e2-kube-api-access-6n8rf\") pod \"rabbitmq-server-0\" (UID: \"1d7961e1-6f96-4707-907d-1fa0c30641e2\") " pod="openstack/rabbitmq-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.041084 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1d7961e1-6f96-4707-907d-1fa0c30641e2-config-data\") pod \"rabbitmq-server-0\" (UID: \"1d7961e1-6f96-4707-907d-1fa0c30641e2\") " pod="openstack/rabbitmq-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.041129 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1d7961e1-6f96-4707-907d-1fa0c30641e2-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"1d7961e1-6f96-4707-907d-1fa0c30641e2\") " pod="openstack/rabbitmq-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.041180 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1d7961e1-6f96-4707-907d-1fa0c30641e2-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"1d7961e1-6f96-4707-907d-1fa0c30641e2\") " pod="openstack/rabbitmq-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.126140 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.127418 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.129217 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.131862 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.132016 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.132804 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.134821 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.136174 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-ksd85" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.136325 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.142274 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1d7961e1-6f96-4707-907d-1fa0c30641e2-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"1d7961e1-6f96-4707-907d-1fa0c30641e2\") " pod="openstack/rabbitmq-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.142305 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1d7961e1-6f96-4707-907d-1fa0c30641e2-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"1d7961e1-6f96-4707-907d-1fa0c30641e2\") " pod="openstack/rabbitmq-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.142326 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1d7961e1-6f96-4707-907d-1fa0c30641e2-server-conf\") pod \"rabbitmq-server-0\" (UID: \"1d7961e1-6f96-4707-907d-1fa0c30641e2\") " pod="openstack/rabbitmq-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.142344 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"1d7961e1-6f96-4707-907d-1fa0c30641e2\") " pod="openstack/rabbitmq-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.142388 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1d7961e1-6f96-4707-907d-1fa0c30641e2-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"1d7961e1-6f96-4707-907d-1fa0c30641e2\") " pod="openstack/rabbitmq-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.142426 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1d7961e1-6f96-4707-907d-1fa0c30641e2-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"1d7961e1-6f96-4707-907d-1fa0c30641e2\") " pod="openstack/rabbitmq-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.142463 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1d7961e1-6f96-4707-907d-1fa0c30641e2-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"1d7961e1-6f96-4707-907d-1fa0c30641e2\") " pod="openstack/rabbitmq-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.142504 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1d7961e1-6f96-4707-907d-1fa0c30641e2-pod-info\") pod \"rabbitmq-server-0\" (UID: \"1d7961e1-6f96-4707-907d-1fa0c30641e2\") " pod="openstack/rabbitmq-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.142526 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6n8rf\" (UniqueName: \"kubernetes.io/projected/1d7961e1-6f96-4707-907d-1fa0c30641e2-kube-api-access-6n8rf\") pod \"rabbitmq-server-0\" (UID: \"1d7961e1-6f96-4707-907d-1fa0c30641e2\") " pod="openstack/rabbitmq-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.142544 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1d7961e1-6f96-4707-907d-1fa0c30641e2-config-data\") pod \"rabbitmq-server-0\" (UID: \"1d7961e1-6f96-4707-907d-1fa0c30641e2\") " pod="openstack/rabbitmq-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.142566 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1d7961e1-6f96-4707-907d-1fa0c30641e2-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"1d7961e1-6f96-4707-907d-1fa0c30641e2\") " pod="openstack/rabbitmq-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.143964 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1d7961e1-6f96-4707-907d-1fa0c30641e2-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"1d7961e1-6f96-4707-907d-1fa0c30641e2\") " pod="openstack/rabbitmq-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.145511 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1d7961e1-6f96-4707-907d-1fa0c30641e2-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"1d7961e1-6f96-4707-907d-1fa0c30641e2\") " pod="openstack/rabbitmq-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.146013 4556 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"1d7961e1-6f96-4707-907d-1fa0c30641e2\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.146526 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1d7961e1-6f96-4707-907d-1fa0c30641e2-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"1d7961e1-6f96-4707-907d-1fa0c30641e2\") " pod="openstack/rabbitmq-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.149503 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1d7961e1-6f96-4707-907d-1fa0c30641e2-server-conf\") pod \"rabbitmq-server-0\" (UID: \"1d7961e1-6f96-4707-907d-1fa0c30641e2\") " pod="openstack/rabbitmq-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.151319 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.152616 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1d7961e1-6f96-4707-907d-1fa0c30641e2-config-data\") pod \"rabbitmq-server-0\" (UID: \"1d7961e1-6f96-4707-907d-1fa0c30641e2\") " pod="openstack/rabbitmq-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.153252 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1d7961e1-6f96-4707-907d-1fa0c30641e2-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"1d7961e1-6f96-4707-907d-1fa0c30641e2\") " pod="openstack/rabbitmq-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.153772 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1d7961e1-6f96-4707-907d-1fa0c30641e2-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"1d7961e1-6f96-4707-907d-1fa0c30641e2\") " pod="openstack/rabbitmq-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.157474 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6n8rf\" (UniqueName: \"kubernetes.io/projected/1d7961e1-6f96-4707-907d-1fa0c30641e2-kube-api-access-6n8rf\") pod \"rabbitmq-server-0\" (UID: \"1d7961e1-6f96-4707-907d-1fa0c30641e2\") " pod="openstack/rabbitmq-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.165553 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1d7961e1-6f96-4707-907d-1fa0c30641e2-pod-info\") pod \"rabbitmq-server-0\" (UID: \"1d7961e1-6f96-4707-907d-1fa0c30641e2\") " pod="openstack/rabbitmq-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.169192 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1d7961e1-6f96-4707-907d-1fa0c30641e2-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"1d7961e1-6f96-4707-907d-1fa0c30641e2\") " pod="openstack/rabbitmq-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.197191 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"1d7961e1-6f96-4707-907d-1fa0c30641e2\") " pod="openstack/rabbitmq-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.228994 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67ff45466c-wdkxp" event={"ID":"fa586b98-2a59-4673-a0bf-f7175f30d356","Type":"ContainerStarted","Data":"031b6d98624d5f4f1ccbf8ef80d5c2c14a3af7a2b1d6eaa11caf0fb3954e993c"} Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.230584 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f54874ffc-rb6x9" event={"ID":"78221461-c2ed-45d9-bd36-aa078d36d757","Type":"ContainerStarted","Data":"5d38656375670fe555a05d1c74743fce293b4370622cd92c9cce18bd0ace0a44"} Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.244278 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/03023501-772f-4798-a953-f56dee586cb1-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"03023501-772f-4798-a953-f56dee586cb1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.244316 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/03023501-772f-4798-a953-f56dee586cb1-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"03023501-772f-4798-a953-f56dee586cb1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.244379 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/03023501-772f-4798-a953-f56dee586cb1-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"03023501-772f-4798-a953-f56dee586cb1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.244431 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"03023501-772f-4798-a953-f56dee586cb1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.244470 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/03023501-772f-4798-a953-f56dee586cb1-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"03023501-772f-4798-a953-f56dee586cb1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.244537 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/03023501-772f-4798-a953-f56dee586cb1-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"03023501-772f-4798-a953-f56dee586cb1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.244559 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/03023501-772f-4798-a953-f56dee586cb1-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"03023501-772f-4798-a953-f56dee586cb1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.244616 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/03023501-772f-4798-a953-f56dee586cb1-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"03023501-772f-4798-a953-f56dee586cb1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.244633 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4q52k\" (UniqueName: \"kubernetes.io/projected/03023501-772f-4798-a953-f56dee586cb1-kube-api-access-4q52k\") pod \"rabbitmq-cell1-server-0\" (UID: \"03023501-772f-4798-a953-f56dee586cb1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.244666 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/03023501-772f-4798-a953-f56dee586cb1-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"03023501-772f-4798-a953-f56dee586cb1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.244728 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/03023501-772f-4798-a953-f56dee586cb1-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"03023501-772f-4798-a953-f56dee586cb1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.346491 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/03023501-772f-4798-a953-f56dee586cb1-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"03023501-772f-4798-a953-f56dee586cb1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.346537 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/03023501-772f-4798-a953-f56dee586cb1-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"03023501-772f-4798-a953-f56dee586cb1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.346575 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/03023501-772f-4798-a953-f56dee586cb1-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"03023501-772f-4798-a953-f56dee586cb1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.346602 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4q52k\" (UniqueName: \"kubernetes.io/projected/03023501-772f-4798-a953-f56dee586cb1-kube-api-access-4q52k\") pod \"rabbitmq-cell1-server-0\" (UID: \"03023501-772f-4798-a953-f56dee586cb1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.346623 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/03023501-772f-4798-a953-f56dee586cb1-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"03023501-772f-4798-a953-f56dee586cb1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.346643 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/03023501-772f-4798-a953-f56dee586cb1-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"03023501-772f-4798-a953-f56dee586cb1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.346668 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/03023501-772f-4798-a953-f56dee586cb1-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"03023501-772f-4798-a953-f56dee586cb1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.346681 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/03023501-772f-4798-a953-f56dee586cb1-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"03023501-772f-4798-a953-f56dee586cb1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.346713 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/03023501-772f-4798-a953-f56dee586cb1-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"03023501-772f-4798-a953-f56dee586cb1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.346743 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"03023501-772f-4798-a953-f56dee586cb1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.346770 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/03023501-772f-4798-a953-f56dee586cb1-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"03023501-772f-4798-a953-f56dee586cb1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.348725 4556 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"03023501-772f-4798-a953-f56dee586cb1\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.349013 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/03023501-772f-4798-a953-f56dee586cb1-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"03023501-772f-4798-a953-f56dee586cb1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.349292 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/03023501-772f-4798-a953-f56dee586cb1-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"03023501-772f-4798-a953-f56dee586cb1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.348732 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/03023501-772f-4798-a953-f56dee586cb1-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"03023501-772f-4798-a953-f56dee586cb1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.350765 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/03023501-772f-4798-a953-f56dee586cb1-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"03023501-772f-4798-a953-f56dee586cb1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.351520 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/03023501-772f-4798-a953-f56dee586cb1-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"03023501-772f-4798-a953-f56dee586cb1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.354060 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/03023501-772f-4798-a953-f56dee586cb1-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"03023501-772f-4798-a953-f56dee586cb1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.354982 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/03023501-772f-4798-a953-f56dee586cb1-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"03023501-772f-4798-a953-f56dee586cb1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.357542 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/03023501-772f-4798-a953-f56dee586cb1-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"03023501-772f-4798-a953-f56dee586cb1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.359606 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/03023501-772f-4798-a953-f56dee586cb1-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"03023501-772f-4798-a953-f56dee586cb1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.365234 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4q52k\" (UniqueName: \"kubernetes.io/projected/03023501-772f-4798-a953-f56dee586cb1-kube-api-access-4q52k\") pod \"rabbitmq-cell1-server-0\" (UID: \"03023501-772f-4798-a953-f56dee586cb1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.365294 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"03023501-772f-4798-a953-f56dee586cb1\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.494409 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 18 09:16:22 crc kubenswrapper[4556]: I0218 09:16:22.531879 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:16:23 crc kubenswrapper[4556]: I0218 09:16:23.299903 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Feb 18 09:16:23 crc kubenswrapper[4556]: I0218 09:16:23.303583 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Feb 18 09:16:23 crc kubenswrapper[4556]: I0218 09:16:23.303707 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Feb 18 09:16:23 crc kubenswrapper[4556]: I0218 09:16:23.312362 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Feb 18 09:16:23 crc kubenswrapper[4556]: I0218 09:16:23.312547 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Feb 18 09:16:23 crc kubenswrapper[4556]: I0218 09:16:23.312858 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-fpf6d" Feb 18 09:16:23 crc kubenswrapper[4556]: I0218 09:16:23.312960 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Feb 18 09:16:23 crc kubenswrapper[4556]: I0218 09:16:23.316683 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Feb 18 09:16:23 crc kubenswrapper[4556]: I0218 09:16:23.482068 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2c504c68-45e0-49dc-90f3-90b91e33551c-operator-scripts\") pod \"openstack-galera-0\" (UID: \"2c504c68-45e0-49dc-90f3-90b91e33551c\") " pod="openstack/openstack-galera-0" Feb 18 09:16:23 crc kubenswrapper[4556]: I0218 09:16:23.482167 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/2c504c68-45e0-49dc-90f3-90b91e33551c-config-data-default\") pod \"openstack-galera-0\" (UID: \"2c504c68-45e0-49dc-90f3-90b91e33551c\") " pod="openstack/openstack-galera-0" Feb 18 09:16:23 crc kubenswrapper[4556]: I0218 09:16:23.482190 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fj79r\" (UniqueName: \"kubernetes.io/projected/2c504c68-45e0-49dc-90f3-90b91e33551c-kube-api-access-fj79r\") pod \"openstack-galera-0\" (UID: \"2c504c68-45e0-49dc-90f3-90b91e33551c\") " pod="openstack/openstack-galera-0" Feb 18 09:16:23 crc kubenswrapper[4556]: I0218 09:16:23.482254 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c504c68-45e0-49dc-90f3-90b91e33551c-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"2c504c68-45e0-49dc-90f3-90b91e33551c\") " pod="openstack/openstack-galera-0" Feb 18 09:16:23 crc kubenswrapper[4556]: I0218 09:16:23.482317 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2c504c68-45e0-49dc-90f3-90b91e33551c-kolla-config\") pod \"openstack-galera-0\" (UID: \"2c504c68-45e0-49dc-90f3-90b91e33551c\") " pod="openstack/openstack-galera-0" Feb 18 09:16:23 crc kubenswrapper[4556]: I0218 09:16:23.482342 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c504c68-45e0-49dc-90f3-90b91e33551c-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"2c504c68-45e0-49dc-90f3-90b91e33551c\") " pod="openstack/openstack-galera-0" Feb 18 09:16:23 crc kubenswrapper[4556]: I0218 09:16:23.482391 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/2c504c68-45e0-49dc-90f3-90b91e33551c-config-data-generated\") pod \"openstack-galera-0\" (UID: \"2c504c68-45e0-49dc-90f3-90b91e33551c\") " pod="openstack/openstack-galera-0" Feb 18 09:16:23 crc kubenswrapper[4556]: I0218 09:16:23.482411 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-0\" (UID: \"2c504c68-45e0-49dc-90f3-90b91e33551c\") " pod="openstack/openstack-galera-0" Feb 18 09:16:23 crc kubenswrapper[4556]: I0218 09:16:23.585583 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/2c504c68-45e0-49dc-90f3-90b91e33551c-config-data-generated\") pod \"openstack-galera-0\" (UID: \"2c504c68-45e0-49dc-90f3-90b91e33551c\") " pod="openstack/openstack-galera-0" Feb 18 09:16:23 crc kubenswrapper[4556]: I0218 09:16:23.585647 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-0\" (UID: \"2c504c68-45e0-49dc-90f3-90b91e33551c\") " pod="openstack/openstack-galera-0" Feb 18 09:16:23 crc kubenswrapper[4556]: I0218 09:16:23.585801 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2c504c68-45e0-49dc-90f3-90b91e33551c-operator-scripts\") pod \"openstack-galera-0\" (UID: \"2c504c68-45e0-49dc-90f3-90b91e33551c\") " pod="openstack/openstack-galera-0" Feb 18 09:16:23 crc kubenswrapper[4556]: I0218 09:16:23.585847 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/2c504c68-45e0-49dc-90f3-90b91e33551c-config-data-default\") pod \"openstack-galera-0\" (UID: \"2c504c68-45e0-49dc-90f3-90b91e33551c\") " pod="openstack/openstack-galera-0" Feb 18 09:16:23 crc kubenswrapper[4556]: I0218 09:16:23.585872 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fj79r\" (UniqueName: \"kubernetes.io/projected/2c504c68-45e0-49dc-90f3-90b91e33551c-kube-api-access-fj79r\") pod \"openstack-galera-0\" (UID: \"2c504c68-45e0-49dc-90f3-90b91e33551c\") " pod="openstack/openstack-galera-0" Feb 18 09:16:23 crc kubenswrapper[4556]: I0218 09:16:23.585957 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/2c504c68-45e0-49dc-90f3-90b91e33551c-config-data-generated\") pod \"openstack-galera-0\" (UID: \"2c504c68-45e0-49dc-90f3-90b91e33551c\") " pod="openstack/openstack-galera-0" Feb 18 09:16:23 crc kubenswrapper[4556]: I0218 09:16:23.586103 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c504c68-45e0-49dc-90f3-90b91e33551c-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"2c504c68-45e0-49dc-90f3-90b91e33551c\") " pod="openstack/openstack-galera-0" Feb 18 09:16:23 crc kubenswrapper[4556]: I0218 09:16:23.586196 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2c504c68-45e0-49dc-90f3-90b91e33551c-kolla-config\") pod \"openstack-galera-0\" (UID: \"2c504c68-45e0-49dc-90f3-90b91e33551c\") " pod="openstack/openstack-galera-0" Feb 18 09:16:23 crc kubenswrapper[4556]: I0218 09:16:23.586255 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c504c68-45e0-49dc-90f3-90b91e33551c-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"2c504c68-45e0-49dc-90f3-90b91e33551c\") " pod="openstack/openstack-galera-0" Feb 18 09:16:23 crc kubenswrapper[4556]: I0218 09:16:23.586089 4556 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-0\" (UID: \"2c504c68-45e0-49dc-90f3-90b91e33551c\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/openstack-galera-0" Feb 18 09:16:23 crc kubenswrapper[4556]: I0218 09:16:23.586901 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2c504c68-45e0-49dc-90f3-90b91e33551c-kolla-config\") pod \"openstack-galera-0\" (UID: \"2c504c68-45e0-49dc-90f3-90b91e33551c\") " pod="openstack/openstack-galera-0" Feb 18 09:16:23 crc kubenswrapper[4556]: I0218 09:16:23.586985 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/2c504c68-45e0-49dc-90f3-90b91e33551c-config-data-default\") pod \"openstack-galera-0\" (UID: \"2c504c68-45e0-49dc-90f3-90b91e33551c\") " pod="openstack/openstack-galera-0" Feb 18 09:16:23 crc kubenswrapper[4556]: I0218 09:16:23.588050 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2c504c68-45e0-49dc-90f3-90b91e33551c-operator-scripts\") pod \"openstack-galera-0\" (UID: \"2c504c68-45e0-49dc-90f3-90b91e33551c\") " pod="openstack/openstack-galera-0" Feb 18 09:16:23 crc kubenswrapper[4556]: I0218 09:16:23.598693 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c504c68-45e0-49dc-90f3-90b91e33551c-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"2c504c68-45e0-49dc-90f3-90b91e33551c\") " pod="openstack/openstack-galera-0" Feb 18 09:16:23 crc kubenswrapper[4556]: I0218 09:16:23.601808 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fj79r\" (UniqueName: \"kubernetes.io/projected/2c504c68-45e0-49dc-90f3-90b91e33551c-kube-api-access-fj79r\") pod \"openstack-galera-0\" (UID: \"2c504c68-45e0-49dc-90f3-90b91e33551c\") " pod="openstack/openstack-galera-0" Feb 18 09:16:23 crc kubenswrapper[4556]: I0218 09:16:23.610680 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c504c68-45e0-49dc-90f3-90b91e33551c-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"2c504c68-45e0-49dc-90f3-90b91e33551c\") " pod="openstack/openstack-galera-0" Feb 18 09:16:23 crc kubenswrapper[4556]: I0218 09:16:23.617436 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-0\" (UID: \"2c504c68-45e0-49dc-90f3-90b91e33551c\") " pod="openstack/openstack-galera-0" Feb 18 09:16:23 crc kubenswrapper[4556]: I0218 09:16:23.626613 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Feb 18 09:16:24 crc kubenswrapper[4556]: I0218 09:16:24.614027 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 18 09:16:24 crc kubenswrapper[4556]: I0218 09:16:24.615502 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Feb 18 09:16:24 crc kubenswrapper[4556]: I0218 09:16:24.620820 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Feb 18 09:16:24 crc kubenswrapper[4556]: I0218 09:16:24.620904 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-7r5ng" Feb 18 09:16:24 crc kubenswrapper[4556]: I0218 09:16:24.621134 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Feb 18 09:16:24 crc kubenswrapper[4556]: I0218 09:16:24.621688 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Feb 18 09:16:24 crc kubenswrapper[4556]: I0218 09:16:24.639219 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 18 09:16:24 crc kubenswrapper[4556]: I0218 09:16:24.802782 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/11b77df6-67d9-4ddb-944b-2ad8b0fada78-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"11b77df6-67d9-4ddb-944b-2ad8b0fada78\") " pod="openstack/openstack-cell1-galera-0" Feb 18 09:16:24 crc kubenswrapper[4556]: I0218 09:16:24.802835 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"11b77df6-67d9-4ddb-944b-2ad8b0fada78\") " pod="openstack/openstack-cell1-galera-0" Feb 18 09:16:24 crc kubenswrapper[4556]: I0218 09:16:24.803491 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/11b77df6-67d9-4ddb-944b-2ad8b0fada78-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"11b77df6-67d9-4ddb-944b-2ad8b0fada78\") " pod="openstack/openstack-cell1-galera-0" Feb 18 09:16:24 crc kubenswrapper[4556]: I0218 09:16:24.803564 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/11b77df6-67d9-4ddb-944b-2ad8b0fada78-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"11b77df6-67d9-4ddb-944b-2ad8b0fada78\") " pod="openstack/openstack-cell1-galera-0" Feb 18 09:16:24 crc kubenswrapper[4556]: I0218 09:16:24.803587 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11b77df6-67d9-4ddb-944b-2ad8b0fada78-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"11b77df6-67d9-4ddb-944b-2ad8b0fada78\") " pod="openstack/openstack-cell1-galera-0" Feb 18 09:16:24 crc kubenswrapper[4556]: I0218 09:16:24.803641 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/11b77df6-67d9-4ddb-944b-2ad8b0fada78-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"11b77df6-67d9-4ddb-944b-2ad8b0fada78\") " pod="openstack/openstack-cell1-galera-0" Feb 18 09:16:24 crc kubenswrapper[4556]: I0218 09:16:24.803667 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/11b77df6-67d9-4ddb-944b-2ad8b0fada78-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"11b77df6-67d9-4ddb-944b-2ad8b0fada78\") " pod="openstack/openstack-cell1-galera-0" Feb 18 09:16:24 crc kubenswrapper[4556]: I0218 09:16:24.803687 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtgw8\" (UniqueName: \"kubernetes.io/projected/11b77df6-67d9-4ddb-944b-2ad8b0fada78-kube-api-access-rtgw8\") pod \"openstack-cell1-galera-0\" (UID: \"11b77df6-67d9-4ddb-944b-2ad8b0fada78\") " pod="openstack/openstack-cell1-galera-0" Feb 18 09:16:24 crc kubenswrapper[4556]: I0218 09:16:24.825221 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Feb 18 09:16:24 crc kubenswrapper[4556]: I0218 09:16:24.825978 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Feb 18 09:16:24 crc kubenswrapper[4556]: I0218 09:16:24.827755 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-sqwc8" Feb 18 09:16:24 crc kubenswrapper[4556]: I0218 09:16:24.829369 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Feb 18 09:16:24 crc kubenswrapper[4556]: I0218 09:16:24.829468 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Feb 18 09:16:24 crc kubenswrapper[4556]: I0218 09:16:24.831274 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Feb 18 09:16:24 crc kubenswrapper[4556]: I0218 09:16:24.906923 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11b77df6-67d9-4ddb-944b-2ad8b0fada78-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"11b77df6-67d9-4ddb-944b-2ad8b0fada78\") " pod="openstack/openstack-cell1-galera-0" Feb 18 09:16:24 crc kubenswrapper[4556]: I0218 09:16:24.906980 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/11b77df6-67d9-4ddb-944b-2ad8b0fada78-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"11b77df6-67d9-4ddb-944b-2ad8b0fada78\") " pod="openstack/openstack-cell1-galera-0" Feb 18 09:16:24 crc kubenswrapper[4556]: I0218 09:16:24.907021 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/11b77df6-67d9-4ddb-944b-2ad8b0fada78-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"11b77df6-67d9-4ddb-944b-2ad8b0fada78\") " pod="openstack/openstack-cell1-galera-0" Feb 18 09:16:24 crc kubenswrapper[4556]: I0218 09:16:24.907056 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtgw8\" (UniqueName: \"kubernetes.io/projected/11b77df6-67d9-4ddb-944b-2ad8b0fada78-kube-api-access-rtgw8\") pod \"openstack-cell1-galera-0\" (UID: \"11b77df6-67d9-4ddb-944b-2ad8b0fada78\") " pod="openstack/openstack-cell1-galera-0" Feb 18 09:16:24 crc kubenswrapper[4556]: I0218 09:16:24.907099 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/11b77df6-67d9-4ddb-944b-2ad8b0fada78-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"11b77df6-67d9-4ddb-944b-2ad8b0fada78\") " pod="openstack/openstack-cell1-galera-0" Feb 18 09:16:24 crc kubenswrapper[4556]: I0218 09:16:24.907143 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"11b77df6-67d9-4ddb-944b-2ad8b0fada78\") " pod="openstack/openstack-cell1-galera-0" Feb 18 09:16:24 crc kubenswrapper[4556]: I0218 09:16:24.907211 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/11b77df6-67d9-4ddb-944b-2ad8b0fada78-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"11b77df6-67d9-4ddb-944b-2ad8b0fada78\") " pod="openstack/openstack-cell1-galera-0" Feb 18 09:16:24 crc kubenswrapper[4556]: I0218 09:16:24.907275 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/11b77df6-67d9-4ddb-944b-2ad8b0fada78-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"11b77df6-67d9-4ddb-944b-2ad8b0fada78\") " pod="openstack/openstack-cell1-galera-0" Feb 18 09:16:24 crc kubenswrapper[4556]: I0218 09:16:24.907588 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/11b77df6-67d9-4ddb-944b-2ad8b0fada78-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"11b77df6-67d9-4ddb-944b-2ad8b0fada78\") " pod="openstack/openstack-cell1-galera-0" Feb 18 09:16:24 crc kubenswrapper[4556]: I0218 09:16:24.907764 4556 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"11b77df6-67d9-4ddb-944b-2ad8b0fada78\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/openstack-cell1-galera-0" Feb 18 09:16:24 crc kubenswrapper[4556]: I0218 09:16:24.908107 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/11b77df6-67d9-4ddb-944b-2ad8b0fada78-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"11b77df6-67d9-4ddb-944b-2ad8b0fada78\") " pod="openstack/openstack-cell1-galera-0" Feb 18 09:16:24 crc kubenswrapper[4556]: I0218 09:16:24.914729 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/11b77df6-67d9-4ddb-944b-2ad8b0fada78-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"11b77df6-67d9-4ddb-944b-2ad8b0fada78\") " pod="openstack/openstack-cell1-galera-0" Feb 18 09:16:24 crc kubenswrapper[4556]: I0218 09:16:24.916177 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/11b77df6-67d9-4ddb-944b-2ad8b0fada78-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"11b77df6-67d9-4ddb-944b-2ad8b0fada78\") " pod="openstack/openstack-cell1-galera-0" Feb 18 09:16:24 crc kubenswrapper[4556]: I0218 09:16:24.920848 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11b77df6-67d9-4ddb-944b-2ad8b0fada78-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"11b77df6-67d9-4ddb-944b-2ad8b0fada78\") " pod="openstack/openstack-cell1-galera-0" Feb 18 09:16:24 crc kubenswrapper[4556]: I0218 09:16:24.924812 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/11b77df6-67d9-4ddb-944b-2ad8b0fada78-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"11b77df6-67d9-4ddb-944b-2ad8b0fada78\") " pod="openstack/openstack-cell1-galera-0" Feb 18 09:16:24 crc kubenswrapper[4556]: I0218 09:16:24.929786 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtgw8\" (UniqueName: \"kubernetes.io/projected/11b77df6-67d9-4ddb-944b-2ad8b0fada78-kube-api-access-rtgw8\") pod \"openstack-cell1-galera-0\" (UID: \"11b77df6-67d9-4ddb-944b-2ad8b0fada78\") " pod="openstack/openstack-cell1-galera-0" Feb 18 09:16:24 crc kubenswrapper[4556]: I0218 09:16:24.944308 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"11b77df6-67d9-4ddb-944b-2ad8b0fada78\") " pod="openstack/openstack-cell1-galera-0" Feb 18 09:16:25 crc kubenswrapper[4556]: I0218 09:16:25.011305 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/68274946-b189-450c-a154-27059f411af3-config-data\") pod \"memcached-0\" (UID: \"68274946-b189-450c-a154-27059f411af3\") " pod="openstack/memcached-0" Feb 18 09:16:25 crc kubenswrapper[4556]: I0218 09:16:25.011651 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpvdj\" (UniqueName: \"kubernetes.io/projected/68274946-b189-450c-a154-27059f411af3-kube-api-access-dpvdj\") pod \"memcached-0\" (UID: \"68274946-b189-450c-a154-27059f411af3\") " pod="openstack/memcached-0" Feb 18 09:16:25 crc kubenswrapper[4556]: I0218 09:16:25.011827 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68274946-b189-450c-a154-27059f411af3-combined-ca-bundle\") pod \"memcached-0\" (UID: \"68274946-b189-450c-a154-27059f411af3\") " pod="openstack/memcached-0" Feb 18 09:16:25 crc kubenswrapper[4556]: I0218 09:16:25.011917 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/68274946-b189-450c-a154-27059f411af3-kolla-config\") pod \"memcached-0\" (UID: \"68274946-b189-450c-a154-27059f411af3\") " pod="openstack/memcached-0" Feb 18 09:16:25 crc kubenswrapper[4556]: I0218 09:16:25.012051 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/68274946-b189-450c-a154-27059f411af3-memcached-tls-certs\") pod \"memcached-0\" (UID: \"68274946-b189-450c-a154-27059f411af3\") " pod="openstack/memcached-0" Feb 18 09:16:25 crc kubenswrapper[4556]: I0218 09:16:25.114336 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/68274946-b189-450c-a154-27059f411af3-memcached-tls-certs\") pod \"memcached-0\" (UID: \"68274946-b189-450c-a154-27059f411af3\") " pod="openstack/memcached-0" Feb 18 09:16:25 crc kubenswrapper[4556]: I0218 09:16:25.114419 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/68274946-b189-450c-a154-27059f411af3-config-data\") pod \"memcached-0\" (UID: \"68274946-b189-450c-a154-27059f411af3\") " pod="openstack/memcached-0" Feb 18 09:16:25 crc kubenswrapper[4556]: I0218 09:16:25.114559 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpvdj\" (UniqueName: \"kubernetes.io/projected/68274946-b189-450c-a154-27059f411af3-kube-api-access-dpvdj\") pod \"memcached-0\" (UID: \"68274946-b189-450c-a154-27059f411af3\") " pod="openstack/memcached-0" Feb 18 09:16:25 crc kubenswrapper[4556]: I0218 09:16:25.114624 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68274946-b189-450c-a154-27059f411af3-combined-ca-bundle\") pod \"memcached-0\" (UID: \"68274946-b189-450c-a154-27059f411af3\") " pod="openstack/memcached-0" Feb 18 09:16:25 crc kubenswrapper[4556]: I0218 09:16:25.114659 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/68274946-b189-450c-a154-27059f411af3-kolla-config\") pod \"memcached-0\" (UID: \"68274946-b189-450c-a154-27059f411af3\") " pod="openstack/memcached-0" Feb 18 09:16:25 crc kubenswrapper[4556]: I0218 09:16:25.115421 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/68274946-b189-450c-a154-27059f411af3-kolla-config\") pod \"memcached-0\" (UID: \"68274946-b189-450c-a154-27059f411af3\") " pod="openstack/memcached-0" Feb 18 09:16:25 crc kubenswrapper[4556]: I0218 09:16:25.116850 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/68274946-b189-450c-a154-27059f411af3-config-data\") pod \"memcached-0\" (UID: \"68274946-b189-450c-a154-27059f411af3\") " pod="openstack/memcached-0" Feb 18 09:16:25 crc kubenswrapper[4556]: I0218 09:16:25.118031 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/68274946-b189-450c-a154-27059f411af3-memcached-tls-certs\") pod \"memcached-0\" (UID: \"68274946-b189-450c-a154-27059f411af3\") " pod="openstack/memcached-0" Feb 18 09:16:25 crc kubenswrapper[4556]: I0218 09:16:25.119969 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68274946-b189-450c-a154-27059f411af3-combined-ca-bundle\") pod \"memcached-0\" (UID: \"68274946-b189-450c-a154-27059f411af3\") " pod="openstack/memcached-0" Feb 18 09:16:25 crc kubenswrapper[4556]: I0218 09:16:25.130721 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpvdj\" (UniqueName: \"kubernetes.io/projected/68274946-b189-450c-a154-27059f411af3-kube-api-access-dpvdj\") pod \"memcached-0\" (UID: \"68274946-b189-450c-a154-27059f411af3\") " pod="openstack/memcached-0" Feb 18 09:16:25 crc kubenswrapper[4556]: I0218 09:16:25.148336 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Feb 18 09:16:25 crc kubenswrapper[4556]: I0218 09:16:25.243677 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Feb 18 09:16:27 crc kubenswrapper[4556]: I0218 09:16:27.171119 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Feb 18 09:16:27 crc kubenswrapper[4556]: I0218 09:16:27.172233 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 18 09:16:27 crc kubenswrapper[4556]: I0218 09:16:27.173984 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-8pklh" Feb 18 09:16:27 crc kubenswrapper[4556]: I0218 09:16:27.177530 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 18 09:16:27 crc kubenswrapper[4556]: I0218 09:16:27.351252 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrmbg\" (UniqueName: \"kubernetes.io/projected/95f33d13-a69d-4add-88a3-21126d52a0ed-kube-api-access-lrmbg\") pod \"kube-state-metrics-0\" (UID: \"95f33d13-a69d-4add-88a3-21126d52a0ed\") " pod="openstack/kube-state-metrics-0" Feb 18 09:16:27 crc kubenswrapper[4556]: I0218 09:16:27.452970 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrmbg\" (UniqueName: \"kubernetes.io/projected/95f33d13-a69d-4add-88a3-21126d52a0ed-kube-api-access-lrmbg\") pod \"kube-state-metrics-0\" (UID: \"95f33d13-a69d-4add-88a3-21126d52a0ed\") " pod="openstack/kube-state-metrics-0" Feb 18 09:16:27 crc kubenswrapper[4556]: I0218 09:16:27.469638 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrmbg\" (UniqueName: \"kubernetes.io/projected/95f33d13-a69d-4add-88a3-21126d52a0ed-kube-api-access-lrmbg\") pod \"kube-state-metrics-0\" (UID: \"95f33d13-a69d-4add-88a3-21126d52a0ed\") " pod="openstack/kube-state-metrics-0" Feb 18 09:16:27 crc kubenswrapper[4556]: I0218 09:16:27.489569 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 18 09:16:30 crc kubenswrapper[4556]: I0218 09:16:30.271660 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-6kkrw"] Feb 18 09:16:30 crc kubenswrapper[4556]: I0218 09:16:30.277594 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-srnvn"] Feb 18 09:16:30 crc kubenswrapper[4556]: I0218 09:16:30.277716 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-6kkrw" Feb 18 09:16:30 crc kubenswrapper[4556]: I0218 09:16:30.282507 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-fsbwt" Feb 18 09:16:30 crc kubenswrapper[4556]: I0218 09:16:30.282734 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Feb 18 09:16:30 crc kubenswrapper[4556]: I0218 09:16:30.282964 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Feb 18 09:16:30 crc kubenswrapper[4556]: I0218 09:16:30.285187 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-6kkrw"] Feb 18 09:16:30 crc kubenswrapper[4556]: I0218 09:16:30.285598 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-srnvn" Feb 18 09:16:30 crc kubenswrapper[4556]: I0218 09:16:30.301175 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-srnvn"] Feb 18 09:16:30 crc kubenswrapper[4556]: I0218 09:16:30.316144 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/40c51577-de87-4bd8-be03-41221bc2f415-scripts\") pod \"ovn-controller-ovs-srnvn\" (UID: \"40c51577-de87-4bd8-be03-41221bc2f415\") " pod="openstack/ovn-controller-ovs-srnvn" Feb 18 09:16:30 crc kubenswrapper[4556]: I0218 09:16:30.316289 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/40c51577-de87-4bd8-be03-41221bc2f415-var-log\") pod \"ovn-controller-ovs-srnvn\" (UID: \"40c51577-de87-4bd8-be03-41221bc2f415\") " pod="openstack/ovn-controller-ovs-srnvn" Feb 18 09:16:30 crc kubenswrapper[4556]: I0218 09:16:30.316332 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/40c51577-de87-4bd8-be03-41221bc2f415-etc-ovs\") pod \"ovn-controller-ovs-srnvn\" (UID: \"40c51577-de87-4bd8-be03-41221bc2f415\") " pod="openstack/ovn-controller-ovs-srnvn" Feb 18 09:16:30 crc kubenswrapper[4556]: I0218 09:16:30.316402 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5f55s\" (UniqueName: \"kubernetes.io/projected/40c51577-de87-4bd8-be03-41221bc2f415-kube-api-access-5f55s\") pod \"ovn-controller-ovs-srnvn\" (UID: \"40c51577-de87-4bd8-be03-41221bc2f415\") " pod="openstack/ovn-controller-ovs-srnvn" Feb 18 09:16:30 crc kubenswrapper[4556]: I0218 09:16:30.316472 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/40c51577-de87-4bd8-be03-41221bc2f415-var-run\") pod \"ovn-controller-ovs-srnvn\" (UID: \"40c51577-de87-4bd8-be03-41221bc2f415\") " pod="openstack/ovn-controller-ovs-srnvn" Feb 18 09:16:30 crc kubenswrapper[4556]: I0218 09:16:30.316527 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/40c51577-de87-4bd8-be03-41221bc2f415-var-lib\") pod \"ovn-controller-ovs-srnvn\" (UID: \"40c51577-de87-4bd8-be03-41221bc2f415\") " pod="openstack/ovn-controller-ovs-srnvn" Feb 18 09:16:30 crc kubenswrapper[4556]: I0218 09:16:30.418331 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/40c51577-de87-4bd8-be03-41221bc2f415-var-log\") pod \"ovn-controller-ovs-srnvn\" (UID: \"40c51577-de87-4bd8-be03-41221bc2f415\") " pod="openstack/ovn-controller-ovs-srnvn" Feb 18 09:16:30 crc kubenswrapper[4556]: I0218 09:16:30.418701 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4906afef-c7e6-4597-a4a5-e9d758917e11-var-run-ovn\") pod \"ovn-controller-6kkrw\" (UID: \"4906afef-c7e6-4597-a4a5-e9d758917e11\") " pod="openstack/ovn-controller-6kkrw" Feb 18 09:16:30 crc kubenswrapper[4556]: I0218 09:16:30.418733 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5plct\" (UniqueName: \"kubernetes.io/projected/4906afef-c7e6-4597-a4a5-e9d758917e11-kube-api-access-5plct\") pod \"ovn-controller-6kkrw\" (UID: \"4906afef-c7e6-4597-a4a5-e9d758917e11\") " pod="openstack/ovn-controller-6kkrw" Feb 18 09:16:30 crc kubenswrapper[4556]: I0218 09:16:30.418764 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/40c51577-de87-4bd8-be03-41221bc2f415-etc-ovs\") pod \"ovn-controller-ovs-srnvn\" (UID: \"40c51577-de87-4bd8-be03-41221bc2f415\") " pod="openstack/ovn-controller-ovs-srnvn" Feb 18 09:16:30 crc kubenswrapper[4556]: I0218 09:16:30.418794 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4906afef-c7e6-4597-a4a5-e9d758917e11-combined-ca-bundle\") pod \"ovn-controller-6kkrw\" (UID: \"4906afef-c7e6-4597-a4a5-e9d758917e11\") " pod="openstack/ovn-controller-6kkrw" Feb 18 09:16:30 crc kubenswrapper[4556]: I0218 09:16:30.418818 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4906afef-c7e6-4597-a4a5-e9d758917e11-var-run\") pod \"ovn-controller-6kkrw\" (UID: \"4906afef-c7e6-4597-a4a5-e9d758917e11\") " pod="openstack/ovn-controller-6kkrw" Feb 18 09:16:30 crc kubenswrapper[4556]: I0218 09:16:30.418870 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/4906afef-c7e6-4597-a4a5-e9d758917e11-ovn-controller-tls-certs\") pod \"ovn-controller-6kkrw\" (UID: \"4906afef-c7e6-4597-a4a5-e9d758917e11\") " pod="openstack/ovn-controller-6kkrw" Feb 18 09:16:30 crc kubenswrapper[4556]: I0218 09:16:30.418885 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4906afef-c7e6-4597-a4a5-e9d758917e11-scripts\") pod \"ovn-controller-6kkrw\" (UID: \"4906afef-c7e6-4597-a4a5-e9d758917e11\") " pod="openstack/ovn-controller-6kkrw" Feb 18 09:16:30 crc kubenswrapper[4556]: I0218 09:16:30.418916 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5f55s\" (UniqueName: \"kubernetes.io/projected/40c51577-de87-4bd8-be03-41221bc2f415-kube-api-access-5f55s\") pod \"ovn-controller-ovs-srnvn\" (UID: \"40c51577-de87-4bd8-be03-41221bc2f415\") " pod="openstack/ovn-controller-ovs-srnvn" Feb 18 09:16:30 crc kubenswrapper[4556]: I0218 09:16:30.418985 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/40c51577-de87-4bd8-be03-41221bc2f415-var-run\") pod \"ovn-controller-ovs-srnvn\" (UID: \"40c51577-de87-4bd8-be03-41221bc2f415\") " pod="openstack/ovn-controller-ovs-srnvn" Feb 18 09:16:30 crc kubenswrapper[4556]: I0218 09:16:30.419054 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/40c51577-de87-4bd8-be03-41221bc2f415-var-lib\") pod \"ovn-controller-ovs-srnvn\" (UID: \"40c51577-de87-4bd8-be03-41221bc2f415\") " pod="openstack/ovn-controller-ovs-srnvn" Feb 18 09:16:30 crc kubenswrapper[4556]: I0218 09:16:30.419104 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/40c51577-de87-4bd8-be03-41221bc2f415-scripts\") pod \"ovn-controller-ovs-srnvn\" (UID: \"40c51577-de87-4bd8-be03-41221bc2f415\") " pod="openstack/ovn-controller-ovs-srnvn" Feb 18 09:16:30 crc kubenswrapper[4556]: I0218 09:16:30.419134 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4906afef-c7e6-4597-a4a5-e9d758917e11-var-log-ovn\") pod \"ovn-controller-6kkrw\" (UID: \"4906afef-c7e6-4597-a4a5-e9d758917e11\") " pod="openstack/ovn-controller-6kkrw" Feb 18 09:16:30 crc kubenswrapper[4556]: I0218 09:16:30.420671 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/40c51577-de87-4bd8-be03-41221bc2f415-var-log\") pod \"ovn-controller-ovs-srnvn\" (UID: \"40c51577-de87-4bd8-be03-41221bc2f415\") " pod="openstack/ovn-controller-ovs-srnvn" Feb 18 09:16:30 crc kubenswrapper[4556]: I0218 09:16:30.420698 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/40c51577-de87-4bd8-be03-41221bc2f415-etc-ovs\") pod \"ovn-controller-ovs-srnvn\" (UID: \"40c51577-de87-4bd8-be03-41221bc2f415\") " pod="openstack/ovn-controller-ovs-srnvn" Feb 18 09:16:30 crc kubenswrapper[4556]: I0218 09:16:30.420741 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/40c51577-de87-4bd8-be03-41221bc2f415-var-lib\") pod \"ovn-controller-ovs-srnvn\" (UID: \"40c51577-de87-4bd8-be03-41221bc2f415\") " pod="openstack/ovn-controller-ovs-srnvn" Feb 18 09:16:30 crc kubenswrapper[4556]: I0218 09:16:30.420821 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/40c51577-de87-4bd8-be03-41221bc2f415-var-run\") pod \"ovn-controller-ovs-srnvn\" (UID: \"40c51577-de87-4bd8-be03-41221bc2f415\") " pod="openstack/ovn-controller-ovs-srnvn" Feb 18 09:16:30 crc kubenswrapper[4556]: I0218 09:16:30.423202 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/40c51577-de87-4bd8-be03-41221bc2f415-scripts\") pod \"ovn-controller-ovs-srnvn\" (UID: \"40c51577-de87-4bd8-be03-41221bc2f415\") " pod="openstack/ovn-controller-ovs-srnvn" Feb 18 09:16:30 crc kubenswrapper[4556]: I0218 09:16:30.442524 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5f55s\" (UniqueName: \"kubernetes.io/projected/40c51577-de87-4bd8-be03-41221bc2f415-kube-api-access-5f55s\") pod \"ovn-controller-ovs-srnvn\" (UID: \"40c51577-de87-4bd8-be03-41221bc2f415\") " pod="openstack/ovn-controller-ovs-srnvn" Feb 18 09:16:30 crc kubenswrapper[4556]: I0218 09:16:30.520690 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4906afef-c7e6-4597-a4a5-e9d758917e11-var-run-ovn\") pod \"ovn-controller-6kkrw\" (UID: \"4906afef-c7e6-4597-a4a5-e9d758917e11\") " pod="openstack/ovn-controller-6kkrw" Feb 18 09:16:30 crc kubenswrapper[4556]: I0218 09:16:30.520736 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5plct\" (UniqueName: \"kubernetes.io/projected/4906afef-c7e6-4597-a4a5-e9d758917e11-kube-api-access-5plct\") pod \"ovn-controller-6kkrw\" (UID: \"4906afef-c7e6-4597-a4a5-e9d758917e11\") " pod="openstack/ovn-controller-6kkrw" Feb 18 09:16:30 crc kubenswrapper[4556]: I0218 09:16:30.520784 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4906afef-c7e6-4597-a4a5-e9d758917e11-combined-ca-bundle\") pod \"ovn-controller-6kkrw\" (UID: \"4906afef-c7e6-4597-a4a5-e9d758917e11\") " pod="openstack/ovn-controller-6kkrw" Feb 18 09:16:30 crc kubenswrapper[4556]: I0218 09:16:30.520808 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4906afef-c7e6-4597-a4a5-e9d758917e11-var-run\") pod \"ovn-controller-6kkrw\" (UID: \"4906afef-c7e6-4597-a4a5-e9d758917e11\") " pod="openstack/ovn-controller-6kkrw" Feb 18 09:16:30 crc kubenswrapper[4556]: I0218 09:16:30.520853 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/4906afef-c7e6-4597-a4a5-e9d758917e11-ovn-controller-tls-certs\") pod \"ovn-controller-6kkrw\" (UID: \"4906afef-c7e6-4597-a4a5-e9d758917e11\") " pod="openstack/ovn-controller-6kkrw" Feb 18 09:16:30 crc kubenswrapper[4556]: I0218 09:16:30.520874 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4906afef-c7e6-4597-a4a5-e9d758917e11-scripts\") pod \"ovn-controller-6kkrw\" (UID: \"4906afef-c7e6-4597-a4a5-e9d758917e11\") " pod="openstack/ovn-controller-6kkrw" Feb 18 09:16:30 crc kubenswrapper[4556]: I0218 09:16:30.520948 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4906afef-c7e6-4597-a4a5-e9d758917e11-var-run-ovn\") pod \"ovn-controller-6kkrw\" (UID: \"4906afef-c7e6-4597-a4a5-e9d758917e11\") " pod="openstack/ovn-controller-6kkrw" Feb 18 09:16:30 crc kubenswrapper[4556]: I0218 09:16:30.520948 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4906afef-c7e6-4597-a4a5-e9d758917e11-var-run\") pod \"ovn-controller-6kkrw\" (UID: \"4906afef-c7e6-4597-a4a5-e9d758917e11\") " pod="openstack/ovn-controller-6kkrw" Feb 18 09:16:30 crc kubenswrapper[4556]: I0218 09:16:30.521982 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4906afef-c7e6-4597-a4a5-e9d758917e11-var-log-ovn\") pod \"ovn-controller-6kkrw\" (UID: \"4906afef-c7e6-4597-a4a5-e9d758917e11\") " pod="openstack/ovn-controller-6kkrw" Feb 18 09:16:30 crc kubenswrapper[4556]: I0218 09:16:30.522139 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4906afef-c7e6-4597-a4a5-e9d758917e11-var-log-ovn\") pod \"ovn-controller-6kkrw\" (UID: \"4906afef-c7e6-4597-a4a5-e9d758917e11\") " pod="openstack/ovn-controller-6kkrw" Feb 18 09:16:30 crc kubenswrapper[4556]: I0218 09:16:30.522971 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4906afef-c7e6-4597-a4a5-e9d758917e11-scripts\") pod \"ovn-controller-6kkrw\" (UID: \"4906afef-c7e6-4597-a4a5-e9d758917e11\") " pod="openstack/ovn-controller-6kkrw" Feb 18 09:16:30 crc kubenswrapper[4556]: I0218 09:16:30.523919 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/4906afef-c7e6-4597-a4a5-e9d758917e11-ovn-controller-tls-certs\") pod \"ovn-controller-6kkrw\" (UID: \"4906afef-c7e6-4597-a4a5-e9d758917e11\") " pod="openstack/ovn-controller-6kkrw" Feb 18 09:16:30 crc kubenswrapper[4556]: I0218 09:16:30.525144 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4906afef-c7e6-4597-a4a5-e9d758917e11-combined-ca-bundle\") pod \"ovn-controller-6kkrw\" (UID: \"4906afef-c7e6-4597-a4a5-e9d758917e11\") " pod="openstack/ovn-controller-6kkrw" Feb 18 09:16:30 crc kubenswrapper[4556]: I0218 09:16:30.541276 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5plct\" (UniqueName: \"kubernetes.io/projected/4906afef-c7e6-4597-a4a5-e9d758917e11-kube-api-access-5plct\") pod \"ovn-controller-6kkrw\" (UID: \"4906afef-c7e6-4597-a4a5-e9d758917e11\") " pod="openstack/ovn-controller-6kkrw" Feb 18 09:16:30 crc kubenswrapper[4556]: I0218 09:16:30.618048 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-6kkrw" Feb 18 09:16:30 crc kubenswrapper[4556]: I0218 09:16:30.648338 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-srnvn" Feb 18 09:16:31 crc kubenswrapper[4556]: I0218 09:16:31.328724 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 18 09:16:31 crc kubenswrapper[4556]: W0218 09:16:31.338528 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod03023501_772f_4798_a953_f56dee586cb1.slice/crio-c24a2bd28a1265077ec721a2e2df3311427addb6fdceba6e220564059fc2a066 WatchSource:0}: Error finding container c24a2bd28a1265077ec721a2e2df3311427addb6fdceba6e220564059fc2a066: Status 404 returned error can't find the container with id c24a2bd28a1265077ec721a2e2df3311427addb6fdceba6e220564059fc2a066 Feb 18 09:16:31 crc kubenswrapper[4556]: I0218 09:16:31.480825 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 18 09:16:31 crc kubenswrapper[4556]: W0218 09:16:31.525541 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1d7961e1_6f96_4707_907d_1fa0c30641e2.slice/crio-25cc6087f58009fb241691501ea8c5e422f1f3c289c1bb7f9ea20ab54db055a6 WatchSource:0}: Error finding container 25cc6087f58009fb241691501ea8c5e422f1f3c289c1bb7f9ea20ab54db055a6: Status 404 returned error can't find the container with id 25cc6087f58009fb241691501ea8c5e422f1f3c289c1bb7f9ea20ab54db055a6 Feb 18 09:16:31 crc kubenswrapper[4556]: I0218 09:16:31.544518 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 18 09:16:31 crc kubenswrapper[4556]: I0218 09:16:31.552763 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Feb 18 09:16:31 crc kubenswrapper[4556]: W0218 09:16:31.603532 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2c504c68_45e0_49dc_90f3_90b91e33551c.slice/crio-4df0b9110b344e27e8f77047822d5d31e6122213af26e34adec8900f1d4f6d61 WatchSource:0}: Error finding container 4df0b9110b344e27e8f77047822d5d31e6122213af26e34adec8900f1d4f6d61: Status 404 returned error can't find the container with id 4df0b9110b344e27e8f77047822d5d31e6122213af26e34adec8900f1d4f6d61 Feb 18 09:16:31 crc kubenswrapper[4556]: W0218 09:16:31.604317 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod11b77df6_67d9_4ddb_944b_2ad8b0fada78.slice/crio-d6b312249ad7cc211e916b2d43f9926550285b7eb86cbcae04531bd36dad81c4 WatchSource:0}: Error finding container d6b312249ad7cc211e916b2d43f9926550285b7eb86cbcae04531bd36dad81c4: Status 404 returned error can't find the container with id d6b312249ad7cc211e916b2d43f9926550285b7eb86cbcae04531bd36dad81c4 Feb 18 09:16:31 crc kubenswrapper[4556]: I0218 09:16:31.632302 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 18 09:16:31 crc kubenswrapper[4556]: I0218 09:16:31.637790 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Feb 18 09:16:31 crc kubenswrapper[4556]: W0218 09:16:31.666310 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod95f33d13_a69d_4add_88a3_21126d52a0ed.slice/crio-d4b199f9ebbe07f1c5a9f8a157f6d56adb74becd56ff76cb650f578ac3dcd5f6 WatchSource:0}: Error finding container d4b199f9ebbe07f1c5a9f8a157f6d56adb74becd56ff76cb650f578ac3dcd5f6: Status 404 returned error can't find the container with id d4b199f9ebbe07f1c5a9f8a157f6d56adb74becd56ff76cb650f578ac3dcd5f6 Feb 18 09:16:31 crc kubenswrapper[4556]: I0218 09:16:31.711751 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-6kkrw"] Feb 18 09:16:31 crc kubenswrapper[4556]: I0218 09:16:31.728107 4556 patch_prober.go:28] interesting pod/machine-config-daemon-f76hs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 09:16:31 crc kubenswrapper[4556]: I0218 09:16:31.728235 4556 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 09:16:31 crc kubenswrapper[4556]: I0218 09:16:31.728287 4556 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" Feb 18 09:16:31 crc kubenswrapper[4556]: I0218 09:16:31.728941 4556 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5eaeb58ab20518b6122148754032f26a9e6729291f4bb5d77b32e55450f0b0ee"} pod="openshift-machine-config-operator/machine-config-daemon-f76hs" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 18 09:16:31 crc kubenswrapper[4556]: I0218 09:16:31.728984 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerName="machine-config-daemon" containerID="cri-o://5eaeb58ab20518b6122148754032f26a9e6729291f4bb5d77b32e55450f0b0ee" gracePeriod=600 Feb 18 09:16:31 crc kubenswrapper[4556]: W0218 09:16:31.743897 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4906afef_c7e6_4597_a4a5_e9d758917e11.slice/crio-58a42936c08e5d0732e44d3f4cda9f3c9304eaf2f15b5808e6da354bc93e8c2b WatchSource:0}: Error finding container 58a42936c08e5d0732e44d3f4cda9f3c9304eaf2f15b5808e6da354bc93e8c2b: Status 404 returned error can't find the container with id 58a42936c08e5d0732e44d3f4cda9f3c9304eaf2f15b5808e6da354bc93e8c2b Feb 18 09:16:31 crc kubenswrapper[4556]: I0218 09:16:31.783480 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-srnvn"] Feb 18 09:16:31 crc kubenswrapper[4556]: W0218 09:16:31.785063 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod40c51577_de87_4bd8_be03_41221bc2f415.slice/crio-2556c01327456513090193364a4d381138c03938f33170eae8bacb3a2f5b0f5e WatchSource:0}: Error finding container 2556c01327456513090193364a4d381138c03938f33170eae8bacb3a2f5b0f5e: Status 404 returned error can't find the container with id 2556c01327456513090193364a4d381138c03938f33170eae8bacb3a2f5b0f5e Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.186811 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.188892 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.191423 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.191596 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.191599 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.191637 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-skrjk" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.193532 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.204945 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.303857 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"2c504c68-45e0-49dc-90f3-90b91e33551c","Type":"ContainerStarted","Data":"4df0b9110b344e27e8f77047822d5d31e6122213af26e34adec8900f1d4f6d61"} Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.305185 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"95f33d13-a69d-4add-88a3-21126d52a0ed","Type":"ContainerStarted","Data":"d4b199f9ebbe07f1c5a9f8a157f6d56adb74becd56ff76cb650f578ac3dcd5f6"} Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.306147 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"03023501-772f-4798-a953-f56dee586cb1","Type":"ContainerStarted","Data":"c24a2bd28a1265077ec721a2e2df3311427addb6fdceba6e220564059fc2a066"} Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.307260 4556 generic.go:334] "Generic (PLEG): container finished" podID="ab4c456b-9c1a-4bd8-a35a-fe73617002c9" containerID="25ff0a03ff53df1eabf20a4b3161ca776f08e8e3ff6e364b72529bd74d799337" exitCode=0 Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.307302 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6fcf94d689-vsztz" event={"ID":"ab4c456b-9c1a-4bd8-a35a-fe73617002c9","Type":"ContainerDied","Data":"25ff0a03ff53df1eabf20a4b3161ca776f08e8e3ff6e364b72529bd74d799337"} Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.309466 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1d7961e1-6f96-4707-907d-1fa0c30641e2","Type":"ContainerStarted","Data":"25cc6087f58009fb241691501ea8c5e422f1f3c289c1bb7f9ea20ab54db055a6"} Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.316870 4556 generic.go:334] "Generic (PLEG): container finished" podID="2fd56f2a-ea59-4b71-8468-56d18eda194a" containerID="0e8dce06934a1d8e8e642a478572d660aaf1657ba0d59076856e01e7fcdfd51c" exitCode=0 Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.316977 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-855cbc58c5-dh8dd" event={"ID":"2fd56f2a-ea59-4b71-8468-56d18eda194a","Type":"ContainerDied","Data":"0e8dce06934a1d8e8e642a478572d660aaf1657ba0d59076856e01e7fcdfd51c"} Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.322654 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"68274946-b189-450c-a154-27059f411af3","Type":"ContainerStarted","Data":"f392a53fb0820661c97a3a0f36a86d02026af3afdc59561d1f06cc4a0df529fa"} Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.324535 4556 generic.go:334] "Generic (PLEG): container finished" podID="78221461-c2ed-45d9-bd36-aa078d36d757" containerID="d24f089f65c98b0bf17cede561b911d8ca46377e4d023b3d5a9df6c7ef7f8e42" exitCode=0 Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.324625 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f54874ffc-rb6x9" event={"ID":"78221461-c2ed-45d9-bd36-aa078d36d757","Type":"ContainerDied","Data":"d24f089f65c98b0bf17cede561b911d8ca46377e4d023b3d5a9df6c7ef7f8e42"} Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.331564 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-srnvn" event={"ID":"40c51577-de87-4bd8-be03-41221bc2f415","Type":"ContainerStarted","Data":"2556c01327456513090193364a4d381138c03938f33170eae8bacb3a2f5b0f5e"} Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.341561 4556 generic.go:334] "Generic (PLEG): container finished" podID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerID="5eaeb58ab20518b6122148754032f26a9e6729291f4bb5d77b32e55450f0b0ee" exitCode=0 Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.341648 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" event={"ID":"8dac7f27-d3d1-4778-9e54-f273035a1d37","Type":"ContainerDied","Data":"5eaeb58ab20518b6122148754032f26a9e6729291f4bb5d77b32e55450f0b0ee"} Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.341677 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" event={"ID":"8dac7f27-d3d1-4778-9e54-f273035a1d37","Type":"ContainerStarted","Data":"7138060d5bb9e6802973b6e4fec551cb2729ba7d9b32c912776fd108e1b87ace"} Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.341697 4556 scope.go:117] "RemoveContainer" containerID="dc67a95c5ea909b9cea5fe464915331393bc298bd4bcd198fcb5ebe8f8ba8a4b" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.347258 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"11b77df6-67d9-4ddb-944b-2ad8b0fada78","Type":"ContainerStarted","Data":"d6b312249ad7cc211e916b2d43f9926550285b7eb86cbcae04531bd36dad81c4"} Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.348568 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-6kkrw" event={"ID":"4906afef-c7e6-4597-a4a5-e9d758917e11","Type":"ContainerStarted","Data":"58a42936c08e5d0732e44d3f4cda9f3c9304eaf2f15b5808e6da354bc93e8c2b"} Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.350836 4556 generic.go:334] "Generic (PLEG): container finished" podID="fa586b98-2a59-4673-a0bf-f7175f30d356" containerID="89933aa5c0aaa3e505249d03754c47c7de8ff057003a4042350fac5d0d7e5eea" exitCode=0 Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.350886 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67ff45466c-wdkxp" event={"ID":"fa586b98-2a59-4673-a0bf-f7175f30d356","Type":"ContainerDied","Data":"89933aa5c0aaa3e505249d03754c47c7de8ff057003a4042350fac5d0d7e5eea"} Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.361430 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10b5d4cb-2beb-405d-83fc-30eb5fd4aaad-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"10b5d4cb-2beb-405d-83fc-30eb5fd4aaad\") " pod="openstack/ovsdbserver-nb-0" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.361508 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/10b5d4cb-2beb-405d-83fc-30eb5fd4aaad-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"10b5d4cb-2beb-405d-83fc-30eb5fd4aaad\") " pod="openstack/ovsdbserver-nb-0" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.361530 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7sknd\" (UniqueName: \"kubernetes.io/projected/10b5d4cb-2beb-405d-83fc-30eb5fd4aaad-kube-api-access-7sknd\") pod \"ovsdbserver-nb-0\" (UID: \"10b5d4cb-2beb-405d-83fc-30eb5fd4aaad\") " pod="openstack/ovsdbserver-nb-0" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.361652 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10b5d4cb-2beb-405d-83fc-30eb5fd4aaad-config\") pod \"ovsdbserver-nb-0\" (UID: \"10b5d4cb-2beb-405d-83fc-30eb5fd4aaad\") " pod="openstack/ovsdbserver-nb-0" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.361746 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/10b5d4cb-2beb-405d-83fc-30eb5fd4aaad-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"10b5d4cb-2beb-405d-83fc-30eb5fd4aaad\") " pod="openstack/ovsdbserver-nb-0" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.361761 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/10b5d4cb-2beb-405d-83fc-30eb5fd4aaad-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"10b5d4cb-2beb-405d-83fc-30eb5fd4aaad\") " pod="openstack/ovsdbserver-nb-0" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.361837 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"10b5d4cb-2beb-405d-83fc-30eb5fd4aaad\") " pod="openstack/ovsdbserver-nb-0" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.361904 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/10b5d4cb-2beb-405d-83fc-30eb5fd4aaad-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"10b5d4cb-2beb-405d-83fc-30eb5fd4aaad\") " pod="openstack/ovsdbserver-nb-0" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.461004 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-j7rpf"] Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.462000 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-j7rpf" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.463476 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/10b5d4cb-2beb-405d-83fc-30eb5fd4aaad-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"10b5d4cb-2beb-405d-83fc-30eb5fd4aaad\") " pod="openstack/ovsdbserver-nb-0" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.463504 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7sknd\" (UniqueName: \"kubernetes.io/projected/10b5d4cb-2beb-405d-83fc-30eb5fd4aaad-kube-api-access-7sknd\") pod \"ovsdbserver-nb-0\" (UID: \"10b5d4cb-2beb-405d-83fc-30eb5fd4aaad\") " pod="openstack/ovsdbserver-nb-0" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.463563 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10b5d4cb-2beb-405d-83fc-30eb5fd4aaad-config\") pod \"ovsdbserver-nb-0\" (UID: \"10b5d4cb-2beb-405d-83fc-30eb5fd4aaad\") " pod="openstack/ovsdbserver-nb-0" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.463620 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/10b5d4cb-2beb-405d-83fc-30eb5fd4aaad-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"10b5d4cb-2beb-405d-83fc-30eb5fd4aaad\") " pod="openstack/ovsdbserver-nb-0" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.463635 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/10b5d4cb-2beb-405d-83fc-30eb5fd4aaad-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"10b5d4cb-2beb-405d-83fc-30eb5fd4aaad\") " pod="openstack/ovsdbserver-nb-0" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.463668 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"10b5d4cb-2beb-405d-83fc-30eb5fd4aaad\") " pod="openstack/ovsdbserver-nb-0" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.463715 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/10b5d4cb-2beb-405d-83fc-30eb5fd4aaad-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"10b5d4cb-2beb-405d-83fc-30eb5fd4aaad\") " pod="openstack/ovsdbserver-nb-0" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.463858 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10b5d4cb-2beb-405d-83fc-30eb5fd4aaad-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"10b5d4cb-2beb-405d-83fc-30eb5fd4aaad\") " pod="openstack/ovsdbserver-nb-0" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.464209 4556 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"10b5d4cb-2beb-405d-83fc-30eb5fd4aaad\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/ovsdbserver-nb-0" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.465237 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/10b5d4cb-2beb-405d-83fc-30eb5fd4aaad-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"10b5d4cb-2beb-405d-83fc-30eb5fd4aaad\") " pod="openstack/ovsdbserver-nb-0" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.465547 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/10b5d4cb-2beb-405d-83fc-30eb5fd4aaad-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"10b5d4cb-2beb-405d-83fc-30eb5fd4aaad\") " pod="openstack/ovsdbserver-nb-0" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.465705 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.465892 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10b5d4cb-2beb-405d-83fc-30eb5fd4aaad-config\") pod \"ovsdbserver-nb-0\" (UID: \"10b5d4cb-2beb-405d-83fc-30eb5fd4aaad\") " pod="openstack/ovsdbserver-nb-0" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.470013 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10b5d4cb-2beb-405d-83fc-30eb5fd4aaad-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"10b5d4cb-2beb-405d-83fc-30eb5fd4aaad\") " pod="openstack/ovsdbserver-nb-0" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.471291 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/10b5d4cb-2beb-405d-83fc-30eb5fd4aaad-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"10b5d4cb-2beb-405d-83fc-30eb5fd4aaad\") " pod="openstack/ovsdbserver-nb-0" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.471857 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/10b5d4cb-2beb-405d-83fc-30eb5fd4aaad-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"10b5d4cb-2beb-405d-83fc-30eb5fd4aaad\") " pod="openstack/ovsdbserver-nb-0" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.486813 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-j7rpf"] Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.526107 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7sknd\" (UniqueName: \"kubernetes.io/projected/10b5d4cb-2beb-405d-83fc-30eb5fd4aaad-kube-api-access-7sknd\") pod \"ovsdbserver-nb-0\" (UID: \"10b5d4cb-2beb-405d-83fc-30eb5fd4aaad\") " pod="openstack/ovsdbserver-nb-0" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.530459 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"10b5d4cb-2beb-405d-83fc-30eb5fd4aaad\") " pod="openstack/ovsdbserver-nb-0" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.565480 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/66575dd2-efe2-4770-a9cd-7afb99e8566c-ovs-rundir\") pod \"ovn-controller-metrics-j7rpf\" (UID: \"66575dd2-efe2-4770-a9cd-7afb99e8566c\") " pod="openstack/ovn-controller-metrics-j7rpf" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.565517 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66575dd2-efe2-4770-a9cd-7afb99e8566c-config\") pod \"ovn-controller-metrics-j7rpf\" (UID: \"66575dd2-efe2-4770-a9cd-7afb99e8566c\") " pod="openstack/ovn-controller-metrics-j7rpf" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.565545 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/66575dd2-efe2-4770-a9cd-7afb99e8566c-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-j7rpf\" (UID: \"66575dd2-efe2-4770-a9cd-7afb99e8566c\") " pod="openstack/ovn-controller-metrics-j7rpf" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.565711 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66575dd2-efe2-4770-a9cd-7afb99e8566c-combined-ca-bundle\") pod \"ovn-controller-metrics-j7rpf\" (UID: \"66575dd2-efe2-4770-a9cd-7afb99e8566c\") " pod="openstack/ovn-controller-metrics-j7rpf" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.565827 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/66575dd2-efe2-4770-a9cd-7afb99e8566c-ovn-rundir\") pod \"ovn-controller-metrics-j7rpf\" (UID: \"66575dd2-efe2-4770-a9cd-7afb99e8566c\") " pod="openstack/ovn-controller-metrics-j7rpf" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.566062 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72w9t\" (UniqueName: \"kubernetes.io/projected/66575dd2-efe2-4770-a9cd-7afb99e8566c-kube-api-access-72w9t\") pod \"ovn-controller-metrics-j7rpf\" (UID: \"66575dd2-efe2-4770-a9cd-7afb99e8566c\") " pod="openstack/ovn-controller-metrics-j7rpf" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.667809 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72w9t\" (UniqueName: \"kubernetes.io/projected/66575dd2-efe2-4770-a9cd-7afb99e8566c-kube-api-access-72w9t\") pod \"ovn-controller-metrics-j7rpf\" (UID: \"66575dd2-efe2-4770-a9cd-7afb99e8566c\") " pod="openstack/ovn-controller-metrics-j7rpf" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.667903 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/66575dd2-efe2-4770-a9cd-7afb99e8566c-ovs-rundir\") pod \"ovn-controller-metrics-j7rpf\" (UID: \"66575dd2-efe2-4770-a9cd-7afb99e8566c\") " pod="openstack/ovn-controller-metrics-j7rpf" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.667929 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66575dd2-efe2-4770-a9cd-7afb99e8566c-config\") pod \"ovn-controller-metrics-j7rpf\" (UID: \"66575dd2-efe2-4770-a9cd-7afb99e8566c\") " pod="openstack/ovn-controller-metrics-j7rpf" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.667990 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/66575dd2-efe2-4770-a9cd-7afb99e8566c-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-j7rpf\" (UID: \"66575dd2-efe2-4770-a9cd-7afb99e8566c\") " pod="openstack/ovn-controller-metrics-j7rpf" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.668021 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66575dd2-efe2-4770-a9cd-7afb99e8566c-combined-ca-bundle\") pod \"ovn-controller-metrics-j7rpf\" (UID: \"66575dd2-efe2-4770-a9cd-7afb99e8566c\") " pod="openstack/ovn-controller-metrics-j7rpf" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.668054 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/66575dd2-efe2-4770-a9cd-7afb99e8566c-ovn-rundir\") pod \"ovn-controller-metrics-j7rpf\" (UID: \"66575dd2-efe2-4770-a9cd-7afb99e8566c\") " pod="openstack/ovn-controller-metrics-j7rpf" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.668203 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/66575dd2-efe2-4770-a9cd-7afb99e8566c-ovs-rundir\") pod \"ovn-controller-metrics-j7rpf\" (UID: \"66575dd2-efe2-4770-a9cd-7afb99e8566c\") " pod="openstack/ovn-controller-metrics-j7rpf" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.668729 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66575dd2-efe2-4770-a9cd-7afb99e8566c-config\") pod \"ovn-controller-metrics-j7rpf\" (UID: \"66575dd2-efe2-4770-a9cd-7afb99e8566c\") " pod="openstack/ovn-controller-metrics-j7rpf" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.669246 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/66575dd2-efe2-4770-a9cd-7afb99e8566c-ovn-rundir\") pod \"ovn-controller-metrics-j7rpf\" (UID: \"66575dd2-efe2-4770-a9cd-7afb99e8566c\") " pod="openstack/ovn-controller-metrics-j7rpf" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.671507 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66575dd2-efe2-4770-a9cd-7afb99e8566c-combined-ca-bundle\") pod \"ovn-controller-metrics-j7rpf\" (UID: \"66575dd2-efe2-4770-a9cd-7afb99e8566c\") " pod="openstack/ovn-controller-metrics-j7rpf" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.672656 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/66575dd2-efe2-4770-a9cd-7afb99e8566c-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-j7rpf\" (UID: \"66575dd2-efe2-4770-a9cd-7afb99e8566c\") " pod="openstack/ovn-controller-metrics-j7rpf" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.681644 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72w9t\" (UniqueName: \"kubernetes.io/projected/66575dd2-efe2-4770-a9cd-7afb99e8566c-kube-api-access-72w9t\") pod \"ovn-controller-metrics-j7rpf\" (UID: \"66575dd2-efe2-4770-a9cd-7afb99e8566c\") " pod="openstack/ovn-controller-metrics-j7rpf" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.707463 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6fcf94d689-vsztz" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.712438 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-855cbc58c5-dh8dd" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.815455 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.815894 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-j7rpf" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.871184 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2cpxc\" (UniqueName: \"kubernetes.io/projected/2fd56f2a-ea59-4b71-8468-56d18eda194a-kube-api-access-2cpxc\") pod \"2fd56f2a-ea59-4b71-8468-56d18eda194a\" (UID: \"2fd56f2a-ea59-4b71-8468-56d18eda194a\") " Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.871289 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab4c456b-9c1a-4bd8-a35a-fe73617002c9-config\") pod \"ab4c456b-9c1a-4bd8-a35a-fe73617002c9\" (UID: \"ab4c456b-9c1a-4bd8-a35a-fe73617002c9\") " Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.871335 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ab4c456b-9c1a-4bd8-a35a-fe73617002c9-dns-svc\") pod \"ab4c456b-9c1a-4bd8-a35a-fe73617002c9\" (UID: \"ab4c456b-9c1a-4bd8-a35a-fe73617002c9\") " Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.871407 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fd56f2a-ea59-4b71-8468-56d18eda194a-config\") pod \"2fd56f2a-ea59-4b71-8468-56d18eda194a\" (UID: \"2fd56f2a-ea59-4b71-8468-56d18eda194a\") " Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.871555 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jz862\" (UniqueName: \"kubernetes.io/projected/ab4c456b-9c1a-4bd8-a35a-fe73617002c9-kube-api-access-jz862\") pod \"ab4c456b-9c1a-4bd8-a35a-fe73617002c9\" (UID: \"ab4c456b-9c1a-4bd8-a35a-fe73617002c9\") " Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.875737 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fd56f2a-ea59-4b71-8468-56d18eda194a-kube-api-access-2cpxc" (OuterVolumeSpecName: "kube-api-access-2cpxc") pod "2fd56f2a-ea59-4b71-8468-56d18eda194a" (UID: "2fd56f2a-ea59-4b71-8468-56d18eda194a"). InnerVolumeSpecName "kube-api-access-2cpxc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.876165 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab4c456b-9c1a-4bd8-a35a-fe73617002c9-kube-api-access-jz862" (OuterVolumeSpecName: "kube-api-access-jz862") pod "ab4c456b-9c1a-4bd8-a35a-fe73617002c9" (UID: "ab4c456b-9c1a-4bd8-a35a-fe73617002c9"). InnerVolumeSpecName "kube-api-access-jz862". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.886792 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab4c456b-9c1a-4bd8-a35a-fe73617002c9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ab4c456b-9c1a-4bd8-a35a-fe73617002c9" (UID: "ab4c456b-9c1a-4bd8-a35a-fe73617002c9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.887674 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab4c456b-9c1a-4bd8-a35a-fe73617002c9-config" (OuterVolumeSpecName: "config") pod "ab4c456b-9c1a-4bd8-a35a-fe73617002c9" (UID: "ab4c456b-9c1a-4bd8-a35a-fe73617002c9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.891332 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2fd56f2a-ea59-4b71-8468-56d18eda194a-config" (OuterVolumeSpecName: "config") pod "2fd56f2a-ea59-4b71-8468-56d18eda194a" (UID: "2fd56f2a-ea59-4b71-8468-56d18eda194a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.973439 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jz862\" (UniqueName: \"kubernetes.io/projected/ab4c456b-9c1a-4bd8-a35a-fe73617002c9-kube-api-access-jz862\") on node \"crc\" DevicePath \"\"" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.973711 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2cpxc\" (UniqueName: \"kubernetes.io/projected/2fd56f2a-ea59-4b71-8468-56d18eda194a-kube-api-access-2cpxc\") on node \"crc\" DevicePath \"\"" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.973724 4556 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab4c456b-9c1a-4bd8-a35a-fe73617002c9-config\") on node \"crc\" DevicePath \"\"" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.973733 4556 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ab4c456b-9c1a-4bd8-a35a-fe73617002c9-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 18 09:16:32 crc kubenswrapper[4556]: I0218 09:16:32.973742 4556 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fd56f2a-ea59-4b71-8468-56d18eda194a-config\") on node \"crc\" DevicePath \"\"" Feb 18 09:16:33 crc kubenswrapper[4556]: I0218 09:16:33.376943 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6fcf94d689-vsztz" event={"ID":"ab4c456b-9c1a-4bd8-a35a-fe73617002c9","Type":"ContainerDied","Data":"99b51f22c992596c809d36264d75223901c85c5053091409733a31c4adad6e49"} Feb 18 09:16:33 crc kubenswrapper[4556]: I0218 09:16:33.377002 4556 scope.go:117] "RemoveContainer" containerID="25ff0a03ff53df1eabf20a4b3161ca776f08e8e3ff6e364b72529bd74d799337" Feb 18 09:16:33 crc kubenswrapper[4556]: I0218 09:16:33.377102 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6fcf94d689-vsztz" Feb 18 09:16:33 crc kubenswrapper[4556]: I0218 09:16:33.388988 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-855cbc58c5-dh8dd" event={"ID":"2fd56f2a-ea59-4b71-8468-56d18eda194a","Type":"ContainerDied","Data":"f8ff36375c709a7773baceded667406673606a77aff7bd08a32b83de0b697e6c"} Feb 18 09:16:33 crc kubenswrapper[4556]: I0218 09:16:33.389033 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-855cbc58c5-dh8dd" Feb 18 09:16:33 crc kubenswrapper[4556]: I0218 09:16:33.472788 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-855cbc58c5-dh8dd"] Feb 18 09:16:33 crc kubenswrapper[4556]: I0218 09:16:33.477140 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-855cbc58c5-dh8dd"] Feb 18 09:16:33 crc kubenswrapper[4556]: I0218 09:16:33.486761 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6fcf94d689-vsztz"] Feb 18 09:16:33 crc kubenswrapper[4556]: I0218 09:16:33.489528 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6fcf94d689-vsztz"] Feb 18 09:16:33 crc kubenswrapper[4556]: I0218 09:16:33.521086 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 18 09:16:33 crc kubenswrapper[4556]: I0218 09:16:33.603335 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-j7rpf"] Feb 18 09:16:33 crc kubenswrapper[4556]: W0218 09:16:33.936732 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod10b5d4cb_2beb_405d_83fc_30eb5fd4aaad.slice/crio-b48bfcb3cae19cc7c49541fc9b0c55a2ed45f44a766c3b74a0abfbf719668d85 WatchSource:0}: Error finding container b48bfcb3cae19cc7c49541fc9b0c55a2ed45f44a766c3b74a0abfbf719668d85: Status 404 returned error can't find the container with id b48bfcb3cae19cc7c49541fc9b0c55a2ed45f44a766c3b74a0abfbf719668d85 Feb 18 09:16:34 crc kubenswrapper[4556]: I0218 09:16:34.399355 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"10b5d4cb-2beb-405d-83fc-30eb5fd4aaad","Type":"ContainerStarted","Data":"b48bfcb3cae19cc7c49541fc9b0c55a2ed45f44a766c3b74a0abfbf719668d85"} Feb 18 09:16:34 crc kubenswrapper[4556]: I0218 09:16:34.402178 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-j7rpf" event={"ID":"66575dd2-efe2-4770-a9cd-7afb99e8566c","Type":"ContainerStarted","Data":"4edb626633d25dcdf2b46638d193a3d003d0651af43746f57ab448edb84a656d"} Feb 18 09:16:34 crc kubenswrapper[4556]: I0218 09:16:34.860607 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 18 09:16:34 crc kubenswrapper[4556]: E0218 09:16:34.861469 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fd56f2a-ea59-4b71-8468-56d18eda194a" containerName="init" Feb 18 09:16:34 crc kubenswrapper[4556]: I0218 09:16:34.861504 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fd56f2a-ea59-4b71-8468-56d18eda194a" containerName="init" Feb 18 09:16:34 crc kubenswrapper[4556]: E0218 09:16:34.861539 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab4c456b-9c1a-4bd8-a35a-fe73617002c9" containerName="init" Feb 18 09:16:34 crc kubenswrapper[4556]: I0218 09:16:34.861547 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab4c456b-9c1a-4bd8-a35a-fe73617002c9" containerName="init" Feb 18 09:16:34 crc kubenswrapper[4556]: I0218 09:16:34.861748 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab4c456b-9c1a-4bd8-a35a-fe73617002c9" containerName="init" Feb 18 09:16:34 crc kubenswrapper[4556]: I0218 09:16:34.861776 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fd56f2a-ea59-4b71-8468-56d18eda194a" containerName="init" Feb 18 09:16:34 crc kubenswrapper[4556]: I0218 09:16:34.863861 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Feb 18 09:16:34 crc kubenswrapper[4556]: I0218 09:16:34.866344 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Feb 18 09:16:34 crc kubenswrapper[4556]: I0218 09:16:34.866666 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Feb 18 09:16:34 crc kubenswrapper[4556]: I0218 09:16:34.866742 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Feb 18 09:16:34 crc kubenswrapper[4556]: I0218 09:16:34.867347 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-jnk5w" Feb 18 09:16:34 crc kubenswrapper[4556]: I0218 09:16:34.879913 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 18 09:16:35 crc kubenswrapper[4556]: I0218 09:16:35.032428 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7840643-68ec-4a2a-b6af-8a9730729077-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"d7840643-68ec-4a2a-b6af-8a9730729077\") " pod="openstack/ovsdbserver-sb-0" Feb 18 09:16:35 crc kubenswrapper[4556]: I0218 09:16:35.032513 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7840643-68ec-4a2a-b6af-8a9730729077-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d7840643-68ec-4a2a-b6af-8a9730729077\") " pod="openstack/ovsdbserver-sb-0" Feb 18 09:16:35 crc kubenswrapper[4556]: I0218 09:16:35.032683 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7840643-68ec-4a2a-b6af-8a9730729077-config\") pod \"ovsdbserver-sb-0\" (UID: \"d7840643-68ec-4a2a-b6af-8a9730729077\") " pod="openstack/ovsdbserver-sb-0" Feb 18 09:16:35 crc kubenswrapper[4556]: I0218 09:16:35.032771 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7840643-68ec-4a2a-b6af-8a9730729077-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d7840643-68ec-4a2a-b6af-8a9730729077\") " pod="openstack/ovsdbserver-sb-0" Feb 18 09:16:35 crc kubenswrapper[4556]: I0218 09:16:35.032853 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"d7840643-68ec-4a2a-b6af-8a9730729077\") " pod="openstack/ovsdbserver-sb-0" Feb 18 09:16:35 crc kubenswrapper[4556]: I0218 09:16:35.032885 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xsr8g\" (UniqueName: \"kubernetes.io/projected/d7840643-68ec-4a2a-b6af-8a9730729077-kube-api-access-xsr8g\") pod \"ovsdbserver-sb-0\" (UID: \"d7840643-68ec-4a2a-b6af-8a9730729077\") " pod="openstack/ovsdbserver-sb-0" Feb 18 09:16:35 crc kubenswrapper[4556]: I0218 09:16:35.032939 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d7840643-68ec-4a2a-b6af-8a9730729077-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"d7840643-68ec-4a2a-b6af-8a9730729077\") " pod="openstack/ovsdbserver-sb-0" Feb 18 09:16:35 crc kubenswrapper[4556]: I0218 09:16:35.032971 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d7840643-68ec-4a2a-b6af-8a9730729077-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"d7840643-68ec-4a2a-b6af-8a9730729077\") " pod="openstack/ovsdbserver-sb-0" Feb 18 09:16:35 crc kubenswrapper[4556]: I0218 09:16:35.135549 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d7840643-68ec-4a2a-b6af-8a9730729077-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"d7840643-68ec-4a2a-b6af-8a9730729077\") " pod="openstack/ovsdbserver-sb-0" Feb 18 09:16:35 crc kubenswrapper[4556]: I0218 09:16:35.135701 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d7840643-68ec-4a2a-b6af-8a9730729077-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"d7840643-68ec-4a2a-b6af-8a9730729077\") " pod="openstack/ovsdbserver-sb-0" Feb 18 09:16:35 crc kubenswrapper[4556]: I0218 09:16:35.135755 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7840643-68ec-4a2a-b6af-8a9730729077-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"d7840643-68ec-4a2a-b6af-8a9730729077\") " pod="openstack/ovsdbserver-sb-0" Feb 18 09:16:35 crc kubenswrapper[4556]: I0218 09:16:35.135799 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7840643-68ec-4a2a-b6af-8a9730729077-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d7840643-68ec-4a2a-b6af-8a9730729077\") " pod="openstack/ovsdbserver-sb-0" Feb 18 09:16:35 crc kubenswrapper[4556]: I0218 09:16:35.135869 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7840643-68ec-4a2a-b6af-8a9730729077-config\") pod \"ovsdbserver-sb-0\" (UID: \"d7840643-68ec-4a2a-b6af-8a9730729077\") " pod="openstack/ovsdbserver-sb-0" Feb 18 09:16:35 crc kubenswrapper[4556]: I0218 09:16:35.135940 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7840643-68ec-4a2a-b6af-8a9730729077-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d7840643-68ec-4a2a-b6af-8a9730729077\") " pod="openstack/ovsdbserver-sb-0" Feb 18 09:16:35 crc kubenswrapper[4556]: I0218 09:16:35.135968 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"d7840643-68ec-4a2a-b6af-8a9730729077\") " pod="openstack/ovsdbserver-sb-0" Feb 18 09:16:35 crc kubenswrapper[4556]: I0218 09:16:35.135990 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xsr8g\" (UniqueName: \"kubernetes.io/projected/d7840643-68ec-4a2a-b6af-8a9730729077-kube-api-access-xsr8g\") pod \"ovsdbserver-sb-0\" (UID: \"d7840643-68ec-4a2a-b6af-8a9730729077\") " pod="openstack/ovsdbserver-sb-0" Feb 18 09:16:35 crc kubenswrapper[4556]: I0218 09:16:35.136748 4556 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"d7840643-68ec-4a2a-b6af-8a9730729077\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/ovsdbserver-sb-0" Feb 18 09:16:35 crc kubenswrapper[4556]: I0218 09:16:35.137683 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d7840643-68ec-4a2a-b6af-8a9730729077-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"d7840643-68ec-4a2a-b6af-8a9730729077\") " pod="openstack/ovsdbserver-sb-0" Feb 18 09:16:35 crc kubenswrapper[4556]: I0218 09:16:35.137693 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d7840643-68ec-4a2a-b6af-8a9730729077-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"d7840643-68ec-4a2a-b6af-8a9730729077\") " pod="openstack/ovsdbserver-sb-0" Feb 18 09:16:35 crc kubenswrapper[4556]: I0218 09:16:35.139027 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7840643-68ec-4a2a-b6af-8a9730729077-config\") pod \"ovsdbserver-sb-0\" (UID: \"d7840643-68ec-4a2a-b6af-8a9730729077\") " pod="openstack/ovsdbserver-sb-0" Feb 18 09:16:35 crc kubenswrapper[4556]: I0218 09:16:35.143548 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7840643-68ec-4a2a-b6af-8a9730729077-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d7840643-68ec-4a2a-b6af-8a9730729077\") " pod="openstack/ovsdbserver-sb-0" Feb 18 09:16:35 crc kubenswrapper[4556]: I0218 09:16:35.144696 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7840643-68ec-4a2a-b6af-8a9730729077-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"d7840643-68ec-4a2a-b6af-8a9730729077\") " pod="openstack/ovsdbserver-sb-0" Feb 18 09:16:35 crc kubenswrapper[4556]: I0218 09:16:35.149835 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7840643-68ec-4a2a-b6af-8a9730729077-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d7840643-68ec-4a2a-b6af-8a9730729077\") " pod="openstack/ovsdbserver-sb-0" Feb 18 09:16:35 crc kubenswrapper[4556]: I0218 09:16:35.152770 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xsr8g\" (UniqueName: \"kubernetes.io/projected/d7840643-68ec-4a2a-b6af-8a9730729077-kube-api-access-xsr8g\") pod \"ovsdbserver-sb-0\" (UID: \"d7840643-68ec-4a2a-b6af-8a9730729077\") " pod="openstack/ovsdbserver-sb-0" Feb 18 09:16:35 crc kubenswrapper[4556]: I0218 09:16:35.170263 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"d7840643-68ec-4a2a-b6af-8a9730729077\") " pod="openstack/ovsdbserver-sb-0" Feb 18 09:16:35 crc kubenswrapper[4556]: I0218 09:16:35.184238 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Feb 18 09:16:35 crc kubenswrapper[4556]: I0218 09:16:35.265892 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67ff45466c-wdkxp"] Feb 18 09:16:35 crc kubenswrapper[4556]: I0218 09:16:35.313840 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2fd56f2a-ea59-4b71-8468-56d18eda194a" path="/var/lib/kubelet/pods/2fd56f2a-ea59-4b71-8468-56d18eda194a/volumes" Feb 18 09:16:35 crc kubenswrapper[4556]: I0218 09:16:35.314577 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab4c456b-9c1a-4bd8-a35a-fe73617002c9" path="/var/lib/kubelet/pods/ab4c456b-9c1a-4bd8-a35a-fe73617002c9/volumes" Feb 18 09:16:35 crc kubenswrapper[4556]: I0218 09:16:35.315204 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-64f7f48db9-mjw87"] Feb 18 09:16:35 crc kubenswrapper[4556]: I0218 09:16:35.316848 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64f7f48db9-mjw87" Feb 18 09:16:35 crc kubenswrapper[4556]: I0218 09:16:35.318545 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Feb 18 09:16:35 crc kubenswrapper[4556]: I0218 09:16:35.336409 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64f7f48db9-mjw87"] Feb 18 09:16:35 crc kubenswrapper[4556]: I0218 09:16:35.441578 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jntbs\" (UniqueName: \"kubernetes.io/projected/30a1b29b-530b-4f64-acb7-a8471fc98fcb-kube-api-access-jntbs\") pod \"dnsmasq-dns-64f7f48db9-mjw87\" (UID: \"30a1b29b-530b-4f64-acb7-a8471fc98fcb\") " pod="openstack/dnsmasq-dns-64f7f48db9-mjw87" Feb 18 09:16:35 crc kubenswrapper[4556]: I0218 09:16:35.441692 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/30a1b29b-530b-4f64-acb7-a8471fc98fcb-ovsdbserver-nb\") pod \"dnsmasq-dns-64f7f48db9-mjw87\" (UID: \"30a1b29b-530b-4f64-acb7-a8471fc98fcb\") " pod="openstack/dnsmasq-dns-64f7f48db9-mjw87" Feb 18 09:16:35 crc kubenswrapper[4556]: I0218 09:16:35.441864 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30a1b29b-530b-4f64-acb7-a8471fc98fcb-config\") pod \"dnsmasq-dns-64f7f48db9-mjw87\" (UID: \"30a1b29b-530b-4f64-acb7-a8471fc98fcb\") " pod="openstack/dnsmasq-dns-64f7f48db9-mjw87" Feb 18 09:16:35 crc kubenswrapper[4556]: I0218 09:16:35.441959 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/30a1b29b-530b-4f64-acb7-a8471fc98fcb-dns-svc\") pod \"dnsmasq-dns-64f7f48db9-mjw87\" (UID: \"30a1b29b-530b-4f64-acb7-a8471fc98fcb\") " pod="openstack/dnsmasq-dns-64f7f48db9-mjw87" Feb 18 09:16:35 crc kubenswrapper[4556]: I0218 09:16:35.544413 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/30a1b29b-530b-4f64-acb7-a8471fc98fcb-ovsdbserver-nb\") pod \"dnsmasq-dns-64f7f48db9-mjw87\" (UID: \"30a1b29b-530b-4f64-acb7-a8471fc98fcb\") " pod="openstack/dnsmasq-dns-64f7f48db9-mjw87" Feb 18 09:16:35 crc kubenswrapper[4556]: I0218 09:16:35.544490 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30a1b29b-530b-4f64-acb7-a8471fc98fcb-config\") pod \"dnsmasq-dns-64f7f48db9-mjw87\" (UID: \"30a1b29b-530b-4f64-acb7-a8471fc98fcb\") " pod="openstack/dnsmasq-dns-64f7f48db9-mjw87" Feb 18 09:16:35 crc kubenswrapper[4556]: I0218 09:16:35.544528 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/30a1b29b-530b-4f64-acb7-a8471fc98fcb-dns-svc\") pod \"dnsmasq-dns-64f7f48db9-mjw87\" (UID: \"30a1b29b-530b-4f64-acb7-a8471fc98fcb\") " pod="openstack/dnsmasq-dns-64f7f48db9-mjw87" Feb 18 09:16:35 crc kubenswrapper[4556]: I0218 09:16:35.544720 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jntbs\" (UniqueName: \"kubernetes.io/projected/30a1b29b-530b-4f64-acb7-a8471fc98fcb-kube-api-access-jntbs\") pod \"dnsmasq-dns-64f7f48db9-mjw87\" (UID: \"30a1b29b-530b-4f64-acb7-a8471fc98fcb\") " pod="openstack/dnsmasq-dns-64f7f48db9-mjw87" Feb 18 09:16:35 crc kubenswrapper[4556]: I0218 09:16:35.546215 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/30a1b29b-530b-4f64-acb7-a8471fc98fcb-ovsdbserver-nb\") pod \"dnsmasq-dns-64f7f48db9-mjw87\" (UID: \"30a1b29b-530b-4f64-acb7-a8471fc98fcb\") " pod="openstack/dnsmasq-dns-64f7f48db9-mjw87" Feb 18 09:16:35 crc kubenswrapper[4556]: I0218 09:16:35.546766 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30a1b29b-530b-4f64-acb7-a8471fc98fcb-config\") pod \"dnsmasq-dns-64f7f48db9-mjw87\" (UID: \"30a1b29b-530b-4f64-acb7-a8471fc98fcb\") " pod="openstack/dnsmasq-dns-64f7f48db9-mjw87" Feb 18 09:16:35 crc kubenswrapper[4556]: I0218 09:16:35.547287 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/30a1b29b-530b-4f64-acb7-a8471fc98fcb-dns-svc\") pod \"dnsmasq-dns-64f7f48db9-mjw87\" (UID: \"30a1b29b-530b-4f64-acb7-a8471fc98fcb\") " pod="openstack/dnsmasq-dns-64f7f48db9-mjw87" Feb 18 09:16:35 crc kubenswrapper[4556]: I0218 09:16:35.563269 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jntbs\" (UniqueName: \"kubernetes.io/projected/30a1b29b-530b-4f64-acb7-a8471fc98fcb-kube-api-access-jntbs\") pod \"dnsmasq-dns-64f7f48db9-mjw87\" (UID: \"30a1b29b-530b-4f64-acb7-a8471fc98fcb\") " pod="openstack/dnsmasq-dns-64f7f48db9-mjw87" Feb 18 09:16:35 crc kubenswrapper[4556]: I0218 09:16:35.630925 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64f7f48db9-mjw87" Feb 18 09:16:36 crc kubenswrapper[4556]: I0218 09:16:36.377076 4556 scope.go:117] "RemoveContainer" containerID="0e8dce06934a1d8e8e642a478572d660aaf1657ba0d59076856e01e7fcdfd51c" Feb 18 09:16:37 crc kubenswrapper[4556]: I0218 09:16:37.449590 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f54874ffc-rb6x9" event={"ID":"78221461-c2ed-45d9-bd36-aa078d36d757","Type":"ContainerStarted","Data":"c86d1f7dfe664be938bec8e672fbb4c9394b560abae6451dfb04d8ee2c44ea24"} Feb 18 09:16:37 crc kubenswrapper[4556]: I0218 09:16:37.450130 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-f54874ffc-rb6x9" Feb 18 09:16:37 crc kubenswrapper[4556]: I0218 09:16:37.473356 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-f54874ffc-rb6x9" podStartSLOduration=7.856419613 podStartE2EDuration="17.473337917s" podCreationTimestamp="2026-02-18 09:16:20 +0000 UTC" firstStartedPulling="2026-02-18 09:16:21.521564588 +0000 UTC m=+738.538525567" lastFinishedPulling="2026-02-18 09:16:31.13848289 +0000 UTC m=+748.155443871" observedRunningTime="2026-02-18 09:16:37.467658546 +0000 UTC m=+754.484619526" watchObservedRunningTime="2026-02-18 09:16:37.473337917 +0000 UTC m=+754.490298898" Feb 18 09:16:37 crc kubenswrapper[4556]: I0218 09:16:37.858050 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 18 09:16:38 crc kubenswrapper[4556]: W0218 09:16:38.798674 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd7840643_68ec_4a2a_b6af_8a9730729077.slice/crio-21ae45a661426bf143c35df465c8a099fdbfe034ca06b6f060e057e7af0ea3e9 WatchSource:0}: Error finding container 21ae45a661426bf143c35df465c8a099fdbfe034ca06b6f060e057e7af0ea3e9: Status 404 returned error can't find the container with id 21ae45a661426bf143c35df465c8a099fdbfe034ca06b6f060e057e7af0ea3e9 Feb 18 09:16:39 crc kubenswrapper[4556]: I0218 09:16:39.182768 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64f7f48db9-mjw87"] Feb 18 09:16:39 crc kubenswrapper[4556]: W0218 09:16:39.343056 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod30a1b29b_530b_4f64_acb7_a8471fc98fcb.slice/crio-5bfbcf946c5e97e1f925409a0045b84af89e75a10d25cdad68731e7561fcb4c4 WatchSource:0}: Error finding container 5bfbcf946c5e97e1f925409a0045b84af89e75a10d25cdad68731e7561fcb4c4: Status 404 returned error can't find the container with id 5bfbcf946c5e97e1f925409a0045b84af89e75a10d25cdad68731e7561fcb4c4 Feb 18 09:16:39 crc kubenswrapper[4556]: I0218 09:16:39.467646 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64f7f48db9-mjw87" event={"ID":"30a1b29b-530b-4f64-acb7-a8471fc98fcb","Type":"ContainerStarted","Data":"5bfbcf946c5e97e1f925409a0045b84af89e75a10d25cdad68731e7561fcb4c4"} Feb 18 09:16:39 crc kubenswrapper[4556]: I0218 09:16:39.472661 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67ff45466c-wdkxp" event={"ID":"fa586b98-2a59-4673-a0bf-f7175f30d356","Type":"ContainerStarted","Data":"4a7ae2e397c41523ce14a310e7946b93b3f6d2798f69070e3aa1990efde95536"} Feb 18 09:16:39 crc kubenswrapper[4556]: I0218 09:16:39.472802 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-67ff45466c-wdkxp" podUID="fa586b98-2a59-4673-a0bf-f7175f30d356" containerName="dnsmasq-dns" containerID="cri-o://4a7ae2e397c41523ce14a310e7946b93b3f6d2798f69070e3aa1990efde95536" gracePeriod=10 Feb 18 09:16:39 crc kubenswrapper[4556]: I0218 09:16:39.473313 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-67ff45466c-wdkxp" Feb 18 09:16:39 crc kubenswrapper[4556]: I0218 09:16:39.476108 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"d7840643-68ec-4a2a-b6af-8a9730729077","Type":"ContainerStarted","Data":"21ae45a661426bf143c35df465c8a099fdbfe034ca06b6f060e057e7af0ea3e9"} Feb 18 09:16:39 crc kubenswrapper[4556]: I0218 09:16:39.500755 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-67ff45466c-wdkxp" podStartSLOduration=9.143491818 podStartE2EDuration="18.500740377s" podCreationTimestamp="2026-02-18 09:16:21 +0000 UTC" firstStartedPulling="2026-02-18 09:16:21.777638043 +0000 UTC m=+738.794599023" lastFinishedPulling="2026-02-18 09:16:31.134886602 +0000 UTC m=+748.151847582" observedRunningTime="2026-02-18 09:16:39.491564144 +0000 UTC m=+756.508525144" watchObservedRunningTime="2026-02-18 09:16:39.500740377 +0000 UTC m=+756.517701347" Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.449826 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67ff45466c-wdkxp" Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.504661 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-j7rpf" event={"ID":"66575dd2-efe2-4770-a9cd-7afb99e8566c","Type":"ContainerStarted","Data":"5689d83edc10e20ca5916bc6920a437330f4a7ba79d88285cb96cd1bc8094646"} Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.510011 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"11b77df6-67d9-4ddb-944b-2ad8b0fada78","Type":"ContainerStarted","Data":"44822dc7d2a57bf07ba75758cbcb009df513ad7efc8100b1e80fb0ea621bb5f6"} Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.511508 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"2c504c68-45e0-49dc-90f3-90b91e33551c","Type":"ContainerStarted","Data":"c17d45078745ffcf58fced4ab990cc3077f44e628505d2a9d2aef5dc32b68ddf"} Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.517186 4556 generic.go:334] "Generic (PLEG): container finished" podID="30a1b29b-530b-4f64-acb7-a8471fc98fcb" containerID="0ed084463a3e474d7452081234fd8a927fb6fb2988840005710c317a2d837f74" exitCode=0 Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.517261 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64f7f48db9-mjw87" event={"ID":"30a1b29b-530b-4f64-acb7-a8471fc98fcb","Type":"ContainerDied","Data":"0ed084463a3e474d7452081234fd8a927fb6fb2988840005710c317a2d837f74"} Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.526422 4556 generic.go:334] "Generic (PLEG): container finished" podID="fa586b98-2a59-4673-a0bf-f7175f30d356" containerID="4a7ae2e397c41523ce14a310e7946b93b3f6d2798f69070e3aa1990efde95536" exitCode=0 Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.526719 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67ff45466c-wdkxp" event={"ID":"fa586b98-2a59-4673-a0bf-f7175f30d356","Type":"ContainerDied","Data":"4a7ae2e397c41523ce14a310e7946b93b3f6d2798f69070e3aa1990efde95536"} Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.526802 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67ff45466c-wdkxp" event={"ID":"fa586b98-2a59-4673-a0bf-f7175f30d356","Type":"ContainerDied","Data":"031b6d98624d5f4f1ccbf8ef80d5c2c14a3af7a2b1d6eaa11caf0fb3954e993c"} Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.526880 4556 scope.go:117] "RemoveContainer" containerID="4a7ae2e397c41523ce14a310e7946b93b3f6d2798f69070e3aa1990efde95536" Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.527542 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67ff45466c-wdkxp" Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.527884 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-j7rpf" podStartSLOduration=3.048369998 podStartE2EDuration="8.527866525s" podCreationTimestamp="2026-02-18 09:16:32 +0000 UTC" firstStartedPulling="2026-02-18 09:16:33.933722603 +0000 UTC m=+750.950683584" lastFinishedPulling="2026-02-18 09:16:39.413219131 +0000 UTC m=+756.430180111" observedRunningTime="2026-02-18 09:16:40.517864224 +0000 UTC m=+757.534825203" watchObservedRunningTime="2026-02-18 09:16:40.527866525 +0000 UTC m=+757.544827506" Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.535819 4556 generic.go:334] "Generic (PLEG): container finished" podID="40c51577-de87-4bd8-be03-41221bc2f415" containerID="0c18112f01b6ff86cb75d679d034ec99af18e974ca08e0e02cbc10dcd396b838" exitCode=0 Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.535988 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-srnvn" event={"ID":"40c51577-de87-4bd8-be03-41221bc2f415","Type":"ContainerDied","Data":"0c18112f01b6ff86cb75d679d034ec99af18e974ca08e0e02cbc10dcd396b838"} Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.556794 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"68274946-b189-450c-a154-27059f411af3","Type":"ContainerStarted","Data":"27418c8117374eb93fb6f69511408726bd46d8184732c48f2d9c81b4492100f5"} Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.557297 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.563712 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"10b5d4cb-2beb-405d-83fc-30eb5fd4aaad","Type":"ContainerStarted","Data":"9b618cdaac602b43530c658ce250936cfc9fa45cb7708bd243f592a16163b81a"} Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.567670 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"95f33d13-a69d-4add-88a3-21126d52a0ed","Type":"ContainerStarted","Data":"4318c39a39db1314fa695847829c132f7eaf428f7ea492aabc0dd8c161ec63ff"} Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.568307 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.577959 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-6kkrw" event={"ID":"4906afef-c7e6-4597-a4a5-e9d758917e11","Type":"ContainerStarted","Data":"eaac26766d6356f6fb282146c0484713cf18e55999668b4f107eb50eb5830211"} Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.578523 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-6kkrw" Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.587929 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"03023501-772f-4798-a953-f56dee586cb1","Type":"ContainerStarted","Data":"7bf62d08939139f2182730d0a4d9b37f729f09b7b5649c1d303400e1b02489b7"} Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.604815 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-6kkrw" podStartSLOduration=3.572559727 podStartE2EDuration="10.604798642s" podCreationTimestamp="2026-02-18 09:16:30 +0000 UTC" firstStartedPulling="2026-02-18 09:16:31.746578277 +0000 UTC m=+748.763539258" lastFinishedPulling="2026-02-18 09:16:38.778817182 +0000 UTC m=+755.795778173" observedRunningTime="2026-02-18 09:16:40.599545875 +0000 UTC m=+757.616506855" watchObservedRunningTime="2026-02-18 09:16:40.604798642 +0000 UTC m=+757.621759622" Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.621923 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=9.519370995 podStartE2EDuration="16.621908442s" podCreationTimestamp="2026-02-18 09:16:24 +0000 UTC" firstStartedPulling="2026-02-18 09:16:31.676315783 +0000 UTC m=+748.693276763" lastFinishedPulling="2026-02-18 09:16:38.77885323 +0000 UTC m=+755.795814210" observedRunningTime="2026-02-18 09:16:40.61803027 +0000 UTC m=+757.634991250" watchObservedRunningTime="2026-02-18 09:16:40.621908442 +0000 UTC m=+757.638869412" Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.641019 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=8.065703391 podStartE2EDuration="13.641003991s" podCreationTimestamp="2026-02-18 09:16:27 +0000 UTC" firstStartedPulling="2026-02-18 09:16:31.670136418 +0000 UTC m=+748.687097398" lastFinishedPulling="2026-02-18 09:16:37.245437018 +0000 UTC m=+754.262397998" observedRunningTime="2026-02-18 09:16:40.633913535 +0000 UTC m=+757.650874515" watchObservedRunningTime="2026-02-18 09:16:40.641003991 +0000 UTC m=+757.657964971" Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.651909 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fa586b98-2a59-4673-a0bf-f7175f30d356-dns-svc\") pod \"fa586b98-2a59-4673-a0bf-f7175f30d356\" (UID: \"fa586b98-2a59-4673-a0bf-f7175f30d356\") " Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.652193 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa586b98-2a59-4673-a0bf-f7175f30d356-config\") pod \"fa586b98-2a59-4673-a0bf-f7175f30d356\" (UID: \"fa586b98-2a59-4673-a0bf-f7175f30d356\") " Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.652227 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wmrpp\" (UniqueName: \"kubernetes.io/projected/fa586b98-2a59-4673-a0bf-f7175f30d356-kube-api-access-wmrpp\") pod \"fa586b98-2a59-4673-a0bf-f7175f30d356\" (UID: \"fa586b98-2a59-4673-a0bf-f7175f30d356\") " Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.668278 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa586b98-2a59-4673-a0bf-f7175f30d356-kube-api-access-wmrpp" (OuterVolumeSpecName: "kube-api-access-wmrpp") pod "fa586b98-2a59-4673-a0bf-f7175f30d356" (UID: "fa586b98-2a59-4673-a0bf-f7175f30d356"). InnerVolumeSpecName "kube-api-access-wmrpp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.676876 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=4.270516645 podStartE2EDuration="9.676862135s" podCreationTimestamp="2026-02-18 09:16:31 +0000 UTC" firstStartedPulling="2026-02-18 09:16:33.949764167 +0000 UTC m=+750.966725147" lastFinishedPulling="2026-02-18 09:16:39.356109658 +0000 UTC m=+756.373070637" observedRunningTime="2026-02-18 09:16:40.663650503 +0000 UTC m=+757.680611484" watchObservedRunningTime="2026-02-18 09:16:40.676862135 +0000 UTC m=+757.693823116" Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.681410 4556 scope.go:117] "RemoveContainer" containerID="89933aa5c0aaa3e505249d03754c47c7de8ff057003a4042350fac5d0d7e5eea" Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.693623 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa586b98-2a59-4673-a0bf-f7175f30d356-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "fa586b98-2a59-4673-a0bf-f7175f30d356" (UID: "fa586b98-2a59-4673-a0bf-f7175f30d356"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.725787 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa586b98-2a59-4673-a0bf-f7175f30d356-config" (OuterVolumeSpecName: "config") pod "fa586b98-2a59-4673-a0bf-f7175f30d356" (UID: "fa586b98-2a59-4673-a0bf-f7175f30d356"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.732645 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f54874ffc-rb6x9"] Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.732886 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-f54874ffc-rb6x9" podUID="78221461-c2ed-45d9-bd36-aa078d36d757" containerName="dnsmasq-dns" containerID="cri-o://c86d1f7dfe664be938bec8e672fbb4c9394b560abae6451dfb04d8ee2c44ea24" gracePeriod=10 Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.737250 4556 scope.go:117] "RemoveContainer" containerID="4a7ae2e397c41523ce14a310e7946b93b3f6d2798f69070e3aa1990efde95536" Feb 18 09:16:40 crc kubenswrapper[4556]: E0218 09:16:40.737746 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a7ae2e397c41523ce14a310e7946b93b3f6d2798f69070e3aa1990efde95536\": container with ID starting with 4a7ae2e397c41523ce14a310e7946b93b3f6d2798f69070e3aa1990efde95536 not found: ID does not exist" containerID="4a7ae2e397c41523ce14a310e7946b93b3f6d2798f69070e3aa1990efde95536" Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.737782 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a7ae2e397c41523ce14a310e7946b93b3f6d2798f69070e3aa1990efde95536"} err="failed to get container status \"4a7ae2e397c41523ce14a310e7946b93b3f6d2798f69070e3aa1990efde95536\": rpc error: code = NotFound desc = could not find container \"4a7ae2e397c41523ce14a310e7946b93b3f6d2798f69070e3aa1990efde95536\": container with ID starting with 4a7ae2e397c41523ce14a310e7946b93b3f6d2798f69070e3aa1990efde95536 not found: ID does not exist" Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.737803 4556 scope.go:117] "RemoveContainer" containerID="89933aa5c0aaa3e505249d03754c47c7de8ff057003a4042350fac5d0d7e5eea" Feb 18 09:16:40 crc kubenswrapper[4556]: E0218 09:16:40.738457 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"89933aa5c0aaa3e505249d03754c47c7de8ff057003a4042350fac5d0d7e5eea\": container with ID starting with 89933aa5c0aaa3e505249d03754c47c7de8ff057003a4042350fac5d0d7e5eea not found: ID does not exist" containerID="89933aa5c0aaa3e505249d03754c47c7de8ff057003a4042350fac5d0d7e5eea" Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.738483 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89933aa5c0aaa3e505249d03754c47c7de8ff057003a4042350fac5d0d7e5eea"} err="failed to get container status \"89933aa5c0aaa3e505249d03754c47c7de8ff057003a4042350fac5d0d7e5eea\": rpc error: code = NotFound desc = could not find container \"89933aa5c0aaa3e505249d03754c47c7de8ff057003a4042350fac5d0d7e5eea\": container with ID starting with 89933aa5c0aaa3e505249d03754c47c7de8ff057003a4042350fac5d0d7e5eea not found: ID does not exist" Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.756790 4556 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa586b98-2a59-4673-a0bf-f7175f30d356-config\") on node \"crc\" DevicePath \"\"" Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.756815 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wmrpp\" (UniqueName: \"kubernetes.io/projected/fa586b98-2a59-4673-a0bf-f7175f30d356-kube-api-access-wmrpp\") on node \"crc\" DevicePath \"\"" Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.756824 4556 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fa586b98-2a59-4673-a0bf-f7175f30d356-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.757289 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-56df986d9c-kdzbf"] Feb 18 09:16:40 crc kubenswrapper[4556]: E0218 09:16:40.757615 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa586b98-2a59-4673-a0bf-f7175f30d356" containerName="dnsmasq-dns" Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.757638 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa586b98-2a59-4673-a0bf-f7175f30d356" containerName="dnsmasq-dns" Feb 18 09:16:40 crc kubenswrapper[4556]: E0218 09:16:40.757647 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa586b98-2a59-4673-a0bf-f7175f30d356" containerName="init" Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.757653 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa586b98-2a59-4673-a0bf-f7175f30d356" containerName="init" Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.757827 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa586b98-2a59-4673-a0bf-f7175f30d356" containerName="dnsmasq-dns" Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.758675 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df986d9c-kdzbf" Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.761224 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.771539 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df986d9c-kdzbf"] Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.857892 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsr9v\" (UniqueName: \"kubernetes.io/projected/cc7299fa-91af-4d80-957d-7a849a12e76c-kube-api-access-jsr9v\") pod \"dnsmasq-dns-56df986d9c-kdzbf\" (UID: \"cc7299fa-91af-4d80-957d-7a849a12e76c\") " pod="openstack/dnsmasq-dns-56df986d9c-kdzbf" Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.858173 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc7299fa-91af-4d80-957d-7a849a12e76c-config\") pod \"dnsmasq-dns-56df986d9c-kdzbf\" (UID: \"cc7299fa-91af-4d80-957d-7a849a12e76c\") " pod="openstack/dnsmasq-dns-56df986d9c-kdzbf" Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.858250 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cc7299fa-91af-4d80-957d-7a849a12e76c-ovsdbserver-sb\") pod \"dnsmasq-dns-56df986d9c-kdzbf\" (UID: \"cc7299fa-91af-4d80-957d-7a849a12e76c\") " pod="openstack/dnsmasq-dns-56df986d9c-kdzbf" Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.858341 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cc7299fa-91af-4d80-957d-7a849a12e76c-dns-svc\") pod \"dnsmasq-dns-56df986d9c-kdzbf\" (UID: \"cc7299fa-91af-4d80-957d-7a849a12e76c\") " pod="openstack/dnsmasq-dns-56df986d9c-kdzbf" Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.858402 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cc7299fa-91af-4d80-957d-7a849a12e76c-ovsdbserver-nb\") pod \"dnsmasq-dns-56df986d9c-kdzbf\" (UID: \"cc7299fa-91af-4d80-957d-7a849a12e76c\") " pod="openstack/dnsmasq-dns-56df986d9c-kdzbf" Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.933718 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67ff45466c-wdkxp"] Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.939659 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-67ff45466c-wdkxp"] Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.960893 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cc7299fa-91af-4d80-957d-7a849a12e76c-dns-svc\") pod \"dnsmasq-dns-56df986d9c-kdzbf\" (UID: \"cc7299fa-91af-4d80-957d-7a849a12e76c\") " pod="openstack/dnsmasq-dns-56df986d9c-kdzbf" Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.960985 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cc7299fa-91af-4d80-957d-7a849a12e76c-ovsdbserver-nb\") pod \"dnsmasq-dns-56df986d9c-kdzbf\" (UID: \"cc7299fa-91af-4d80-957d-7a849a12e76c\") " pod="openstack/dnsmasq-dns-56df986d9c-kdzbf" Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.961073 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsr9v\" (UniqueName: \"kubernetes.io/projected/cc7299fa-91af-4d80-957d-7a849a12e76c-kube-api-access-jsr9v\") pod \"dnsmasq-dns-56df986d9c-kdzbf\" (UID: \"cc7299fa-91af-4d80-957d-7a849a12e76c\") " pod="openstack/dnsmasq-dns-56df986d9c-kdzbf" Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.961091 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc7299fa-91af-4d80-957d-7a849a12e76c-config\") pod \"dnsmasq-dns-56df986d9c-kdzbf\" (UID: \"cc7299fa-91af-4d80-957d-7a849a12e76c\") " pod="openstack/dnsmasq-dns-56df986d9c-kdzbf" Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.961201 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cc7299fa-91af-4d80-957d-7a849a12e76c-ovsdbserver-sb\") pod \"dnsmasq-dns-56df986d9c-kdzbf\" (UID: \"cc7299fa-91af-4d80-957d-7a849a12e76c\") " pod="openstack/dnsmasq-dns-56df986d9c-kdzbf" Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.962783 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cc7299fa-91af-4d80-957d-7a849a12e76c-ovsdbserver-nb\") pod \"dnsmasq-dns-56df986d9c-kdzbf\" (UID: \"cc7299fa-91af-4d80-957d-7a849a12e76c\") " pod="openstack/dnsmasq-dns-56df986d9c-kdzbf" Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.963385 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cc7299fa-91af-4d80-957d-7a849a12e76c-dns-svc\") pod \"dnsmasq-dns-56df986d9c-kdzbf\" (UID: \"cc7299fa-91af-4d80-957d-7a849a12e76c\") " pod="openstack/dnsmasq-dns-56df986d9c-kdzbf" Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.963405 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc7299fa-91af-4d80-957d-7a849a12e76c-config\") pod \"dnsmasq-dns-56df986d9c-kdzbf\" (UID: \"cc7299fa-91af-4d80-957d-7a849a12e76c\") " pod="openstack/dnsmasq-dns-56df986d9c-kdzbf" Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.963939 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cc7299fa-91af-4d80-957d-7a849a12e76c-ovsdbserver-sb\") pod \"dnsmasq-dns-56df986d9c-kdzbf\" (UID: \"cc7299fa-91af-4d80-957d-7a849a12e76c\") " pod="openstack/dnsmasq-dns-56df986d9c-kdzbf" Feb 18 09:16:40 crc kubenswrapper[4556]: I0218 09:16:40.991871 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsr9v\" (UniqueName: \"kubernetes.io/projected/cc7299fa-91af-4d80-957d-7a849a12e76c-kube-api-access-jsr9v\") pod \"dnsmasq-dns-56df986d9c-kdzbf\" (UID: \"cc7299fa-91af-4d80-957d-7a849a12e76c\") " pod="openstack/dnsmasq-dns-56df986d9c-kdzbf" Feb 18 09:16:41 crc kubenswrapper[4556]: I0218 09:16:41.216559 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df986d9c-kdzbf" Feb 18 09:16:41 crc kubenswrapper[4556]: I0218 09:16:41.235675 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f54874ffc-rb6x9" Feb 18 09:16:41 crc kubenswrapper[4556]: I0218 09:16:41.290927 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa586b98-2a59-4673-a0bf-f7175f30d356" path="/var/lib/kubelet/pods/fa586b98-2a59-4673-a0bf-f7175f30d356/volumes" Feb 18 09:16:41 crc kubenswrapper[4556]: I0218 09:16:41.368478 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78221461-c2ed-45d9-bd36-aa078d36d757-config\") pod \"78221461-c2ed-45d9-bd36-aa078d36d757\" (UID: \"78221461-c2ed-45d9-bd36-aa078d36d757\") " Feb 18 09:16:41 crc kubenswrapper[4556]: I0218 09:16:41.368949 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pph7d\" (UniqueName: \"kubernetes.io/projected/78221461-c2ed-45d9-bd36-aa078d36d757-kube-api-access-pph7d\") pod \"78221461-c2ed-45d9-bd36-aa078d36d757\" (UID: \"78221461-c2ed-45d9-bd36-aa078d36d757\") " Feb 18 09:16:41 crc kubenswrapper[4556]: I0218 09:16:41.369057 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/78221461-c2ed-45d9-bd36-aa078d36d757-dns-svc\") pod \"78221461-c2ed-45d9-bd36-aa078d36d757\" (UID: \"78221461-c2ed-45d9-bd36-aa078d36d757\") " Feb 18 09:16:41 crc kubenswrapper[4556]: I0218 09:16:41.374447 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78221461-c2ed-45d9-bd36-aa078d36d757-kube-api-access-pph7d" (OuterVolumeSpecName: "kube-api-access-pph7d") pod "78221461-c2ed-45d9-bd36-aa078d36d757" (UID: "78221461-c2ed-45d9-bd36-aa078d36d757"). InnerVolumeSpecName "kube-api-access-pph7d". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:16:41 crc kubenswrapper[4556]: I0218 09:16:41.396672 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78221461-c2ed-45d9-bd36-aa078d36d757-config" (OuterVolumeSpecName: "config") pod "78221461-c2ed-45d9-bd36-aa078d36d757" (UID: "78221461-c2ed-45d9-bd36-aa078d36d757"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:16:41 crc kubenswrapper[4556]: I0218 09:16:41.431931 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78221461-c2ed-45d9-bd36-aa078d36d757-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "78221461-c2ed-45d9-bd36-aa078d36d757" (UID: "78221461-c2ed-45d9-bd36-aa078d36d757"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:16:41 crc kubenswrapper[4556]: I0218 09:16:41.472471 4556 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78221461-c2ed-45d9-bd36-aa078d36d757-config\") on node \"crc\" DevicePath \"\"" Feb 18 09:16:41 crc kubenswrapper[4556]: I0218 09:16:41.472505 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pph7d\" (UniqueName: \"kubernetes.io/projected/78221461-c2ed-45d9-bd36-aa078d36d757-kube-api-access-pph7d\") on node \"crc\" DevicePath \"\"" Feb 18 09:16:41 crc kubenswrapper[4556]: I0218 09:16:41.472517 4556 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/78221461-c2ed-45d9-bd36-aa078d36d757-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 18 09:16:41 crc kubenswrapper[4556]: I0218 09:16:41.596823 4556 generic.go:334] "Generic (PLEG): container finished" podID="78221461-c2ed-45d9-bd36-aa078d36d757" containerID="c86d1f7dfe664be938bec8e672fbb4c9394b560abae6451dfb04d8ee2c44ea24" exitCode=0 Feb 18 09:16:41 crc kubenswrapper[4556]: I0218 09:16:41.596905 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f54874ffc-rb6x9" Feb 18 09:16:41 crc kubenswrapper[4556]: I0218 09:16:41.596920 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f54874ffc-rb6x9" event={"ID":"78221461-c2ed-45d9-bd36-aa078d36d757","Type":"ContainerDied","Data":"c86d1f7dfe664be938bec8e672fbb4c9394b560abae6451dfb04d8ee2c44ea24"} Feb 18 09:16:41 crc kubenswrapper[4556]: I0218 09:16:41.597264 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f54874ffc-rb6x9" event={"ID":"78221461-c2ed-45d9-bd36-aa078d36d757","Type":"ContainerDied","Data":"5d38656375670fe555a05d1c74743fce293b4370622cd92c9cce18bd0ace0a44"} Feb 18 09:16:41 crc kubenswrapper[4556]: I0218 09:16:41.597286 4556 scope.go:117] "RemoveContainer" containerID="c86d1f7dfe664be938bec8e672fbb4c9394b560abae6451dfb04d8ee2c44ea24" Feb 18 09:16:41 crc kubenswrapper[4556]: I0218 09:16:41.599209 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"10b5d4cb-2beb-405d-83fc-30eb5fd4aaad","Type":"ContainerStarted","Data":"e60b695038ed8157fdd9323c32d6234ab26860be944777026cad4fc0f0ef167b"} Feb 18 09:16:41 crc kubenswrapper[4556]: I0218 09:16:41.601832 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64f7f48db9-mjw87" event={"ID":"30a1b29b-530b-4f64-acb7-a8471fc98fcb","Type":"ContainerStarted","Data":"833d74497abfc08cdfa0f0c32d15a296083d4ae387fbc6542d6c8810ddc63fa8"} Feb 18 09:16:41 crc kubenswrapper[4556]: I0218 09:16:41.601944 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-64f7f48db9-mjw87" Feb 18 09:16:41 crc kubenswrapper[4556]: I0218 09:16:41.605141 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-srnvn" event={"ID":"40c51577-de87-4bd8-be03-41221bc2f415","Type":"ContainerStarted","Data":"d4db4d4cb56ec3b96f9161f404bf149b765663fa0426e6afcc983f89aa243c24"} Feb 18 09:16:41 crc kubenswrapper[4556]: I0218 09:16:41.605186 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-srnvn" event={"ID":"40c51577-de87-4bd8-be03-41221bc2f415","Type":"ContainerStarted","Data":"2b42f7db4b7e1de6bc66a46cf2b422080cfe5fc40bba0e7b51a19bb917af0ca4"} Feb 18 09:16:41 crc kubenswrapper[4556]: I0218 09:16:41.605301 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-srnvn" Feb 18 09:16:41 crc kubenswrapper[4556]: I0218 09:16:41.605325 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-srnvn" Feb 18 09:16:41 crc kubenswrapper[4556]: I0218 09:16:41.608297 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1d7961e1-6f96-4707-907d-1fa0c30641e2","Type":"ContainerStarted","Data":"98251fd58b2a29ac981c57fb4437b319daf966b2b07731b525037e5dccf051cf"} Feb 18 09:16:41 crc kubenswrapper[4556]: I0218 09:16:41.611023 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"d7840643-68ec-4a2a-b6af-8a9730729077","Type":"ContainerStarted","Data":"2b2d1d8119df62a19391914f61a0fb31c01b8476a48298e12b43e9bf826c0059"} Feb 18 09:16:41 crc kubenswrapper[4556]: I0218 09:16:41.611046 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"d7840643-68ec-4a2a-b6af-8a9730729077","Type":"ContainerStarted","Data":"7477eeef5440f4dcf12d4a5c4dfcc1d5bb8f3a870e93c5a7464fb06e8845e922"} Feb 18 09:16:41 crc kubenswrapper[4556]: I0218 09:16:41.614511 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df986d9c-kdzbf"] Feb 18 09:16:41 crc kubenswrapper[4556]: I0218 09:16:41.615971 4556 scope.go:117] "RemoveContainer" containerID="d24f089f65c98b0bf17cede561b911d8ca46377e4d023b3d5a9df6c7ef7f8e42" Feb 18 09:16:41 crc kubenswrapper[4556]: W0218 09:16:41.624985 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcc7299fa_91af_4d80_957d_7a849a12e76c.slice/crio-76029ec76a7668e60ecfd1c097ec379a82c468a7f72edf7ff95f62468d2600b1 WatchSource:0}: Error finding container 76029ec76a7668e60ecfd1c097ec379a82c468a7f72edf7ff95f62468d2600b1: Status 404 returned error can't find the container with id 76029ec76a7668e60ecfd1c097ec379a82c468a7f72edf7ff95f62468d2600b1 Feb 18 09:16:41 crc kubenswrapper[4556]: I0218 09:16:41.625829 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-64f7f48db9-mjw87" podStartSLOduration=6.625819995 podStartE2EDuration="6.625819995s" podCreationTimestamp="2026-02-18 09:16:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:16:41.620361531 +0000 UTC m=+758.637322511" watchObservedRunningTime="2026-02-18 09:16:41.625819995 +0000 UTC m=+758.642780975" Feb 18 09:16:41 crc kubenswrapper[4556]: I0218 09:16:41.637138 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=6.225133426 podStartE2EDuration="8.637129977s" podCreationTimestamp="2026-02-18 09:16:33 +0000 UTC" firstStartedPulling="2026-02-18 09:16:38.804342851 +0000 UTC m=+755.821303820" lastFinishedPulling="2026-02-18 09:16:41.216339392 +0000 UTC m=+758.233300371" observedRunningTime="2026-02-18 09:16:41.634811068 +0000 UTC m=+758.651772048" watchObservedRunningTime="2026-02-18 09:16:41.637129977 +0000 UTC m=+758.654090956" Feb 18 09:16:41 crc kubenswrapper[4556]: I0218 09:16:41.649092 4556 scope.go:117] "RemoveContainer" containerID="c86d1f7dfe664be938bec8e672fbb4c9394b560abae6451dfb04d8ee2c44ea24" Feb 18 09:16:41 crc kubenswrapper[4556]: I0218 09:16:41.650166 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-srnvn" podStartSLOduration=6.020778188 podStartE2EDuration="11.650137683s" podCreationTimestamp="2026-02-18 09:16:30 +0000 UTC" firstStartedPulling="2026-02-18 09:16:31.787702824 +0000 UTC m=+748.804663804" lastFinishedPulling="2026-02-18 09:16:37.417062318 +0000 UTC m=+754.434023299" observedRunningTime="2026-02-18 09:16:41.64883331 +0000 UTC m=+758.665794290" watchObservedRunningTime="2026-02-18 09:16:41.650137683 +0000 UTC m=+758.667098663" Feb 18 09:16:41 crc kubenswrapper[4556]: E0218 09:16:41.650465 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c86d1f7dfe664be938bec8e672fbb4c9394b560abae6451dfb04d8ee2c44ea24\": container with ID starting with c86d1f7dfe664be938bec8e672fbb4c9394b560abae6451dfb04d8ee2c44ea24 not found: ID does not exist" containerID="c86d1f7dfe664be938bec8e672fbb4c9394b560abae6451dfb04d8ee2c44ea24" Feb 18 09:16:41 crc kubenswrapper[4556]: I0218 09:16:41.650503 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c86d1f7dfe664be938bec8e672fbb4c9394b560abae6451dfb04d8ee2c44ea24"} err="failed to get container status \"c86d1f7dfe664be938bec8e672fbb4c9394b560abae6451dfb04d8ee2c44ea24\": rpc error: code = NotFound desc = could not find container \"c86d1f7dfe664be938bec8e672fbb4c9394b560abae6451dfb04d8ee2c44ea24\": container with ID starting with c86d1f7dfe664be938bec8e672fbb4c9394b560abae6451dfb04d8ee2c44ea24 not found: ID does not exist" Feb 18 09:16:41 crc kubenswrapper[4556]: I0218 09:16:41.650525 4556 scope.go:117] "RemoveContainer" containerID="d24f089f65c98b0bf17cede561b911d8ca46377e4d023b3d5a9df6c7ef7f8e42" Feb 18 09:16:41 crc kubenswrapper[4556]: E0218 09:16:41.650798 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d24f089f65c98b0bf17cede561b911d8ca46377e4d023b3d5a9df6c7ef7f8e42\": container with ID starting with d24f089f65c98b0bf17cede561b911d8ca46377e4d023b3d5a9df6c7ef7f8e42 not found: ID does not exist" containerID="d24f089f65c98b0bf17cede561b911d8ca46377e4d023b3d5a9df6c7ef7f8e42" Feb 18 09:16:41 crc kubenswrapper[4556]: I0218 09:16:41.650845 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d24f089f65c98b0bf17cede561b911d8ca46377e4d023b3d5a9df6c7ef7f8e42"} err="failed to get container status \"d24f089f65c98b0bf17cede561b911d8ca46377e4d023b3d5a9df6c7ef7f8e42\": rpc error: code = NotFound desc = could not find container \"d24f089f65c98b0bf17cede561b911d8ca46377e4d023b3d5a9df6c7ef7f8e42\": container with ID starting with d24f089f65c98b0bf17cede561b911d8ca46377e4d023b3d5a9df6c7ef7f8e42 not found: ID does not exist" Feb 18 09:16:41 crc kubenswrapper[4556]: I0218 09:16:41.660440 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f54874ffc-rb6x9"] Feb 18 09:16:41 crc kubenswrapper[4556]: I0218 09:16:41.663825 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-f54874ffc-rb6x9"] Feb 18 09:16:41 crc kubenswrapper[4556]: I0218 09:16:41.816241 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Feb 18 09:16:42 crc kubenswrapper[4556]: I0218 09:16:42.624179 4556 generic.go:334] "Generic (PLEG): container finished" podID="cc7299fa-91af-4d80-957d-7a849a12e76c" containerID="0e00134d256432d75bb3b13230f55719bdd3db03254837a9dfd979e4e3b6947a" exitCode=0 Feb 18 09:16:42 crc kubenswrapper[4556]: I0218 09:16:42.624238 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df986d9c-kdzbf" event={"ID":"cc7299fa-91af-4d80-957d-7a849a12e76c","Type":"ContainerDied","Data":"0e00134d256432d75bb3b13230f55719bdd3db03254837a9dfd979e4e3b6947a"} Feb 18 09:16:42 crc kubenswrapper[4556]: I0218 09:16:42.624265 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df986d9c-kdzbf" event={"ID":"cc7299fa-91af-4d80-957d-7a849a12e76c","Type":"ContainerStarted","Data":"76029ec76a7668e60ecfd1c097ec379a82c468a7f72edf7ff95f62468d2600b1"} Feb 18 09:16:42 crc kubenswrapper[4556]: I0218 09:16:42.815846 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Feb 18 09:16:43 crc kubenswrapper[4556]: I0218 09:16:43.294055 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78221461-c2ed-45d9-bd36-aa078d36d757" path="/var/lib/kubelet/pods/78221461-c2ed-45d9-bd36-aa078d36d757/volumes" Feb 18 09:16:43 crc kubenswrapper[4556]: I0218 09:16:43.630996 4556 generic.go:334] "Generic (PLEG): container finished" podID="11b77df6-67d9-4ddb-944b-2ad8b0fada78" containerID="44822dc7d2a57bf07ba75758cbcb009df513ad7efc8100b1e80fb0ea621bb5f6" exitCode=0 Feb 18 09:16:43 crc kubenswrapper[4556]: I0218 09:16:43.631065 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"11b77df6-67d9-4ddb-944b-2ad8b0fada78","Type":"ContainerDied","Data":"44822dc7d2a57bf07ba75758cbcb009df513ad7efc8100b1e80fb0ea621bb5f6"} Feb 18 09:16:43 crc kubenswrapper[4556]: I0218 09:16:43.636457 4556 generic.go:334] "Generic (PLEG): container finished" podID="2c504c68-45e0-49dc-90f3-90b91e33551c" containerID="c17d45078745ffcf58fced4ab990cc3077f44e628505d2a9d2aef5dc32b68ddf" exitCode=0 Feb 18 09:16:43 crc kubenswrapper[4556]: I0218 09:16:43.636524 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"2c504c68-45e0-49dc-90f3-90b91e33551c","Type":"ContainerDied","Data":"c17d45078745ffcf58fced4ab990cc3077f44e628505d2a9d2aef5dc32b68ddf"} Feb 18 09:16:43 crc kubenswrapper[4556]: I0218 09:16:43.639361 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df986d9c-kdzbf" event={"ID":"cc7299fa-91af-4d80-957d-7a849a12e76c","Type":"ContainerStarted","Data":"274f996804719440b04e23c51db947fad7337fbbecee12fc1ac0fcb7489aa6af"} Feb 18 09:16:43 crc kubenswrapper[4556]: I0218 09:16:43.639451 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-56df986d9c-kdzbf" Feb 18 09:16:43 crc kubenswrapper[4556]: I0218 09:16:43.694485 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-56df986d9c-kdzbf" podStartSLOduration=3.694423746 podStartE2EDuration="3.694423746s" podCreationTimestamp="2026-02-18 09:16:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:16:43.679955963 +0000 UTC m=+760.696916943" watchObservedRunningTime="2026-02-18 09:16:43.694423746 +0000 UTC m=+760.711384716" Feb 18 09:16:44 crc kubenswrapper[4556]: I0218 09:16:44.185204 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Feb 18 09:16:44 crc kubenswrapper[4556]: I0218 09:16:44.212591 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Feb 18 09:16:44 crc kubenswrapper[4556]: I0218 09:16:44.645337 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"11b77df6-67d9-4ddb-944b-2ad8b0fada78","Type":"ContainerStarted","Data":"da67bc9d9ffd7eb3ac00becf0b3814624c556b396873e940dccd6e882a231d11"} Feb 18 09:16:44 crc kubenswrapper[4556]: I0218 09:16:44.647260 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"2c504c68-45e0-49dc-90f3-90b91e33551c","Type":"ContainerStarted","Data":"a40d19bf8ab46aa5723ddd411440a514bebfd2282724421278400ac7f698a5c6"} Feb 18 09:16:44 crc kubenswrapper[4556]: I0218 09:16:44.647331 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Feb 18 09:16:44 crc kubenswrapper[4556]: I0218 09:16:44.662144 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=13.903025962 podStartE2EDuration="21.662133697s" podCreationTimestamp="2026-02-18 09:16:23 +0000 UTC" firstStartedPulling="2026-02-18 09:16:31.608383788 +0000 UTC m=+748.625344769" lastFinishedPulling="2026-02-18 09:16:39.367491524 +0000 UTC m=+756.384452504" observedRunningTime="2026-02-18 09:16:44.660069129 +0000 UTC m=+761.677030109" watchObservedRunningTime="2026-02-18 09:16:44.662133697 +0000 UTC m=+761.679094676" Feb 18 09:16:44 crc kubenswrapper[4556]: I0218 09:16:44.677690 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=14.897370975 podStartE2EDuration="22.677671329s" podCreationTimestamp="2026-02-18 09:16:22 +0000 UTC" firstStartedPulling="2026-02-18 09:16:31.606701312 +0000 UTC m=+748.623662293" lastFinishedPulling="2026-02-18 09:16:39.387001666 +0000 UTC m=+756.403962647" observedRunningTime="2026-02-18 09:16:44.673555288 +0000 UTC m=+761.690516279" watchObservedRunningTime="2026-02-18 09:16:44.677671329 +0000 UTC m=+761.694632309" Feb 18 09:16:44 crc kubenswrapper[4556]: I0218 09:16:44.846138 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Feb 18 09:16:45 crc kubenswrapper[4556]: I0218 09:16:45.149723 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Feb 18 09:16:45 crc kubenswrapper[4556]: I0218 09:16:45.245080 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Feb 18 09:16:45 crc kubenswrapper[4556]: I0218 09:16:45.245180 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Feb 18 09:16:45 crc kubenswrapper[4556]: I0218 09:16:45.632854 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-64f7f48db9-mjw87" Feb 18 09:16:45 crc kubenswrapper[4556]: I0218 09:16:45.698513 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Feb 18 09:16:47 crc kubenswrapper[4556]: I0218 09:16:47.299634 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Feb 18 09:16:47 crc kubenswrapper[4556]: I0218 09:16:47.495226 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Feb 18 09:16:47 crc kubenswrapper[4556]: I0218 09:16:47.531328 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df986d9c-kdzbf"] Feb 18 09:16:47 crc kubenswrapper[4556]: I0218 09:16:47.531588 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-56df986d9c-kdzbf" podUID="cc7299fa-91af-4d80-957d-7a849a12e76c" containerName="dnsmasq-dns" containerID="cri-o://274f996804719440b04e23c51db947fad7337fbbecee12fc1ac0fcb7489aa6af" gracePeriod=10 Feb 18 09:16:47 crc kubenswrapper[4556]: I0218 09:16:47.535495 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-56df986d9c-kdzbf" Feb 18 09:16:47 crc kubenswrapper[4556]: I0218 09:16:47.573526 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-66b577f8c-9dkbr"] Feb 18 09:16:47 crc kubenswrapper[4556]: E0218 09:16:47.579790 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78221461-c2ed-45d9-bd36-aa078d36d757" containerName="dnsmasq-dns" Feb 18 09:16:47 crc kubenswrapper[4556]: I0218 09:16:47.579822 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="78221461-c2ed-45d9-bd36-aa078d36d757" containerName="dnsmasq-dns" Feb 18 09:16:47 crc kubenswrapper[4556]: E0218 09:16:47.579865 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78221461-c2ed-45d9-bd36-aa078d36d757" containerName="init" Feb 18 09:16:47 crc kubenswrapper[4556]: I0218 09:16:47.579872 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="78221461-c2ed-45d9-bd36-aa078d36d757" containerName="init" Feb 18 09:16:47 crc kubenswrapper[4556]: I0218 09:16:47.580112 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="78221461-c2ed-45d9-bd36-aa078d36d757" containerName="dnsmasq-dns" Feb 18 09:16:47 crc kubenswrapper[4556]: I0218 09:16:47.583853 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66b577f8c-9dkbr" Feb 18 09:16:47 crc kubenswrapper[4556]: I0218 09:16:47.593203 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-66b577f8c-9dkbr"] Feb 18 09:16:47 crc kubenswrapper[4556]: I0218 09:16:47.680913 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f-config\") pod \"dnsmasq-dns-66b577f8c-9dkbr\" (UID: \"b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f\") " pod="openstack/dnsmasq-dns-66b577f8c-9dkbr" Feb 18 09:16:47 crc kubenswrapper[4556]: I0218 09:16:47.681002 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f-ovsdbserver-sb\") pod \"dnsmasq-dns-66b577f8c-9dkbr\" (UID: \"b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f\") " pod="openstack/dnsmasq-dns-66b577f8c-9dkbr" Feb 18 09:16:47 crc kubenswrapper[4556]: I0218 09:16:47.681076 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6kpfc\" (UniqueName: \"kubernetes.io/projected/b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f-kube-api-access-6kpfc\") pod \"dnsmasq-dns-66b577f8c-9dkbr\" (UID: \"b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f\") " pod="openstack/dnsmasq-dns-66b577f8c-9dkbr" Feb 18 09:16:47 crc kubenswrapper[4556]: I0218 09:16:47.681123 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f-ovsdbserver-nb\") pod \"dnsmasq-dns-66b577f8c-9dkbr\" (UID: \"b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f\") " pod="openstack/dnsmasq-dns-66b577f8c-9dkbr" Feb 18 09:16:47 crc kubenswrapper[4556]: I0218 09:16:47.681147 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f-dns-svc\") pod \"dnsmasq-dns-66b577f8c-9dkbr\" (UID: \"b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f\") " pod="openstack/dnsmasq-dns-66b577f8c-9dkbr" Feb 18 09:16:47 crc kubenswrapper[4556]: I0218 09:16:47.684834 4556 generic.go:334] "Generic (PLEG): container finished" podID="cc7299fa-91af-4d80-957d-7a849a12e76c" containerID="274f996804719440b04e23c51db947fad7337fbbecee12fc1ac0fcb7489aa6af" exitCode=0 Feb 18 09:16:47 crc kubenswrapper[4556]: I0218 09:16:47.685010 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df986d9c-kdzbf" event={"ID":"cc7299fa-91af-4d80-957d-7a849a12e76c","Type":"ContainerDied","Data":"274f996804719440b04e23c51db947fad7337fbbecee12fc1ac0fcb7489aa6af"} Feb 18 09:16:47 crc kubenswrapper[4556]: I0218 09:16:47.775809 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Feb 18 09:16:47 crc kubenswrapper[4556]: I0218 09:16:47.784240 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6kpfc\" (UniqueName: \"kubernetes.io/projected/b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f-kube-api-access-6kpfc\") pod \"dnsmasq-dns-66b577f8c-9dkbr\" (UID: \"b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f\") " pod="openstack/dnsmasq-dns-66b577f8c-9dkbr" Feb 18 09:16:47 crc kubenswrapper[4556]: I0218 09:16:47.784515 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f-ovsdbserver-nb\") pod \"dnsmasq-dns-66b577f8c-9dkbr\" (UID: \"b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f\") " pod="openstack/dnsmasq-dns-66b577f8c-9dkbr" Feb 18 09:16:47 crc kubenswrapper[4556]: I0218 09:16:47.784551 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f-dns-svc\") pod \"dnsmasq-dns-66b577f8c-9dkbr\" (UID: \"b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f\") " pod="openstack/dnsmasq-dns-66b577f8c-9dkbr" Feb 18 09:16:47 crc kubenswrapper[4556]: I0218 09:16:47.784669 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f-config\") pod \"dnsmasq-dns-66b577f8c-9dkbr\" (UID: \"b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f\") " pod="openstack/dnsmasq-dns-66b577f8c-9dkbr" Feb 18 09:16:47 crc kubenswrapper[4556]: I0218 09:16:47.784726 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f-ovsdbserver-sb\") pod \"dnsmasq-dns-66b577f8c-9dkbr\" (UID: \"b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f\") " pod="openstack/dnsmasq-dns-66b577f8c-9dkbr" Feb 18 09:16:47 crc kubenswrapper[4556]: I0218 09:16:47.786095 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f-config\") pod \"dnsmasq-dns-66b577f8c-9dkbr\" (UID: \"b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f\") " pod="openstack/dnsmasq-dns-66b577f8c-9dkbr" Feb 18 09:16:47 crc kubenswrapper[4556]: I0218 09:16:47.786106 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f-dns-svc\") pod \"dnsmasq-dns-66b577f8c-9dkbr\" (UID: \"b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f\") " pod="openstack/dnsmasq-dns-66b577f8c-9dkbr" Feb 18 09:16:47 crc kubenswrapper[4556]: I0218 09:16:47.786563 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f-ovsdbserver-sb\") pod \"dnsmasq-dns-66b577f8c-9dkbr\" (UID: \"b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f\") " pod="openstack/dnsmasq-dns-66b577f8c-9dkbr" Feb 18 09:16:47 crc kubenswrapper[4556]: I0218 09:16:47.786589 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f-ovsdbserver-nb\") pod \"dnsmasq-dns-66b577f8c-9dkbr\" (UID: \"b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f\") " pod="openstack/dnsmasq-dns-66b577f8c-9dkbr" Feb 18 09:16:47 crc kubenswrapper[4556]: I0218 09:16:47.816217 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6kpfc\" (UniqueName: \"kubernetes.io/projected/b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f-kube-api-access-6kpfc\") pod \"dnsmasq-dns-66b577f8c-9dkbr\" (UID: \"b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f\") " pod="openstack/dnsmasq-dns-66b577f8c-9dkbr" Feb 18 09:16:47 crc kubenswrapper[4556]: I0218 09:16:47.936947 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df986d9c-kdzbf" Feb 18 09:16:47 crc kubenswrapper[4556]: I0218 09:16:47.976215 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66b577f8c-9dkbr" Feb 18 09:16:48 crc kubenswrapper[4556]: I0218 09:16:48.089998 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cc7299fa-91af-4d80-957d-7a849a12e76c-ovsdbserver-nb\") pod \"cc7299fa-91af-4d80-957d-7a849a12e76c\" (UID: \"cc7299fa-91af-4d80-957d-7a849a12e76c\") " Feb 18 09:16:48 crc kubenswrapper[4556]: I0218 09:16:48.090175 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cc7299fa-91af-4d80-957d-7a849a12e76c-dns-svc\") pod \"cc7299fa-91af-4d80-957d-7a849a12e76c\" (UID: \"cc7299fa-91af-4d80-957d-7a849a12e76c\") " Feb 18 09:16:48 crc kubenswrapper[4556]: I0218 09:16:48.090283 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jsr9v\" (UniqueName: \"kubernetes.io/projected/cc7299fa-91af-4d80-957d-7a849a12e76c-kube-api-access-jsr9v\") pod \"cc7299fa-91af-4d80-957d-7a849a12e76c\" (UID: \"cc7299fa-91af-4d80-957d-7a849a12e76c\") " Feb 18 09:16:48 crc kubenswrapper[4556]: I0218 09:16:48.090310 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc7299fa-91af-4d80-957d-7a849a12e76c-config\") pod \"cc7299fa-91af-4d80-957d-7a849a12e76c\" (UID: \"cc7299fa-91af-4d80-957d-7a849a12e76c\") " Feb 18 09:16:48 crc kubenswrapper[4556]: I0218 09:16:48.090400 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cc7299fa-91af-4d80-957d-7a849a12e76c-ovsdbserver-sb\") pod \"cc7299fa-91af-4d80-957d-7a849a12e76c\" (UID: \"cc7299fa-91af-4d80-957d-7a849a12e76c\") " Feb 18 09:16:48 crc kubenswrapper[4556]: I0218 09:16:48.100205 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc7299fa-91af-4d80-957d-7a849a12e76c-kube-api-access-jsr9v" (OuterVolumeSpecName: "kube-api-access-jsr9v") pod "cc7299fa-91af-4d80-957d-7a849a12e76c" (UID: "cc7299fa-91af-4d80-957d-7a849a12e76c"). InnerVolumeSpecName "kube-api-access-jsr9v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:16:48 crc kubenswrapper[4556]: I0218 09:16:48.130051 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc7299fa-91af-4d80-957d-7a849a12e76c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "cc7299fa-91af-4d80-957d-7a849a12e76c" (UID: "cc7299fa-91af-4d80-957d-7a849a12e76c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:16:48 crc kubenswrapper[4556]: I0218 09:16:48.132387 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc7299fa-91af-4d80-957d-7a849a12e76c-config" (OuterVolumeSpecName: "config") pod "cc7299fa-91af-4d80-957d-7a849a12e76c" (UID: "cc7299fa-91af-4d80-957d-7a849a12e76c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:16:48 crc kubenswrapper[4556]: I0218 09:16:48.146141 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc7299fa-91af-4d80-957d-7a849a12e76c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "cc7299fa-91af-4d80-957d-7a849a12e76c" (UID: "cc7299fa-91af-4d80-957d-7a849a12e76c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:16:48 crc kubenswrapper[4556]: I0218 09:16:48.150740 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc7299fa-91af-4d80-957d-7a849a12e76c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "cc7299fa-91af-4d80-957d-7a849a12e76c" (UID: "cc7299fa-91af-4d80-957d-7a849a12e76c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:16:48 crc kubenswrapper[4556]: I0218 09:16:48.193668 4556 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cc7299fa-91af-4d80-957d-7a849a12e76c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 18 09:16:48 crc kubenswrapper[4556]: I0218 09:16:48.193704 4556 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cc7299fa-91af-4d80-957d-7a849a12e76c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 18 09:16:48 crc kubenswrapper[4556]: I0218 09:16:48.193714 4556 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cc7299fa-91af-4d80-957d-7a849a12e76c-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 18 09:16:48 crc kubenswrapper[4556]: I0218 09:16:48.193723 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jsr9v\" (UniqueName: \"kubernetes.io/projected/cc7299fa-91af-4d80-957d-7a849a12e76c-kube-api-access-jsr9v\") on node \"crc\" DevicePath \"\"" Feb 18 09:16:48 crc kubenswrapper[4556]: I0218 09:16:48.193738 4556 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc7299fa-91af-4d80-957d-7a849a12e76c-config\") on node \"crc\" DevicePath \"\"" Feb 18 09:16:48 crc kubenswrapper[4556]: I0218 09:16:48.367125 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-66b577f8c-9dkbr"] Feb 18 09:16:48 crc kubenswrapper[4556]: I0218 09:16:48.690771 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66b577f8c-9dkbr" event={"ID":"b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f","Type":"ContainerStarted","Data":"87f39b4d6f60421cb6f3ab8da95f31310358b095971c339e77a5eff518431062"} Feb 18 09:16:48 crc kubenswrapper[4556]: I0218 09:16:48.692187 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df986d9c-kdzbf" event={"ID":"cc7299fa-91af-4d80-957d-7a849a12e76c","Type":"ContainerDied","Data":"76029ec76a7668e60ecfd1c097ec379a82c468a7f72edf7ff95f62468d2600b1"} Feb 18 09:16:48 crc kubenswrapper[4556]: I0218 09:16:48.692245 4556 scope.go:117] "RemoveContainer" containerID="274f996804719440b04e23c51db947fad7337fbbecee12fc1ac0fcb7489aa6af" Feb 18 09:16:48 crc kubenswrapper[4556]: I0218 09:16:48.692389 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df986d9c-kdzbf" Feb 18 09:16:48 crc kubenswrapper[4556]: I0218 09:16:48.712569 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Feb 18 09:16:48 crc kubenswrapper[4556]: E0218 09:16:48.712865 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc7299fa-91af-4d80-957d-7a849a12e76c" containerName="init" Feb 18 09:16:48 crc kubenswrapper[4556]: I0218 09:16:48.712881 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc7299fa-91af-4d80-957d-7a849a12e76c" containerName="init" Feb 18 09:16:48 crc kubenswrapper[4556]: E0218 09:16:48.712890 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc7299fa-91af-4d80-957d-7a849a12e76c" containerName="dnsmasq-dns" Feb 18 09:16:48 crc kubenswrapper[4556]: I0218 09:16:48.712897 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc7299fa-91af-4d80-957d-7a849a12e76c" containerName="dnsmasq-dns" Feb 18 09:16:48 crc kubenswrapper[4556]: I0218 09:16:48.713027 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc7299fa-91af-4d80-957d-7a849a12e76c" containerName="dnsmasq-dns" Feb 18 09:16:48 crc kubenswrapper[4556]: I0218 09:16:48.723375 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Feb 18 09:16:48 crc kubenswrapper[4556]: I0218 09:16:48.732391 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Feb 18 09:16:48 crc kubenswrapper[4556]: I0218 09:16:48.732419 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Feb 18 09:16:48 crc kubenswrapper[4556]: I0218 09:16:48.732952 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-tmptk" Feb 18 09:16:48 crc kubenswrapper[4556]: I0218 09:16:48.733354 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Feb 18 09:16:48 crc kubenswrapper[4556]: I0218 09:16:48.736129 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df986d9c-kdzbf"] Feb 18 09:16:48 crc kubenswrapper[4556]: I0218 09:16:48.753882 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-56df986d9c-kdzbf"] Feb 18 09:16:48 crc kubenswrapper[4556]: I0218 09:16:48.759414 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Feb 18 09:16:48 crc kubenswrapper[4556]: I0218 09:16:48.812694 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/da163339-653b-4057-9c2f-332eb6957f40-lock\") pod \"swift-storage-0\" (UID: \"da163339-653b-4057-9c2f-332eb6957f40\") " pod="openstack/swift-storage-0" Feb 18 09:16:48 crc kubenswrapper[4556]: I0218 09:16:48.812800 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"da163339-653b-4057-9c2f-332eb6957f40\") " pod="openstack/swift-storage-0" Feb 18 09:16:48 crc kubenswrapper[4556]: I0218 09:16:48.813000 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/da163339-653b-4057-9c2f-332eb6957f40-etc-swift\") pod \"swift-storage-0\" (UID: \"da163339-653b-4057-9c2f-332eb6957f40\") " pod="openstack/swift-storage-0" Feb 18 09:16:48 crc kubenswrapper[4556]: I0218 09:16:48.813255 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2ntr\" (UniqueName: \"kubernetes.io/projected/da163339-653b-4057-9c2f-332eb6957f40-kube-api-access-k2ntr\") pod \"swift-storage-0\" (UID: \"da163339-653b-4057-9c2f-332eb6957f40\") " pod="openstack/swift-storage-0" Feb 18 09:16:48 crc kubenswrapper[4556]: I0218 09:16:48.813349 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/da163339-653b-4057-9c2f-332eb6957f40-cache\") pod \"swift-storage-0\" (UID: \"da163339-653b-4057-9c2f-332eb6957f40\") " pod="openstack/swift-storage-0" Feb 18 09:16:48 crc kubenswrapper[4556]: I0218 09:16:48.813419 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da163339-653b-4057-9c2f-332eb6957f40-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"da163339-653b-4057-9c2f-332eb6957f40\") " pod="openstack/swift-storage-0" Feb 18 09:16:48 crc kubenswrapper[4556]: I0218 09:16:48.915115 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"da163339-653b-4057-9c2f-332eb6957f40\") " pod="openstack/swift-storage-0" Feb 18 09:16:48 crc kubenswrapper[4556]: I0218 09:16:48.915179 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/da163339-653b-4057-9c2f-332eb6957f40-etc-swift\") pod \"swift-storage-0\" (UID: \"da163339-653b-4057-9c2f-332eb6957f40\") " pod="openstack/swift-storage-0" Feb 18 09:16:48 crc kubenswrapper[4556]: I0218 09:16:48.915256 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2ntr\" (UniqueName: \"kubernetes.io/projected/da163339-653b-4057-9c2f-332eb6957f40-kube-api-access-k2ntr\") pod \"swift-storage-0\" (UID: \"da163339-653b-4057-9c2f-332eb6957f40\") " pod="openstack/swift-storage-0" Feb 18 09:16:48 crc kubenswrapper[4556]: I0218 09:16:48.915310 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/da163339-653b-4057-9c2f-332eb6957f40-cache\") pod \"swift-storage-0\" (UID: \"da163339-653b-4057-9c2f-332eb6957f40\") " pod="openstack/swift-storage-0" Feb 18 09:16:48 crc kubenswrapper[4556]: I0218 09:16:48.915337 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da163339-653b-4057-9c2f-332eb6957f40-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"da163339-653b-4057-9c2f-332eb6957f40\") " pod="openstack/swift-storage-0" Feb 18 09:16:48 crc kubenswrapper[4556]: E0218 09:16:48.915392 4556 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 18 09:16:48 crc kubenswrapper[4556]: E0218 09:16:48.915415 4556 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 18 09:16:48 crc kubenswrapper[4556]: I0218 09:16:48.915430 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/da163339-653b-4057-9c2f-332eb6957f40-lock\") pod \"swift-storage-0\" (UID: \"da163339-653b-4057-9c2f-332eb6957f40\") " pod="openstack/swift-storage-0" Feb 18 09:16:48 crc kubenswrapper[4556]: E0218 09:16:48.915470 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/da163339-653b-4057-9c2f-332eb6957f40-etc-swift podName:da163339-653b-4057-9c2f-332eb6957f40 nodeName:}" failed. No retries permitted until 2026-02-18 09:16:49.415451486 +0000 UTC m=+766.432412466 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/da163339-653b-4057-9c2f-332eb6957f40-etc-swift") pod "swift-storage-0" (UID: "da163339-653b-4057-9c2f-332eb6957f40") : configmap "swift-ring-files" not found Feb 18 09:16:48 crc kubenswrapper[4556]: I0218 09:16:48.915498 4556 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"da163339-653b-4057-9c2f-332eb6957f40\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/swift-storage-0" Feb 18 09:16:48 crc kubenswrapper[4556]: I0218 09:16:48.915876 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/da163339-653b-4057-9c2f-332eb6957f40-lock\") pod \"swift-storage-0\" (UID: \"da163339-653b-4057-9c2f-332eb6957f40\") " pod="openstack/swift-storage-0" Feb 18 09:16:48 crc kubenswrapper[4556]: I0218 09:16:48.915959 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/da163339-653b-4057-9c2f-332eb6957f40-cache\") pod \"swift-storage-0\" (UID: \"da163339-653b-4057-9c2f-332eb6957f40\") " pod="openstack/swift-storage-0" Feb 18 09:16:48 crc kubenswrapper[4556]: I0218 09:16:48.923463 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da163339-653b-4057-9c2f-332eb6957f40-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"da163339-653b-4057-9c2f-332eb6957f40\") " pod="openstack/swift-storage-0" Feb 18 09:16:48 crc kubenswrapper[4556]: I0218 09:16:48.930129 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2ntr\" (UniqueName: \"kubernetes.io/projected/da163339-653b-4057-9c2f-332eb6957f40-kube-api-access-k2ntr\") pod \"swift-storage-0\" (UID: \"da163339-653b-4057-9c2f-332eb6957f40\") " pod="openstack/swift-storage-0" Feb 18 09:16:48 crc kubenswrapper[4556]: I0218 09:16:48.933368 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"da163339-653b-4057-9c2f-332eb6957f40\") " pod="openstack/swift-storage-0" Feb 18 09:16:49 crc kubenswrapper[4556]: I0218 09:16:49.244442 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-9z5cc"] Feb 18 09:16:49 crc kubenswrapper[4556]: I0218 09:16:49.246282 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-9z5cc" Feb 18 09:16:49 crc kubenswrapper[4556]: I0218 09:16:49.248350 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Feb 18 09:16:49 crc kubenswrapper[4556]: I0218 09:16:49.248562 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Feb 18 09:16:49 crc kubenswrapper[4556]: I0218 09:16:49.248978 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Feb 18 09:16:49 crc kubenswrapper[4556]: I0218 09:16:49.251100 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-9z5cc"] Feb 18 09:16:49 crc kubenswrapper[4556]: I0218 09:16:49.291279 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc7299fa-91af-4d80-957d-7a849a12e76c" path="/var/lib/kubelet/pods/cc7299fa-91af-4d80-957d-7a849a12e76c/volumes" Feb 18 09:16:49 crc kubenswrapper[4556]: I0218 09:16:49.323279 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/03ccffdf-f11c-4788-9e22-0b0661665c5d-etc-swift\") pod \"swift-ring-rebalance-9z5cc\" (UID: \"03ccffdf-f11c-4788-9e22-0b0661665c5d\") " pod="openstack/swift-ring-rebalance-9z5cc" Feb 18 09:16:49 crc kubenswrapper[4556]: I0218 09:16:49.323317 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/03ccffdf-f11c-4788-9e22-0b0661665c5d-ring-data-devices\") pod \"swift-ring-rebalance-9z5cc\" (UID: \"03ccffdf-f11c-4788-9e22-0b0661665c5d\") " pod="openstack/swift-ring-rebalance-9z5cc" Feb 18 09:16:49 crc kubenswrapper[4556]: I0218 09:16:49.323397 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/03ccffdf-f11c-4788-9e22-0b0661665c5d-dispersionconf\") pod \"swift-ring-rebalance-9z5cc\" (UID: \"03ccffdf-f11c-4788-9e22-0b0661665c5d\") " pod="openstack/swift-ring-rebalance-9z5cc" Feb 18 09:16:49 crc kubenswrapper[4556]: I0218 09:16:49.323441 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/03ccffdf-f11c-4788-9e22-0b0661665c5d-swiftconf\") pod \"swift-ring-rebalance-9z5cc\" (UID: \"03ccffdf-f11c-4788-9e22-0b0661665c5d\") " pod="openstack/swift-ring-rebalance-9z5cc" Feb 18 09:16:49 crc kubenswrapper[4556]: I0218 09:16:49.323455 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhwx8\" (UniqueName: \"kubernetes.io/projected/03ccffdf-f11c-4788-9e22-0b0661665c5d-kube-api-access-zhwx8\") pod \"swift-ring-rebalance-9z5cc\" (UID: \"03ccffdf-f11c-4788-9e22-0b0661665c5d\") " pod="openstack/swift-ring-rebalance-9z5cc" Feb 18 09:16:49 crc kubenswrapper[4556]: I0218 09:16:49.323492 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03ccffdf-f11c-4788-9e22-0b0661665c5d-combined-ca-bundle\") pod \"swift-ring-rebalance-9z5cc\" (UID: \"03ccffdf-f11c-4788-9e22-0b0661665c5d\") " pod="openstack/swift-ring-rebalance-9z5cc" Feb 18 09:16:49 crc kubenswrapper[4556]: I0218 09:16:49.323517 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/03ccffdf-f11c-4788-9e22-0b0661665c5d-scripts\") pod \"swift-ring-rebalance-9z5cc\" (UID: \"03ccffdf-f11c-4788-9e22-0b0661665c5d\") " pod="openstack/swift-ring-rebalance-9z5cc" Feb 18 09:16:49 crc kubenswrapper[4556]: I0218 09:16:49.424867 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/03ccffdf-f11c-4788-9e22-0b0661665c5d-swiftconf\") pod \"swift-ring-rebalance-9z5cc\" (UID: \"03ccffdf-f11c-4788-9e22-0b0661665c5d\") " pod="openstack/swift-ring-rebalance-9z5cc" Feb 18 09:16:49 crc kubenswrapper[4556]: I0218 09:16:49.424904 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhwx8\" (UniqueName: \"kubernetes.io/projected/03ccffdf-f11c-4788-9e22-0b0661665c5d-kube-api-access-zhwx8\") pod \"swift-ring-rebalance-9z5cc\" (UID: \"03ccffdf-f11c-4788-9e22-0b0661665c5d\") " pod="openstack/swift-ring-rebalance-9z5cc" Feb 18 09:16:49 crc kubenswrapper[4556]: I0218 09:16:49.424946 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03ccffdf-f11c-4788-9e22-0b0661665c5d-combined-ca-bundle\") pod \"swift-ring-rebalance-9z5cc\" (UID: \"03ccffdf-f11c-4788-9e22-0b0661665c5d\") " pod="openstack/swift-ring-rebalance-9z5cc" Feb 18 09:16:49 crc kubenswrapper[4556]: I0218 09:16:49.424967 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/03ccffdf-f11c-4788-9e22-0b0661665c5d-scripts\") pod \"swift-ring-rebalance-9z5cc\" (UID: \"03ccffdf-f11c-4788-9e22-0b0661665c5d\") " pod="openstack/swift-ring-rebalance-9z5cc" Feb 18 09:16:49 crc kubenswrapper[4556]: I0218 09:16:49.425057 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/03ccffdf-f11c-4788-9e22-0b0661665c5d-etc-swift\") pod \"swift-ring-rebalance-9z5cc\" (UID: \"03ccffdf-f11c-4788-9e22-0b0661665c5d\") " pod="openstack/swift-ring-rebalance-9z5cc" Feb 18 09:16:49 crc kubenswrapper[4556]: I0218 09:16:49.425077 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/03ccffdf-f11c-4788-9e22-0b0661665c5d-ring-data-devices\") pod \"swift-ring-rebalance-9z5cc\" (UID: \"03ccffdf-f11c-4788-9e22-0b0661665c5d\") " pod="openstack/swift-ring-rebalance-9z5cc" Feb 18 09:16:49 crc kubenswrapper[4556]: I0218 09:16:49.425106 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/da163339-653b-4057-9c2f-332eb6957f40-etc-swift\") pod \"swift-storage-0\" (UID: \"da163339-653b-4057-9c2f-332eb6957f40\") " pod="openstack/swift-storage-0" Feb 18 09:16:49 crc kubenswrapper[4556]: I0218 09:16:49.425138 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/03ccffdf-f11c-4788-9e22-0b0661665c5d-dispersionconf\") pod \"swift-ring-rebalance-9z5cc\" (UID: \"03ccffdf-f11c-4788-9e22-0b0661665c5d\") " pod="openstack/swift-ring-rebalance-9z5cc" Feb 18 09:16:49 crc kubenswrapper[4556]: I0218 09:16:49.426605 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/03ccffdf-f11c-4788-9e22-0b0661665c5d-scripts\") pod \"swift-ring-rebalance-9z5cc\" (UID: \"03ccffdf-f11c-4788-9e22-0b0661665c5d\") " pod="openstack/swift-ring-rebalance-9z5cc" Feb 18 09:16:49 crc kubenswrapper[4556]: I0218 09:16:49.427109 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/03ccffdf-f11c-4788-9e22-0b0661665c5d-ring-data-devices\") pod \"swift-ring-rebalance-9z5cc\" (UID: \"03ccffdf-f11c-4788-9e22-0b0661665c5d\") " pod="openstack/swift-ring-rebalance-9z5cc" Feb 18 09:16:49 crc kubenswrapper[4556]: E0218 09:16:49.427231 4556 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 18 09:16:49 crc kubenswrapper[4556]: E0218 09:16:49.427286 4556 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 18 09:16:49 crc kubenswrapper[4556]: E0218 09:16:49.427388 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/da163339-653b-4057-9c2f-332eb6957f40-etc-swift podName:da163339-653b-4057-9c2f-332eb6957f40 nodeName:}" failed. No retries permitted until 2026-02-18 09:16:50.427354587 +0000 UTC m=+767.444315566 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/da163339-653b-4057-9c2f-332eb6957f40-etc-swift") pod "swift-storage-0" (UID: "da163339-653b-4057-9c2f-332eb6957f40") : configmap "swift-ring-files" not found Feb 18 09:16:49 crc kubenswrapper[4556]: I0218 09:16:49.428358 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/03ccffdf-f11c-4788-9e22-0b0661665c5d-etc-swift\") pod \"swift-ring-rebalance-9z5cc\" (UID: \"03ccffdf-f11c-4788-9e22-0b0661665c5d\") " pod="openstack/swift-ring-rebalance-9z5cc" Feb 18 09:16:49 crc kubenswrapper[4556]: I0218 09:16:49.430526 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/03ccffdf-f11c-4788-9e22-0b0661665c5d-dispersionconf\") pod \"swift-ring-rebalance-9z5cc\" (UID: \"03ccffdf-f11c-4788-9e22-0b0661665c5d\") " pod="openstack/swift-ring-rebalance-9z5cc" Feb 18 09:16:49 crc kubenswrapper[4556]: I0218 09:16:49.432506 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/03ccffdf-f11c-4788-9e22-0b0661665c5d-swiftconf\") pod \"swift-ring-rebalance-9z5cc\" (UID: \"03ccffdf-f11c-4788-9e22-0b0661665c5d\") " pod="openstack/swift-ring-rebalance-9z5cc" Feb 18 09:16:49 crc kubenswrapper[4556]: I0218 09:16:49.432564 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03ccffdf-f11c-4788-9e22-0b0661665c5d-combined-ca-bundle\") pod \"swift-ring-rebalance-9z5cc\" (UID: \"03ccffdf-f11c-4788-9e22-0b0661665c5d\") " pod="openstack/swift-ring-rebalance-9z5cc" Feb 18 09:16:49 crc kubenswrapper[4556]: I0218 09:16:49.439962 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhwx8\" (UniqueName: \"kubernetes.io/projected/03ccffdf-f11c-4788-9e22-0b0661665c5d-kube-api-access-zhwx8\") pod \"swift-ring-rebalance-9z5cc\" (UID: \"03ccffdf-f11c-4788-9e22-0b0661665c5d\") " pod="openstack/swift-ring-rebalance-9z5cc" Feb 18 09:16:49 crc kubenswrapper[4556]: I0218 09:16:49.562123 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-9z5cc" Feb 18 09:16:50 crc kubenswrapper[4556]: I0218 09:16:50.218699 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Feb 18 09:16:50 crc kubenswrapper[4556]: I0218 09:16:50.332573 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Feb 18 09:16:50 crc kubenswrapper[4556]: I0218 09:16:50.333818 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Feb 18 09:16:50 crc kubenswrapper[4556]: I0218 09:16:50.338747 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Feb 18 09:16:50 crc kubenswrapper[4556]: I0218 09:16:50.338747 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Feb 18 09:16:50 crc kubenswrapper[4556]: I0218 09:16:50.339015 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-2cvxd" Feb 18 09:16:50 crc kubenswrapper[4556]: I0218 09:16:50.339428 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Feb 18 09:16:50 crc kubenswrapper[4556]: I0218 09:16:50.344243 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Feb 18 09:16:50 crc kubenswrapper[4556]: I0218 09:16:50.444485 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ef64adb0-c929-4a3f-8aa6-d490b45ba5e1-scripts\") pod \"ovn-northd-0\" (UID: \"ef64adb0-c929-4a3f-8aa6-d490b45ba5e1\") " pod="openstack/ovn-northd-0" Feb 18 09:16:50 crc kubenswrapper[4556]: I0218 09:16:50.444538 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f82m9\" (UniqueName: \"kubernetes.io/projected/ef64adb0-c929-4a3f-8aa6-d490b45ba5e1-kube-api-access-f82m9\") pod \"ovn-northd-0\" (UID: \"ef64adb0-c929-4a3f-8aa6-d490b45ba5e1\") " pod="openstack/ovn-northd-0" Feb 18 09:16:50 crc kubenswrapper[4556]: I0218 09:16:50.444640 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef64adb0-c929-4a3f-8aa6-d490b45ba5e1-config\") pod \"ovn-northd-0\" (UID: \"ef64adb0-c929-4a3f-8aa6-d490b45ba5e1\") " pod="openstack/ovn-northd-0" Feb 18 09:16:50 crc kubenswrapper[4556]: I0218 09:16:50.444698 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef64adb0-c929-4a3f-8aa6-d490b45ba5e1-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"ef64adb0-c929-4a3f-8aa6-d490b45ba5e1\") " pod="openstack/ovn-northd-0" Feb 18 09:16:50 crc kubenswrapper[4556]: I0218 09:16:50.444726 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef64adb0-c929-4a3f-8aa6-d490b45ba5e1-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"ef64adb0-c929-4a3f-8aa6-d490b45ba5e1\") " pod="openstack/ovn-northd-0" Feb 18 09:16:50 crc kubenswrapper[4556]: I0218 09:16:50.444749 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef64adb0-c929-4a3f-8aa6-d490b45ba5e1-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"ef64adb0-c929-4a3f-8aa6-d490b45ba5e1\") " pod="openstack/ovn-northd-0" Feb 18 09:16:50 crc kubenswrapper[4556]: I0218 09:16:50.444791 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/da163339-653b-4057-9c2f-332eb6957f40-etc-swift\") pod \"swift-storage-0\" (UID: \"da163339-653b-4057-9c2f-332eb6957f40\") " pod="openstack/swift-storage-0" Feb 18 09:16:50 crc kubenswrapper[4556]: I0218 09:16:50.444926 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ef64adb0-c929-4a3f-8aa6-d490b45ba5e1-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"ef64adb0-c929-4a3f-8aa6-d490b45ba5e1\") " pod="openstack/ovn-northd-0" Feb 18 09:16:50 crc kubenswrapper[4556]: E0218 09:16:50.445144 4556 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 18 09:16:50 crc kubenswrapper[4556]: E0218 09:16:50.445219 4556 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 18 09:16:50 crc kubenswrapper[4556]: E0218 09:16:50.445342 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/da163339-653b-4057-9c2f-332eb6957f40-etc-swift podName:da163339-653b-4057-9c2f-332eb6957f40 nodeName:}" failed. No retries permitted until 2026-02-18 09:16:52.445308189 +0000 UTC m=+769.462269169 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/da163339-653b-4057-9c2f-332eb6957f40-etc-swift") pod "swift-storage-0" (UID: "da163339-653b-4057-9c2f-332eb6957f40") : configmap "swift-ring-files" not found Feb 18 09:16:50 crc kubenswrapper[4556]: I0218 09:16:50.547020 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ef64adb0-c929-4a3f-8aa6-d490b45ba5e1-scripts\") pod \"ovn-northd-0\" (UID: \"ef64adb0-c929-4a3f-8aa6-d490b45ba5e1\") " pod="openstack/ovn-northd-0" Feb 18 09:16:50 crc kubenswrapper[4556]: I0218 09:16:50.547376 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f82m9\" (UniqueName: \"kubernetes.io/projected/ef64adb0-c929-4a3f-8aa6-d490b45ba5e1-kube-api-access-f82m9\") pod \"ovn-northd-0\" (UID: \"ef64adb0-c929-4a3f-8aa6-d490b45ba5e1\") " pod="openstack/ovn-northd-0" Feb 18 09:16:50 crc kubenswrapper[4556]: I0218 09:16:50.547463 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef64adb0-c929-4a3f-8aa6-d490b45ba5e1-config\") pod \"ovn-northd-0\" (UID: \"ef64adb0-c929-4a3f-8aa6-d490b45ba5e1\") " pod="openstack/ovn-northd-0" Feb 18 09:16:50 crc kubenswrapper[4556]: I0218 09:16:50.547516 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef64adb0-c929-4a3f-8aa6-d490b45ba5e1-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"ef64adb0-c929-4a3f-8aa6-d490b45ba5e1\") " pod="openstack/ovn-northd-0" Feb 18 09:16:50 crc kubenswrapper[4556]: I0218 09:16:50.547540 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef64adb0-c929-4a3f-8aa6-d490b45ba5e1-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"ef64adb0-c929-4a3f-8aa6-d490b45ba5e1\") " pod="openstack/ovn-northd-0" Feb 18 09:16:50 crc kubenswrapper[4556]: I0218 09:16:50.547564 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef64adb0-c929-4a3f-8aa6-d490b45ba5e1-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"ef64adb0-c929-4a3f-8aa6-d490b45ba5e1\") " pod="openstack/ovn-northd-0" Feb 18 09:16:50 crc kubenswrapper[4556]: I0218 09:16:50.547624 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ef64adb0-c929-4a3f-8aa6-d490b45ba5e1-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"ef64adb0-c929-4a3f-8aa6-d490b45ba5e1\") " pod="openstack/ovn-northd-0" Feb 18 09:16:50 crc kubenswrapper[4556]: I0218 09:16:50.547943 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ef64adb0-c929-4a3f-8aa6-d490b45ba5e1-scripts\") pod \"ovn-northd-0\" (UID: \"ef64adb0-c929-4a3f-8aa6-d490b45ba5e1\") " pod="openstack/ovn-northd-0" Feb 18 09:16:50 crc kubenswrapper[4556]: I0218 09:16:50.548109 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ef64adb0-c929-4a3f-8aa6-d490b45ba5e1-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"ef64adb0-c929-4a3f-8aa6-d490b45ba5e1\") " pod="openstack/ovn-northd-0" Feb 18 09:16:50 crc kubenswrapper[4556]: I0218 09:16:50.548641 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef64adb0-c929-4a3f-8aa6-d490b45ba5e1-config\") pod \"ovn-northd-0\" (UID: \"ef64adb0-c929-4a3f-8aa6-d490b45ba5e1\") " pod="openstack/ovn-northd-0" Feb 18 09:16:50 crc kubenswrapper[4556]: I0218 09:16:50.553122 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef64adb0-c929-4a3f-8aa6-d490b45ba5e1-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"ef64adb0-c929-4a3f-8aa6-d490b45ba5e1\") " pod="openstack/ovn-northd-0" Feb 18 09:16:50 crc kubenswrapper[4556]: I0218 09:16:50.554802 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef64adb0-c929-4a3f-8aa6-d490b45ba5e1-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"ef64adb0-c929-4a3f-8aa6-d490b45ba5e1\") " pod="openstack/ovn-northd-0" Feb 18 09:16:50 crc kubenswrapper[4556]: I0218 09:16:50.554812 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef64adb0-c929-4a3f-8aa6-d490b45ba5e1-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"ef64adb0-c929-4a3f-8aa6-d490b45ba5e1\") " pod="openstack/ovn-northd-0" Feb 18 09:16:50 crc kubenswrapper[4556]: I0218 09:16:50.561666 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f82m9\" (UniqueName: \"kubernetes.io/projected/ef64adb0-c929-4a3f-8aa6-d490b45ba5e1-kube-api-access-f82m9\") pod \"ovn-northd-0\" (UID: \"ef64adb0-c929-4a3f-8aa6-d490b45ba5e1\") " pod="openstack/ovn-northd-0" Feb 18 09:16:50 crc kubenswrapper[4556]: I0218 09:16:50.654068 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Feb 18 09:16:52 crc kubenswrapper[4556]: I0218 09:16:52.123417 4556 scope.go:117] "RemoveContainer" containerID="0e00134d256432d75bb3b13230f55719bdd3db03254837a9dfd979e4e3b6947a" Feb 18 09:16:52 crc kubenswrapper[4556]: I0218 09:16:52.490622 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/da163339-653b-4057-9c2f-332eb6957f40-etc-swift\") pod \"swift-storage-0\" (UID: \"da163339-653b-4057-9c2f-332eb6957f40\") " pod="openstack/swift-storage-0" Feb 18 09:16:52 crc kubenswrapper[4556]: E0218 09:16:52.490828 4556 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 18 09:16:52 crc kubenswrapper[4556]: E0218 09:16:52.490855 4556 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 18 09:16:52 crc kubenswrapper[4556]: E0218 09:16:52.490932 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/da163339-653b-4057-9c2f-332eb6957f40-etc-swift podName:da163339-653b-4057-9c2f-332eb6957f40 nodeName:}" failed. No retries permitted until 2026-02-18 09:16:56.490915115 +0000 UTC m=+773.507876095 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/da163339-653b-4057-9c2f-332eb6957f40-etc-swift") pod "swift-storage-0" (UID: "da163339-653b-4057-9c2f-332eb6957f40") : configmap "swift-ring-files" not found Feb 18 09:16:52 crc kubenswrapper[4556]: I0218 09:16:52.522459 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Feb 18 09:16:52 crc kubenswrapper[4556]: W0218 09:16:52.523721 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podef64adb0_c929_4a3f_8aa6_d490b45ba5e1.slice/crio-e6c7e67153c85c78febee377630a913725f32577b7edd15b7cdcc8e1a3c3b015 WatchSource:0}: Error finding container e6c7e67153c85c78febee377630a913725f32577b7edd15b7cdcc8e1a3c3b015: Status 404 returned error can't find the container with id e6c7e67153c85c78febee377630a913725f32577b7edd15b7cdcc8e1a3c3b015 Feb 18 09:16:52 crc kubenswrapper[4556]: I0218 09:16:52.526915 4556 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 18 09:16:52 crc kubenswrapper[4556]: I0218 09:16:52.592363 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-9z5cc"] Feb 18 09:16:52 crc kubenswrapper[4556]: W0218 09:16:52.594726 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod03ccffdf_f11c_4788_9e22_0b0661665c5d.slice/crio-0233d450ed2b7409752b7b01e9727cb23c5995dfa206a754bf8045f48287dabd WatchSource:0}: Error finding container 0233d450ed2b7409752b7b01e9727cb23c5995dfa206a754bf8045f48287dabd: Status 404 returned error can't find the container with id 0233d450ed2b7409752b7b01e9727cb23c5995dfa206a754bf8045f48287dabd Feb 18 09:16:52 crc kubenswrapper[4556]: I0218 09:16:52.726452 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"ef64adb0-c929-4a3f-8aa6-d490b45ba5e1","Type":"ContainerStarted","Data":"e6c7e67153c85c78febee377630a913725f32577b7edd15b7cdcc8e1a3c3b015"} Feb 18 09:16:52 crc kubenswrapper[4556]: I0218 09:16:52.727518 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-9z5cc" event={"ID":"03ccffdf-f11c-4788-9e22-0b0661665c5d","Type":"ContainerStarted","Data":"0233d450ed2b7409752b7b01e9727cb23c5995dfa206a754bf8045f48287dabd"} Feb 18 09:16:53 crc kubenswrapper[4556]: I0218 09:16:53.627804 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Feb 18 09:16:53 crc kubenswrapper[4556]: I0218 09:16:53.628176 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Feb 18 09:16:53 crc kubenswrapper[4556]: I0218 09:16:53.703098 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-lbxfb"] Feb 18 09:16:53 crc kubenswrapper[4556]: I0218 09:16:53.704265 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-lbxfb" Feb 18 09:16:53 crc kubenswrapper[4556]: I0218 09:16:53.706176 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-mariadb-root-db-secret" Feb 18 09:16:53 crc kubenswrapper[4556]: I0218 09:16:53.709561 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-lbxfb"] Feb 18 09:16:53 crc kubenswrapper[4556]: I0218 09:16:53.724832 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Feb 18 09:16:53 crc kubenswrapper[4556]: I0218 09:16:53.745518 4556 generic.go:334] "Generic (PLEG): container finished" podID="b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f" containerID="8f567d29c8442ee45a00a6b7f4c2a7e49859c1fe116422ede582ac534a996100" exitCode=0 Feb 18 09:16:53 crc kubenswrapper[4556]: I0218 09:16:53.745604 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66b577f8c-9dkbr" event={"ID":"b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f","Type":"ContainerDied","Data":"8f567d29c8442ee45a00a6b7f4c2a7e49859c1fe116422ede582ac534a996100"} Feb 18 09:16:53 crc kubenswrapper[4556]: I0218 09:16:53.821004 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Feb 18 09:16:53 crc kubenswrapper[4556]: I0218 09:16:53.822108 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7pkxz\" (UniqueName: \"kubernetes.io/projected/0537334d-14e5-4ff1-ab08-4c4e7a882c9a-kube-api-access-7pkxz\") pod \"root-account-create-update-lbxfb\" (UID: \"0537334d-14e5-4ff1-ab08-4c4e7a882c9a\") " pod="openstack/root-account-create-update-lbxfb" Feb 18 09:16:53 crc kubenswrapper[4556]: I0218 09:16:53.822214 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0537334d-14e5-4ff1-ab08-4c4e7a882c9a-operator-scripts\") pod \"root-account-create-update-lbxfb\" (UID: \"0537334d-14e5-4ff1-ab08-4c4e7a882c9a\") " pod="openstack/root-account-create-update-lbxfb" Feb 18 09:16:53 crc kubenswrapper[4556]: I0218 09:16:53.929290 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7pkxz\" (UniqueName: \"kubernetes.io/projected/0537334d-14e5-4ff1-ab08-4c4e7a882c9a-kube-api-access-7pkxz\") pod \"root-account-create-update-lbxfb\" (UID: \"0537334d-14e5-4ff1-ab08-4c4e7a882c9a\") " pod="openstack/root-account-create-update-lbxfb" Feb 18 09:16:53 crc kubenswrapper[4556]: I0218 09:16:53.929450 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0537334d-14e5-4ff1-ab08-4c4e7a882c9a-operator-scripts\") pod \"root-account-create-update-lbxfb\" (UID: \"0537334d-14e5-4ff1-ab08-4c4e7a882c9a\") " pod="openstack/root-account-create-update-lbxfb" Feb 18 09:16:53 crc kubenswrapper[4556]: I0218 09:16:53.931909 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0537334d-14e5-4ff1-ab08-4c4e7a882c9a-operator-scripts\") pod \"root-account-create-update-lbxfb\" (UID: \"0537334d-14e5-4ff1-ab08-4c4e7a882c9a\") " pod="openstack/root-account-create-update-lbxfb" Feb 18 09:16:53 crc kubenswrapper[4556]: I0218 09:16:53.953749 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7pkxz\" (UniqueName: \"kubernetes.io/projected/0537334d-14e5-4ff1-ab08-4c4e7a882c9a-kube-api-access-7pkxz\") pod \"root-account-create-update-lbxfb\" (UID: \"0537334d-14e5-4ff1-ab08-4c4e7a882c9a\") " pod="openstack/root-account-create-update-lbxfb" Feb 18 09:16:54 crc kubenswrapper[4556]: I0218 09:16:54.030527 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-lbxfb" Feb 18 09:16:54 crc kubenswrapper[4556]: I0218 09:16:54.590018 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-lbxfb"] Feb 18 09:16:54 crc kubenswrapper[4556]: W0218 09:16:54.599745 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0537334d_14e5_4ff1_ab08_4c4e7a882c9a.slice/crio-b6fc0becf112d0cf2a50d883daae1e0cd5ea68f49d82e8ed944126faa9bcf621 WatchSource:0}: Error finding container b6fc0becf112d0cf2a50d883daae1e0cd5ea68f49d82e8ed944126faa9bcf621: Status 404 returned error can't find the container with id b6fc0becf112d0cf2a50d883daae1e0cd5ea68f49d82e8ed944126faa9bcf621 Feb 18 09:16:54 crc kubenswrapper[4556]: I0218 09:16:54.764313 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-lbxfb" event={"ID":"0537334d-14e5-4ff1-ab08-4c4e7a882c9a","Type":"ContainerStarted","Data":"d1da7830e114e7c9d08657a35bdcbad92e1e4bf6c2ebaeedb0d76f08ad71fa0d"} Feb 18 09:16:54 crc kubenswrapper[4556]: I0218 09:16:54.765328 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-lbxfb" event={"ID":"0537334d-14e5-4ff1-ab08-4c4e7a882c9a","Type":"ContainerStarted","Data":"b6fc0becf112d0cf2a50d883daae1e0cd5ea68f49d82e8ed944126faa9bcf621"} Feb 18 09:16:54 crc kubenswrapper[4556]: I0218 09:16:54.769616 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66b577f8c-9dkbr" event={"ID":"b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f","Type":"ContainerStarted","Data":"a8dadc6867286da49a84752ef6856ada290207ac296cb9cfefb9d2498343692e"} Feb 18 09:16:54 crc kubenswrapper[4556]: I0218 09:16:54.769701 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-66b577f8c-9dkbr" Feb 18 09:16:54 crc kubenswrapper[4556]: I0218 09:16:54.772653 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"ef64adb0-c929-4a3f-8aa6-d490b45ba5e1","Type":"ContainerStarted","Data":"94f93077301a023c923b54427c23fdff858bb3755ea6e68c24ebf9fc75a255fb"} Feb 18 09:16:54 crc kubenswrapper[4556]: I0218 09:16:54.772692 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"ef64adb0-c929-4a3f-8aa6-d490b45ba5e1","Type":"ContainerStarted","Data":"a1d1836bbdc12e4bb0c9dca2a62dd7ba955b826e8d64b90f0884b96303f33168"} Feb 18 09:16:54 crc kubenswrapper[4556]: I0218 09:16:54.772705 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Feb 18 09:16:54 crc kubenswrapper[4556]: I0218 09:16:54.782207 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/root-account-create-update-lbxfb" podStartSLOduration=1.782188701 podStartE2EDuration="1.782188701s" podCreationTimestamp="2026-02-18 09:16:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:16:54.7774105 +0000 UTC m=+771.794371481" watchObservedRunningTime="2026-02-18 09:16:54.782188701 +0000 UTC m=+771.799149701" Feb 18 09:16:54 crc kubenswrapper[4556]: I0218 09:16:54.797934 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=3.117198738 podStartE2EDuration="4.797922944s" podCreationTimestamp="2026-02-18 09:16:50 +0000 UTC" firstStartedPulling="2026-02-18 09:16:52.526686726 +0000 UTC m=+769.543647707" lastFinishedPulling="2026-02-18 09:16:54.207410933 +0000 UTC m=+771.224371913" observedRunningTime="2026-02-18 09:16:54.793186183 +0000 UTC m=+771.810147163" watchObservedRunningTime="2026-02-18 09:16:54.797922944 +0000 UTC m=+771.814883924" Feb 18 09:16:54 crc kubenswrapper[4556]: I0218 09:16:54.810223 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-66b577f8c-9dkbr" podStartSLOduration=7.810206534 podStartE2EDuration="7.810206534s" podCreationTimestamp="2026-02-18 09:16:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:16:54.807135015 +0000 UTC m=+771.824095996" watchObservedRunningTime="2026-02-18 09:16:54.810206534 +0000 UTC m=+771.827167514" Feb 18 09:16:55 crc kubenswrapper[4556]: I0218 09:16:55.292177 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-8n5wp"] Feb 18 09:16:55 crc kubenswrapper[4556]: I0218 09:16:55.293351 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-8n5wp"] Feb 18 09:16:55 crc kubenswrapper[4556]: I0218 09:16:55.293420 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-8n5wp" Feb 18 09:16:55 crc kubenswrapper[4556]: I0218 09:16:55.378554 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-de68-account-create-update-f26q6"] Feb 18 09:16:55 crc kubenswrapper[4556]: I0218 09:16:55.380187 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-de68-account-create-update-f26q6" Feb 18 09:16:55 crc kubenswrapper[4556]: I0218 09:16:55.385592 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Feb 18 09:16:55 crc kubenswrapper[4556]: I0218 09:16:55.403261 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-de68-account-create-update-f26q6"] Feb 18 09:16:55 crc kubenswrapper[4556]: I0218 09:16:55.459730 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxvxx\" (UniqueName: \"kubernetes.io/projected/b43cd61b-beae-4ef8-8648-d014d045e56c-kube-api-access-fxvxx\") pod \"glance-db-create-8n5wp\" (UID: \"b43cd61b-beae-4ef8-8648-d014d045e56c\") " pod="openstack/glance-db-create-8n5wp" Feb 18 09:16:55 crc kubenswrapper[4556]: I0218 09:16:55.459774 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b43cd61b-beae-4ef8-8648-d014d045e56c-operator-scripts\") pod \"glance-db-create-8n5wp\" (UID: \"b43cd61b-beae-4ef8-8648-d014d045e56c\") " pod="openstack/glance-db-create-8n5wp" Feb 18 09:16:55 crc kubenswrapper[4556]: I0218 09:16:55.459819 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8f7157d7-e589-4d41-955e-73ad86618b83-operator-scripts\") pod \"glance-de68-account-create-update-f26q6\" (UID: \"8f7157d7-e589-4d41-955e-73ad86618b83\") " pod="openstack/glance-de68-account-create-update-f26q6" Feb 18 09:16:55 crc kubenswrapper[4556]: I0218 09:16:55.459856 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rsrc5\" (UniqueName: \"kubernetes.io/projected/8f7157d7-e589-4d41-955e-73ad86618b83-kube-api-access-rsrc5\") pod \"glance-de68-account-create-update-f26q6\" (UID: \"8f7157d7-e589-4d41-955e-73ad86618b83\") " pod="openstack/glance-de68-account-create-update-f26q6" Feb 18 09:16:55 crc kubenswrapper[4556]: I0218 09:16:55.561895 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxvxx\" (UniqueName: \"kubernetes.io/projected/b43cd61b-beae-4ef8-8648-d014d045e56c-kube-api-access-fxvxx\") pod \"glance-db-create-8n5wp\" (UID: \"b43cd61b-beae-4ef8-8648-d014d045e56c\") " pod="openstack/glance-db-create-8n5wp" Feb 18 09:16:55 crc kubenswrapper[4556]: I0218 09:16:55.561942 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b43cd61b-beae-4ef8-8648-d014d045e56c-operator-scripts\") pod \"glance-db-create-8n5wp\" (UID: \"b43cd61b-beae-4ef8-8648-d014d045e56c\") " pod="openstack/glance-db-create-8n5wp" Feb 18 09:16:55 crc kubenswrapper[4556]: I0218 09:16:55.561973 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8f7157d7-e589-4d41-955e-73ad86618b83-operator-scripts\") pod \"glance-de68-account-create-update-f26q6\" (UID: \"8f7157d7-e589-4d41-955e-73ad86618b83\") " pod="openstack/glance-de68-account-create-update-f26q6" Feb 18 09:16:55 crc kubenswrapper[4556]: I0218 09:16:55.562000 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rsrc5\" (UniqueName: \"kubernetes.io/projected/8f7157d7-e589-4d41-955e-73ad86618b83-kube-api-access-rsrc5\") pod \"glance-de68-account-create-update-f26q6\" (UID: \"8f7157d7-e589-4d41-955e-73ad86618b83\") " pod="openstack/glance-de68-account-create-update-f26q6" Feb 18 09:16:55 crc kubenswrapper[4556]: I0218 09:16:55.563072 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b43cd61b-beae-4ef8-8648-d014d045e56c-operator-scripts\") pod \"glance-db-create-8n5wp\" (UID: \"b43cd61b-beae-4ef8-8648-d014d045e56c\") " pod="openstack/glance-db-create-8n5wp" Feb 18 09:16:55 crc kubenswrapper[4556]: I0218 09:16:55.563559 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8f7157d7-e589-4d41-955e-73ad86618b83-operator-scripts\") pod \"glance-de68-account-create-update-f26q6\" (UID: \"8f7157d7-e589-4d41-955e-73ad86618b83\") " pod="openstack/glance-de68-account-create-update-f26q6" Feb 18 09:16:55 crc kubenswrapper[4556]: I0218 09:16:55.580620 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxvxx\" (UniqueName: \"kubernetes.io/projected/b43cd61b-beae-4ef8-8648-d014d045e56c-kube-api-access-fxvxx\") pod \"glance-db-create-8n5wp\" (UID: \"b43cd61b-beae-4ef8-8648-d014d045e56c\") " pod="openstack/glance-db-create-8n5wp" Feb 18 09:16:55 crc kubenswrapper[4556]: I0218 09:16:55.581983 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rsrc5\" (UniqueName: \"kubernetes.io/projected/8f7157d7-e589-4d41-955e-73ad86618b83-kube-api-access-rsrc5\") pod \"glance-de68-account-create-update-f26q6\" (UID: \"8f7157d7-e589-4d41-955e-73ad86618b83\") " pod="openstack/glance-de68-account-create-update-f26q6" Feb 18 09:16:55 crc kubenswrapper[4556]: I0218 09:16:55.612990 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-8n5wp" Feb 18 09:16:55 crc kubenswrapper[4556]: I0218 09:16:55.714263 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-de68-account-create-update-f26q6" Feb 18 09:16:55 crc kubenswrapper[4556]: I0218 09:16:55.786003 4556 generic.go:334] "Generic (PLEG): container finished" podID="0537334d-14e5-4ff1-ab08-4c4e7a882c9a" containerID="d1da7830e114e7c9d08657a35bdcbad92e1e4bf6c2ebaeedb0d76f08ad71fa0d" exitCode=0 Feb 18 09:16:55 crc kubenswrapper[4556]: I0218 09:16:55.786089 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-lbxfb" event={"ID":"0537334d-14e5-4ff1-ab08-4c4e7a882c9a","Type":"ContainerDied","Data":"d1da7830e114e7c9d08657a35bdcbad92e1e4bf6c2ebaeedb0d76f08ad71fa0d"} Feb 18 09:16:56 crc kubenswrapper[4556]: I0218 09:16:56.189036 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-94t8h"] Feb 18 09:16:56 crc kubenswrapper[4556]: I0218 09:16:56.190206 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-94t8h" Feb 18 09:16:56 crc kubenswrapper[4556]: I0218 09:16:56.194881 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-94t8h"] Feb 18 09:16:56 crc kubenswrapper[4556]: I0218 09:16:56.286303 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ce838404-6d81-41ca-bdbf-e9af260a0d55-operator-scripts\") pod \"keystone-db-create-94t8h\" (UID: \"ce838404-6d81-41ca-bdbf-e9af260a0d55\") " pod="openstack/keystone-db-create-94t8h" Feb 18 09:16:56 crc kubenswrapper[4556]: I0218 09:16:56.286608 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldfwb\" (UniqueName: \"kubernetes.io/projected/ce838404-6d81-41ca-bdbf-e9af260a0d55-kube-api-access-ldfwb\") pod \"keystone-db-create-94t8h\" (UID: \"ce838404-6d81-41ca-bdbf-e9af260a0d55\") " pod="openstack/keystone-db-create-94t8h" Feb 18 09:16:56 crc kubenswrapper[4556]: I0218 09:16:56.308295 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-101a-account-create-update-t5fnf"] Feb 18 09:16:56 crc kubenswrapper[4556]: I0218 09:16:56.309924 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-101a-account-create-update-t5fnf" Feb 18 09:16:56 crc kubenswrapper[4556]: I0218 09:16:56.312390 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Feb 18 09:16:56 crc kubenswrapper[4556]: I0218 09:16:56.321426 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-101a-account-create-update-t5fnf"] Feb 18 09:16:56 crc kubenswrapper[4556]: I0218 09:16:56.388139 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/45a79e16-1b19-4d21-9edb-ff35eb78d357-operator-scripts\") pod \"keystone-101a-account-create-update-t5fnf\" (UID: \"45a79e16-1b19-4d21-9edb-ff35eb78d357\") " pod="openstack/keystone-101a-account-create-update-t5fnf" Feb 18 09:16:56 crc kubenswrapper[4556]: I0218 09:16:56.388266 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ce838404-6d81-41ca-bdbf-e9af260a0d55-operator-scripts\") pod \"keystone-db-create-94t8h\" (UID: \"ce838404-6d81-41ca-bdbf-e9af260a0d55\") " pod="openstack/keystone-db-create-94t8h" Feb 18 09:16:56 crc kubenswrapper[4556]: I0218 09:16:56.388307 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwwdn\" (UniqueName: \"kubernetes.io/projected/45a79e16-1b19-4d21-9edb-ff35eb78d357-kube-api-access-nwwdn\") pod \"keystone-101a-account-create-update-t5fnf\" (UID: \"45a79e16-1b19-4d21-9edb-ff35eb78d357\") " pod="openstack/keystone-101a-account-create-update-t5fnf" Feb 18 09:16:56 crc kubenswrapper[4556]: I0218 09:16:56.388395 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldfwb\" (UniqueName: \"kubernetes.io/projected/ce838404-6d81-41ca-bdbf-e9af260a0d55-kube-api-access-ldfwb\") pod \"keystone-db-create-94t8h\" (UID: \"ce838404-6d81-41ca-bdbf-e9af260a0d55\") " pod="openstack/keystone-db-create-94t8h" Feb 18 09:16:56 crc kubenswrapper[4556]: I0218 09:16:56.389912 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ce838404-6d81-41ca-bdbf-e9af260a0d55-operator-scripts\") pod \"keystone-db-create-94t8h\" (UID: \"ce838404-6d81-41ca-bdbf-e9af260a0d55\") " pod="openstack/keystone-db-create-94t8h" Feb 18 09:16:56 crc kubenswrapper[4556]: I0218 09:16:56.411281 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldfwb\" (UniqueName: \"kubernetes.io/projected/ce838404-6d81-41ca-bdbf-e9af260a0d55-kube-api-access-ldfwb\") pod \"keystone-db-create-94t8h\" (UID: \"ce838404-6d81-41ca-bdbf-e9af260a0d55\") " pod="openstack/keystone-db-create-94t8h" Feb 18 09:16:56 crc kubenswrapper[4556]: I0218 09:16:56.415049 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-fqkrx"] Feb 18 09:16:56 crc kubenswrapper[4556]: I0218 09:16:56.416880 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-fqkrx" Feb 18 09:16:56 crc kubenswrapper[4556]: I0218 09:16:56.425772 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-fqkrx"] Feb 18 09:16:56 crc kubenswrapper[4556]: I0218 09:16:56.491257 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/45a79e16-1b19-4d21-9edb-ff35eb78d357-operator-scripts\") pod \"keystone-101a-account-create-update-t5fnf\" (UID: \"45a79e16-1b19-4d21-9edb-ff35eb78d357\") " pod="openstack/keystone-101a-account-create-update-t5fnf" Feb 18 09:16:56 crc kubenswrapper[4556]: I0218 09:16:56.491319 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/da163339-653b-4057-9c2f-332eb6957f40-etc-swift\") pod \"swift-storage-0\" (UID: \"da163339-653b-4057-9c2f-332eb6957f40\") " pod="openstack/swift-storage-0" Feb 18 09:16:56 crc kubenswrapper[4556]: I0218 09:16:56.491356 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwwdn\" (UniqueName: \"kubernetes.io/projected/45a79e16-1b19-4d21-9edb-ff35eb78d357-kube-api-access-nwwdn\") pod \"keystone-101a-account-create-update-t5fnf\" (UID: \"45a79e16-1b19-4d21-9edb-ff35eb78d357\") " pod="openstack/keystone-101a-account-create-update-t5fnf" Feb 18 09:16:56 crc kubenswrapper[4556]: E0218 09:16:56.491553 4556 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 18 09:16:56 crc kubenswrapper[4556]: E0218 09:16:56.491588 4556 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 18 09:16:56 crc kubenswrapper[4556]: E0218 09:16:56.491641 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/da163339-653b-4057-9c2f-332eb6957f40-etc-swift podName:da163339-653b-4057-9c2f-332eb6957f40 nodeName:}" failed. No retries permitted until 2026-02-18 09:17:04.4916241 +0000 UTC m=+781.508585110 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/da163339-653b-4057-9c2f-332eb6957f40-etc-swift") pod "swift-storage-0" (UID: "da163339-653b-4057-9c2f-332eb6957f40") : configmap "swift-ring-files" not found Feb 18 09:16:56 crc kubenswrapper[4556]: I0218 09:16:56.492007 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/45a79e16-1b19-4d21-9edb-ff35eb78d357-operator-scripts\") pod \"keystone-101a-account-create-update-t5fnf\" (UID: \"45a79e16-1b19-4d21-9edb-ff35eb78d357\") " pod="openstack/keystone-101a-account-create-update-t5fnf" Feb 18 09:16:56 crc kubenswrapper[4556]: I0218 09:16:56.508302 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-94t8h" Feb 18 09:16:56 crc kubenswrapper[4556]: I0218 09:16:56.519522 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwwdn\" (UniqueName: \"kubernetes.io/projected/45a79e16-1b19-4d21-9edb-ff35eb78d357-kube-api-access-nwwdn\") pod \"keystone-101a-account-create-update-t5fnf\" (UID: \"45a79e16-1b19-4d21-9edb-ff35eb78d357\") " pod="openstack/keystone-101a-account-create-update-t5fnf" Feb 18 09:16:56 crc kubenswrapper[4556]: I0218 09:16:56.520477 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-be94-account-create-update-rqlh4"] Feb 18 09:16:56 crc kubenswrapper[4556]: I0218 09:16:56.522494 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-be94-account-create-update-rqlh4" Feb 18 09:16:56 crc kubenswrapper[4556]: I0218 09:16:56.524536 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Feb 18 09:16:56 crc kubenswrapper[4556]: I0218 09:16:56.527746 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-be94-account-create-update-rqlh4"] Feb 18 09:16:56 crc kubenswrapper[4556]: I0218 09:16:56.592619 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkplh\" (UniqueName: \"kubernetes.io/projected/0f905f0b-32d9-40bc-8c19-9cfb9e4f739b-kube-api-access-rkplh\") pod \"placement-db-create-fqkrx\" (UID: \"0f905f0b-32d9-40bc-8c19-9cfb9e4f739b\") " pod="openstack/placement-db-create-fqkrx" Feb 18 09:16:56 crc kubenswrapper[4556]: I0218 09:16:56.592754 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0f905f0b-32d9-40bc-8c19-9cfb9e4f739b-operator-scripts\") pod \"placement-db-create-fqkrx\" (UID: \"0f905f0b-32d9-40bc-8c19-9cfb9e4f739b\") " pod="openstack/placement-db-create-fqkrx" Feb 18 09:16:56 crc kubenswrapper[4556]: I0218 09:16:56.631693 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-101a-account-create-update-t5fnf" Feb 18 09:16:56 crc kubenswrapper[4556]: I0218 09:16:56.694519 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/03d46588-9781-48c2-9ce0-a56681c223c9-operator-scripts\") pod \"placement-be94-account-create-update-rqlh4\" (UID: \"03d46588-9781-48c2-9ce0-a56681c223c9\") " pod="openstack/placement-be94-account-create-update-rqlh4" Feb 18 09:16:56 crc kubenswrapper[4556]: I0218 09:16:56.694680 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkplh\" (UniqueName: \"kubernetes.io/projected/0f905f0b-32d9-40bc-8c19-9cfb9e4f739b-kube-api-access-rkplh\") pod \"placement-db-create-fqkrx\" (UID: \"0f905f0b-32d9-40bc-8c19-9cfb9e4f739b\") " pod="openstack/placement-db-create-fqkrx" Feb 18 09:16:56 crc kubenswrapper[4556]: I0218 09:16:56.694788 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zsvcf\" (UniqueName: \"kubernetes.io/projected/03d46588-9781-48c2-9ce0-a56681c223c9-kube-api-access-zsvcf\") pod \"placement-be94-account-create-update-rqlh4\" (UID: \"03d46588-9781-48c2-9ce0-a56681c223c9\") " pod="openstack/placement-be94-account-create-update-rqlh4" Feb 18 09:16:56 crc kubenswrapper[4556]: I0218 09:16:56.694871 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0f905f0b-32d9-40bc-8c19-9cfb9e4f739b-operator-scripts\") pod \"placement-db-create-fqkrx\" (UID: \"0f905f0b-32d9-40bc-8c19-9cfb9e4f739b\") " pod="openstack/placement-db-create-fqkrx" Feb 18 09:16:56 crc kubenswrapper[4556]: I0218 09:16:56.695989 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0f905f0b-32d9-40bc-8c19-9cfb9e4f739b-operator-scripts\") pod \"placement-db-create-fqkrx\" (UID: \"0f905f0b-32d9-40bc-8c19-9cfb9e4f739b\") " pod="openstack/placement-db-create-fqkrx" Feb 18 09:16:56 crc kubenswrapper[4556]: I0218 09:16:56.710553 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkplh\" (UniqueName: \"kubernetes.io/projected/0f905f0b-32d9-40bc-8c19-9cfb9e4f739b-kube-api-access-rkplh\") pod \"placement-db-create-fqkrx\" (UID: \"0f905f0b-32d9-40bc-8c19-9cfb9e4f739b\") " pod="openstack/placement-db-create-fqkrx" Feb 18 09:16:56 crc kubenswrapper[4556]: I0218 09:16:56.747122 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-fqkrx" Feb 18 09:16:56 crc kubenswrapper[4556]: I0218 09:16:56.796859 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zsvcf\" (UniqueName: \"kubernetes.io/projected/03d46588-9781-48c2-9ce0-a56681c223c9-kube-api-access-zsvcf\") pod \"placement-be94-account-create-update-rqlh4\" (UID: \"03d46588-9781-48c2-9ce0-a56681c223c9\") " pod="openstack/placement-be94-account-create-update-rqlh4" Feb 18 09:16:56 crc kubenswrapper[4556]: I0218 09:16:56.796946 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/03d46588-9781-48c2-9ce0-a56681c223c9-operator-scripts\") pod \"placement-be94-account-create-update-rqlh4\" (UID: \"03d46588-9781-48c2-9ce0-a56681c223c9\") " pod="openstack/placement-be94-account-create-update-rqlh4" Feb 18 09:16:56 crc kubenswrapper[4556]: I0218 09:16:56.797613 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/03d46588-9781-48c2-9ce0-a56681c223c9-operator-scripts\") pod \"placement-be94-account-create-update-rqlh4\" (UID: \"03d46588-9781-48c2-9ce0-a56681c223c9\") " pod="openstack/placement-be94-account-create-update-rqlh4" Feb 18 09:16:56 crc kubenswrapper[4556]: I0218 09:16:56.820071 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zsvcf\" (UniqueName: \"kubernetes.io/projected/03d46588-9781-48c2-9ce0-a56681c223c9-kube-api-access-zsvcf\") pod \"placement-be94-account-create-update-rqlh4\" (UID: \"03d46588-9781-48c2-9ce0-a56681c223c9\") " pod="openstack/placement-be94-account-create-update-rqlh4" Feb 18 09:16:56 crc kubenswrapper[4556]: I0218 09:16:56.859309 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-be94-account-create-update-rqlh4" Feb 18 09:16:57 crc kubenswrapper[4556]: I0218 09:16:57.057407 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-lbxfb" Feb 18 09:16:57 crc kubenswrapper[4556]: I0218 09:16:57.203778 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7pkxz\" (UniqueName: \"kubernetes.io/projected/0537334d-14e5-4ff1-ab08-4c4e7a882c9a-kube-api-access-7pkxz\") pod \"0537334d-14e5-4ff1-ab08-4c4e7a882c9a\" (UID: \"0537334d-14e5-4ff1-ab08-4c4e7a882c9a\") " Feb 18 09:16:57 crc kubenswrapper[4556]: I0218 09:16:57.204044 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0537334d-14e5-4ff1-ab08-4c4e7a882c9a-operator-scripts\") pod \"0537334d-14e5-4ff1-ab08-4c4e7a882c9a\" (UID: \"0537334d-14e5-4ff1-ab08-4c4e7a882c9a\") " Feb 18 09:16:57 crc kubenswrapper[4556]: I0218 09:16:57.204710 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0537334d-14e5-4ff1-ab08-4c4e7a882c9a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0537334d-14e5-4ff1-ab08-4c4e7a882c9a" (UID: "0537334d-14e5-4ff1-ab08-4c4e7a882c9a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:16:57 crc kubenswrapper[4556]: I0218 09:16:57.207399 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0537334d-14e5-4ff1-ab08-4c4e7a882c9a-kube-api-access-7pkxz" (OuterVolumeSpecName: "kube-api-access-7pkxz") pod "0537334d-14e5-4ff1-ab08-4c4e7a882c9a" (UID: "0537334d-14e5-4ff1-ab08-4c4e7a882c9a"). InnerVolumeSpecName "kube-api-access-7pkxz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:16:57 crc kubenswrapper[4556]: I0218 09:16:57.306398 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7pkxz\" (UniqueName: \"kubernetes.io/projected/0537334d-14e5-4ff1-ab08-4c4e7a882c9a-kube-api-access-7pkxz\") on node \"crc\" DevicePath \"\"" Feb 18 09:16:57 crc kubenswrapper[4556]: I0218 09:16:57.306421 4556 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0537334d-14e5-4ff1-ab08-4c4e7a882c9a-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 18 09:16:57 crc kubenswrapper[4556]: I0218 09:16:57.448061 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-8n5wp"] Feb 18 09:16:57 crc kubenswrapper[4556]: I0218 09:16:57.456804 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-de68-account-create-update-f26q6"] Feb 18 09:16:57 crc kubenswrapper[4556]: I0218 09:16:57.537563 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-101a-account-create-update-t5fnf"] Feb 18 09:16:57 crc kubenswrapper[4556]: I0218 09:16:57.541739 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-fqkrx"] Feb 18 09:16:57 crc kubenswrapper[4556]: I0218 09:16:57.641475 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-94t8h"] Feb 18 09:16:57 crc kubenswrapper[4556]: W0218 09:16:57.653442 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podce838404_6d81_41ca_bdbf_e9af260a0d55.slice/crio-361d9fdf8bdd4a91bb40ad117a5c7804993509173db06e574161f3fd6c1f3fa1 WatchSource:0}: Error finding container 361d9fdf8bdd4a91bb40ad117a5c7804993509173db06e574161f3fd6c1f3fa1: Status 404 returned error can't find the container with id 361d9fdf8bdd4a91bb40ad117a5c7804993509173db06e574161f3fd6c1f3fa1 Feb 18 09:16:57 crc kubenswrapper[4556]: I0218 09:16:57.670982 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-be94-account-create-update-rqlh4"] Feb 18 09:16:57 crc kubenswrapper[4556]: W0218 09:16:57.683167 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod03d46588_9781_48c2_9ce0_a56681c223c9.slice/crio-e367860a749be3bea330f67d7236cbfba9304084578fc79875b5f3d3454eb27a WatchSource:0}: Error finding container e367860a749be3bea330f67d7236cbfba9304084578fc79875b5f3d3454eb27a: Status 404 returned error can't find the container with id e367860a749be3bea330f67d7236cbfba9304084578fc79875b5f3d3454eb27a Feb 18 09:16:57 crc kubenswrapper[4556]: I0218 09:16:57.819253 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-de68-account-create-update-f26q6" event={"ID":"8f7157d7-e589-4d41-955e-73ad86618b83","Type":"ContainerStarted","Data":"3933b7ed7b0d41f7f8756aeadc1b735fa65bb0952d7a007f2153c06e853c694b"} Feb 18 09:16:57 crc kubenswrapper[4556]: I0218 09:16:57.819315 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-de68-account-create-update-f26q6" event={"ID":"8f7157d7-e589-4d41-955e-73ad86618b83","Type":"ContainerStarted","Data":"e1dfee45090ef16c79c151d549b948ab4c93a8acd2703ac697f66380644c5cb0"} Feb 18 09:16:57 crc kubenswrapper[4556]: I0218 09:16:57.821453 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-be94-account-create-update-rqlh4" event={"ID":"03d46588-9781-48c2-9ce0-a56681c223c9","Type":"ContainerStarted","Data":"e367860a749be3bea330f67d7236cbfba9304084578fc79875b5f3d3454eb27a"} Feb 18 09:16:57 crc kubenswrapper[4556]: I0218 09:16:57.823828 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-94t8h" event={"ID":"ce838404-6d81-41ca-bdbf-e9af260a0d55","Type":"ContainerStarted","Data":"361d9fdf8bdd4a91bb40ad117a5c7804993509173db06e574161f3fd6c1f3fa1"} Feb 18 09:16:57 crc kubenswrapper[4556]: I0218 09:16:57.830277 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-8n5wp" event={"ID":"b43cd61b-beae-4ef8-8648-d014d045e56c","Type":"ContainerStarted","Data":"61012520598e8019f9b5f64bf562fa2834f8935a374317944e30a9c6c3ca3a3c"} Feb 18 09:16:57 crc kubenswrapper[4556]: I0218 09:16:57.830308 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-8n5wp" event={"ID":"b43cd61b-beae-4ef8-8648-d014d045e56c","Type":"ContainerStarted","Data":"b2356cd316d50d613e71ff38b506a9cab8f2cf3ed885b981a1ca9c0d40fc0db9"} Feb 18 09:16:57 crc kubenswrapper[4556]: I0218 09:16:57.833936 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-9z5cc" event={"ID":"03ccffdf-f11c-4788-9e22-0b0661665c5d","Type":"ContainerStarted","Data":"dba6fea7c4013cc379aaaea6e6615b0125f4f7197cb5831bee26f4a54ef5bd6c"} Feb 18 09:16:57 crc kubenswrapper[4556]: I0218 09:16:57.838913 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-101a-account-create-update-t5fnf" event={"ID":"45a79e16-1b19-4d21-9edb-ff35eb78d357","Type":"ContainerStarted","Data":"b40a41ffea6945c4c57ebb77f95f33e5dea3a98e76eb30b2a22fcdaf4d9d3cbf"} Feb 18 09:16:57 crc kubenswrapper[4556]: I0218 09:16:57.838947 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-101a-account-create-update-t5fnf" event={"ID":"45a79e16-1b19-4d21-9edb-ff35eb78d357","Type":"ContainerStarted","Data":"c2da0e2ea0fefcfa566beefaf3543e90b019f021cdb6cb303ea4248202f60453"} Feb 18 09:16:57 crc kubenswrapper[4556]: I0218 09:16:57.841472 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-de68-account-create-update-f26q6" podStartSLOduration=2.841449804 podStartE2EDuration="2.841449804s" podCreationTimestamp="2026-02-18 09:16:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:16:57.834950865 +0000 UTC m=+774.851911845" watchObservedRunningTime="2026-02-18 09:16:57.841449804 +0000 UTC m=+774.858410783" Feb 18 09:16:57 crc kubenswrapper[4556]: I0218 09:16:57.845915 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-lbxfb" Feb 18 09:16:57 crc kubenswrapper[4556]: I0218 09:16:57.846650 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-lbxfb" event={"ID":"0537334d-14e5-4ff1-ab08-4c4e7a882c9a","Type":"ContainerDied","Data":"b6fc0becf112d0cf2a50d883daae1e0cd5ea68f49d82e8ed944126faa9bcf621"} Feb 18 09:16:57 crc kubenswrapper[4556]: I0218 09:16:57.846694 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b6fc0becf112d0cf2a50d883daae1e0cd5ea68f49d82e8ed944126faa9bcf621" Feb 18 09:16:57 crc kubenswrapper[4556]: I0218 09:16:57.859147 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-fqkrx" event={"ID":"0f905f0b-32d9-40bc-8c19-9cfb9e4f739b","Type":"ContainerStarted","Data":"b6e2dd2b3609e35ad44eba00703cf72f2e7c5e572459442f7fd721df57ee204e"} Feb 18 09:16:57 crc kubenswrapper[4556]: I0218 09:16:57.859222 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-fqkrx" event={"ID":"0f905f0b-32d9-40bc-8c19-9cfb9e4f739b","Type":"ContainerStarted","Data":"ea9b6cacd688ce4c821724b217147a0f2a424c58c2a97c255888c41644dee5c4"} Feb 18 09:16:57 crc kubenswrapper[4556]: I0218 09:16:57.872296 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-create-8n5wp" podStartSLOduration=2.872272282 podStartE2EDuration="2.872272282s" podCreationTimestamp="2026-02-18 09:16:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:16:57.85418762 +0000 UTC m=+774.871148600" watchObservedRunningTime="2026-02-18 09:16:57.872272282 +0000 UTC m=+774.889233262" Feb 18 09:16:57 crc kubenswrapper[4556]: I0218 09:16:57.886375 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-9z5cc" podStartSLOduration=4.4433739039999995 podStartE2EDuration="8.88635633s" podCreationTimestamp="2026-02-18 09:16:49 +0000 UTC" firstStartedPulling="2026-02-18 09:16:52.596988374 +0000 UTC m=+769.613949354" lastFinishedPulling="2026-02-18 09:16:57.0399708 +0000 UTC m=+774.056931780" observedRunningTime="2026-02-18 09:16:57.881187632 +0000 UTC m=+774.898148613" watchObservedRunningTime="2026-02-18 09:16:57.88635633 +0000 UTC m=+774.903317310" Feb 18 09:16:57 crc kubenswrapper[4556]: I0218 09:16:57.916937 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-fqkrx" podStartSLOduration=1.916904981 podStartE2EDuration="1.916904981s" podCreationTimestamp="2026-02-18 09:16:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:16:57.898379748 +0000 UTC m=+774.915340728" watchObservedRunningTime="2026-02-18 09:16:57.916904981 +0000 UTC m=+774.933865961" Feb 18 09:16:57 crc kubenswrapper[4556]: I0218 09:16:57.923448 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-101a-account-create-update-t5fnf" podStartSLOduration=1.9234402579999998 podStartE2EDuration="1.923440258s" podCreationTimestamp="2026-02-18 09:16:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:16:57.909823211 +0000 UTC m=+774.926784191" watchObservedRunningTime="2026-02-18 09:16:57.923440258 +0000 UTC m=+774.940401238" Feb 18 09:16:58 crc kubenswrapper[4556]: I0218 09:16:58.868376 4556 generic.go:334] "Generic (PLEG): container finished" podID="45a79e16-1b19-4d21-9edb-ff35eb78d357" containerID="b40a41ffea6945c4c57ebb77f95f33e5dea3a98e76eb30b2a22fcdaf4d9d3cbf" exitCode=0 Feb 18 09:16:58 crc kubenswrapper[4556]: I0218 09:16:58.868732 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-101a-account-create-update-t5fnf" event={"ID":"45a79e16-1b19-4d21-9edb-ff35eb78d357","Type":"ContainerDied","Data":"b40a41ffea6945c4c57ebb77f95f33e5dea3a98e76eb30b2a22fcdaf4d9d3cbf"} Feb 18 09:16:58 crc kubenswrapper[4556]: I0218 09:16:58.871051 4556 generic.go:334] "Generic (PLEG): container finished" podID="0f905f0b-32d9-40bc-8c19-9cfb9e4f739b" containerID="b6e2dd2b3609e35ad44eba00703cf72f2e7c5e572459442f7fd721df57ee204e" exitCode=0 Feb 18 09:16:58 crc kubenswrapper[4556]: I0218 09:16:58.871097 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-fqkrx" event={"ID":"0f905f0b-32d9-40bc-8c19-9cfb9e4f739b","Type":"ContainerDied","Data":"b6e2dd2b3609e35ad44eba00703cf72f2e7c5e572459442f7fd721df57ee204e"} Feb 18 09:16:58 crc kubenswrapper[4556]: I0218 09:16:58.872859 4556 generic.go:334] "Generic (PLEG): container finished" podID="8f7157d7-e589-4d41-955e-73ad86618b83" containerID="3933b7ed7b0d41f7f8756aeadc1b735fa65bb0952d7a007f2153c06e853c694b" exitCode=0 Feb 18 09:16:58 crc kubenswrapper[4556]: I0218 09:16:58.872900 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-de68-account-create-update-f26q6" event={"ID":"8f7157d7-e589-4d41-955e-73ad86618b83","Type":"ContainerDied","Data":"3933b7ed7b0d41f7f8756aeadc1b735fa65bb0952d7a007f2153c06e853c694b"} Feb 18 09:16:58 crc kubenswrapper[4556]: I0218 09:16:58.874380 4556 generic.go:334] "Generic (PLEG): container finished" podID="03d46588-9781-48c2-9ce0-a56681c223c9" containerID="e053ef5cf1cf63f13f45392300ae28da28077133bd12a872132c875a0fc3bab0" exitCode=0 Feb 18 09:16:58 crc kubenswrapper[4556]: I0218 09:16:58.874420 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-be94-account-create-update-rqlh4" event={"ID":"03d46588-9781-48c2-9ce0-a56681c223c9","Type":"ContainerDied","Data":"e053ef5cf1cf63f13f45392300ae28da28077133bd12a872132c875a0fc3bab0"} Feb 18 09:16:58 crc kubenswrapper[4556]: I0218 09:16:58.875821 4556 generic.go:334] "Generic (PLEG): container finished" podID="b43cd61b-beae-4ef8-8648-d014d045e56c" containerID="61012520598e8019f9b5f64bf562fa2834f8935a374317944e30a9c6c3ca3a3c" exitCode=0 Feb 18 09:16:58 crc kubenswrapper[4556]: I0218 09:16:58.875859 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-8n5wp" event={"ID":"b43cd61b-beae-4ef8-8648-d014d045e56c","Type":"ContainerDied","Data":"61012520598e8019f9b5f64bf562fa2834f8935a374317944e30a9c6c3ca3a3c"} Feb 18 09:16:58 crc kubenswrapper[4556]: I0218 09:16:58.878087 4556 generic.go:334] "Generic (PLEG): container finished" podID="ce838404-6d81-41ca-bdbf-e9af260a0d55" containerID="513593e18b400b5a6800acf343dac7b309280f2ceafe96d0317252c54847bc60" exitCode=0 Feb 18 09:16:58 crc kubenswrapper[4556]: I0218 09:16:58.878956 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-94t8h" event={"ID":"ce838404-6d81-41ca-bdbf-e9af260a0d55","Type":"ContainerDied","Data":"513593e18b400b5a6800acf343dac7b309280f2ceafe96d0317252c54847bc60"} Feb 18 09:17:00 crc kubenswrapper[4556]: I0218 09:17:00.236583 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-de68-account-create-update-f26q6" Feb 18 09:17:00 crc kubenswrapper[4556]: I0218 09:17:00.382838 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rsrc5\" (UniqueName: \"kubernetes.io/projected/8f7157d7-e589-4d41-955e-73ad86618b83-kube-api-access-rsrc5\") pod \"8f7157d7-e589-4d41-955e-73ad86618b83\" (UID: \"8f7157d7-e589-4d41-955e-73ad86618b83\") " Feb 18 09:17:00 crc kubenswrapper[4556]: I0218 09:17:00.383039 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8f7157d7-e589-4d41-955e-73ad86618b83-operator-scripts\") pod \"8f7157d7-e589-4d41-955e-73ad86618b83\" (UID: \"8f7157d7-e589-4d41-955e-73ad86618b83\") " Feb 18 09:17:00 crc kubenswrapper[4556]: I0218 09:17:00.385071 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f7157d7-e589-4d41-955e-73ad86618b83-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8f7157d7-e589-4d41-955e-73ad86618b83" (UID: "8f7157d7-e589-4d41-955e-73ad86618b83"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:17:00 crc kubenswrapper[4556]: I0218 09:17:00.393875 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f7157d7-e589-4d41-955e-73ad86618b83-kube-api-access-rsrc5" (OuterVolumeSpecName: "kube-api-access-rsrc5") pod "8f7157d7-e589-4d41-955e-73ad86618b83" (UID: "8f7157d7-e589-4d41-955e-73ad86618b83"). InnerVolumeSpecName "kube-api-access-rsrc5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:17:00 crc kubenswrapper[4556]: I0218 09:17:00.475349 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-be94-account-create-update-rqlh4" Feb 18 09:17:00 crc kubenswrapper[4556]: I0218 09:17:00.480862 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-94t8h" Feb 18 09:17:00 crc kubenswrapper[4556]: I0218 09:17:00.486958 4556 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8f7157d7-e589-4d41-955e-73ad86618b83-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:00 crc kubenswrapper[4556]: I0218 09:17:00.486985 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rsrc5\" (UniqueName: \"kubernetes.io/projected/8f7157d7-e589-4d41-955e-73ad86618b83-kube-api-access-rsrc5\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:00 crc kubenswrapper[4556]: I0218 09:17:00.487668 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-fqkrx" Feb 18 09:17:00 crc kubenswrapper[4556]: I0218 09:17:00.490408 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-8n5wp" Feb 18 09:17:00 crc kubenswrapper[4556]: I0218 09:17:00.496930 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-101a-account-create-update-t5fnf" Feb 18 09:17:00 crc kubenswrapper[4556]: I0218 09:17:00.588503 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/03d46588-9781-48c2-9ce0-a56681c223c9-operator-scripts\") pod \"03d46588-9781-48c2-9ce0-a56681c223c9\" (UID: \"03d46588-9781-48c2-9ce0-a56681c223c9\") " Feb 18 09:17:00 crc kubenswrapper[4556]: I0218 09:17:00.588818 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0f905f0b-32d9-40bc-8c19-9cfb9e4f739b-operator-scripts\") pod \"0f905f0b-32d9-40bc-8c19-9cfb9e4f739b\" (UID: \"0f905f0b-32d9-40bc-8c19-9cfb9e4f739b\") " Feb 18 09:17:00 crc kubenswrapper[4556]: I0218 09:17:00.588920 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zsvcf\" (UniqueName: \"kubernetes.io/projected/03d46588-9781-48c2-9ce0-a56681c223c9-kube-api-access-zsvcf\") pod \"03d46588-9781-48c2-9ce0-a56681c223c9\" (UID: \"03d46588-9781-48c2-9ce0-a56681c223c9\") " Feb 18 09:17:00 crc kubenswrapper[4556]: I0218 09:17:00.588968 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rkplh\" (UniqueName: \"kubernetes.io/projected/0f905f0b-32d9-40bc-8c19-9cfb9e4f739b-kube-api-access-rkplh\") pod \"0f905f0b-32d9-40bc-8c19-9cfb9e4f739b\" (UID: \"0f905f0b-32d9-40bc-8c19-9cfb9e4f739b\") " Feb 18 09:17:00 crc kubenswrapper[4556]: I0218 09:17:00.588997 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ldfwb\" (UniqueName: \"kubernetes.io/projected/ce838404-6d81-41ca-bdbf-e9af260a0d55-kube-api-access-ldfwb\") pod \"ce838404-6d81-41ca-bdbf-e9af260a0d55\" (UID: \"ce838404-6d81-41ca-bdbf-e9af260a0d55\") " Feb 18 09:17:00 crc kubenswrapper[4556]: I0218 09:17:00.589058 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ce838404-6d81-41ca-bdbf-e9af260a0d55-operator-scripts\") pod \"ce838404-6d81-41ca-bdbf-e9af260a0d55\" (UID: \"ce838404-6d81-41ca-bdbf-e9af260a0d55\") " Feb 18 09:17:00 crc kubenswrapper[4556]: I0218 09:17:00.589503 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f905f0b-32d9-40bc-8c19-9cfb9e4f739b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0f905f0b-32d9-40bc-8c19-9cfb9e4f739b" (UID: "0f905f0b-32d9-40bc-8c19-9cfb9e4f739b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:17:00 crc kubenswrapper[4556]: I0218 09:17:00.589896 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce838404-6d81-41ca-bdbf-e9af260a0d55-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ce838404-6d81-41ca-bdbf-e9af260a0d55" (UID: "ce838404-6d81-41ca-bdbf-e9af260a0d55"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:17:00 crc kubenswrapper[4556]: I0218 09:17:00.589918 4556 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0f905f0b-32d9-40bc-8c19-9cfb9e4f739b-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:00 crc kubenswrapper[4556]: I0218 09:17:00.590029 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03d46588-9781-48c2-9ce0-a56681c223c9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "03d46588-9781-48c2-9ce0-a56681c223c9" (UID: "03d46588-9781-48c2-9ce0-a56681c223c9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:17:00 crc kubenswrapper[4556]: I0218 09:17:00.592575 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03d46588-9781-48c2-9ce0-a56681c223c9-kube-api-access-zsvcf" (OuterVolumeSpecName: "kube-api-access-zsvcf") pod "03d46588-9781-48c2-9ce0-a56681c223c9" (UID: "03d46588-9781-48c2-9ce0-a56681c223c9"). InnerVolumeSpecName "kube-api-access-zsvcf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:17:00 crc kubenswrapper[4556]: I0218 09:17:00.592703 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce838404-6d81-41ca-bdbf-e9af260a0d55-kube-api-access-ldfwb" (OuterVolumeSpecName: "kube-api-access-ldfwb") pod "ce838404-6d81-41ca-bdbf-e9af260a0d55" (UID: "ce838404-6d81-41ca-bdbf-e9af260a0d55"). InnerVolumeSpecName "kube-api-access-ldfwb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:17:00 crc kubenswrapper[4556]: I0218 09:17:00.592985 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f905f0b-32d9-40bc-8c19-9cfb9e4f739b-kube-api-access-rkplh" (OuterVolumeSpecName: "kube-api-access-rkplh") pod "0f905f0b-32d9-40bc-8c19-9cfb9e4f739b" (UID: "0f905f0b-32d9-40bc-8c19-9cfb9e4f739b"). InnerVolumeSpecName "kube-api-access-rkplh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:17:00 crc kubenswrapper[4556]: I0218 09:17:00.690922 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b43cd61b-beae-4ef8-8648-d014d045e56c-operator-scripts\") pod \"b43cd61b-beae-4ef8-8648-d014d045e56c\" (UID: \"b43cd61b-beae-4ef8-8648-d014d045e56c\") " Feb 18 09:17:00 crc kubenswrapper[4556]: I0218 09:17:00.691139 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/45a79e16-1b19-4d21-9edb-ff35eb78d357-operator-scripts\") pod \"45a79e16-1b19-4d21-9edb-ff35eb78d357\" (UID: \"45a79e16-1b19-4d21-9edb-ff35eb78d357\") " Feb 18 09:17:00 crc kubenswrapper[4556]: I0218 09:17:00.691217 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nwwdn\" (UniqueName: \"kubernetes.io/projected/45a79e16-1b19-4d21-9edb-ff35eb78d357-kube-api-access-nwwdn\") pod \"45a79e16-1b19-4d21-9edb-ff35eb78d357\" (UID: \"45a79e16-1b19-4d21-9edb-ff35eb78d357\") " Feb 18 09:17:00 crc kubenswrapper[4556]: I0218 09:17:00.691295 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fxvxx\" (UniqueName: \"kubernetes.io/projected/b43cd61b-beae-4ef8-8648-d014d045e56c-kube-api-access-fxvxx\") pod \"b43cd61b-beae-4ef8-8648-d014d045e56c\" (UID: \"b43cd61b-beae-4ef8-8648-d014d045e56c\") " Feb 18 09:17:00 crc kubenswrapper[4556]: I0218 09:17:00.691598 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b43cd61b-beae-4ef8-8648-d014d045e56c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b43cd61b-beae-4ef8-8648-d014d045e56c" (UID: "b43cd61b-beae-4ef8-8648-d014d045e56c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:17:00 crc kubenswrapper[4556]: I0218 09:17:00.691635 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/45a79e16-1b19-4d21-9edb-ff35eb78d357-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "45a79e16-1b19-4d21-9edb-ff35eb78d357" (UID: "45a79e16-1b19-4d21-9edb-ff35eb78d357"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:17:00 crc kubenswrapper[4556]: I0218 09:17:00.692141 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zsvcf\" (UniqueName: \"kubernetes.io/projected/03d46588-9781-48c2-9ce0-a56681c223c9-kube-api-access-zsvcf\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:00 crc kubenswrapper[4556]: I0218 09:17:00.692180 4556 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b43cd61b-beae-4ef8-8648-d014d045e56c-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:00 crc kubenswrapper[4556]: I0218 09:17:00.692191 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rkplh\" (UniqueName: \"kubernetes.io/projected/0f905f0b-32d9-40bc-8c19-9cfb9e4f739b-kube-api-access-rkplh\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:00 crc kubenswrapper[4556]: I0218 09:17:00.692202 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ldfwb\" (UniqueName: \"kubernetes.io/projected/ce838404-6d81-41ca-bdbf-e9af260a0d55-kube-api-access-ldfwb\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:00 crc kubenswrapper[4556]: I0218 09:17:00.692211 4556 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ce838404-6d81-41ca-bdbf-e9af260a0d55-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:00 crc kubenswrapper[4556]: I0218 09:17:00.692220 4556 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/45a79e16-1b19-4d21-9edb-ff35eb78d357-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:00 crc kubenswrapper[4556]: I0218 09:17:00.692229 4556 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/03d46588-9781-48c2-9ce0-a56681c223c9-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:00 crc kubenswrapper[4556]: I0218 09:17:00.694338 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45a79e16-1b19-4d21-9edb-ff35eb78d357-kube-api-access-nwwdn" (OuterVolumeSpecName: "kube-api-access-nwwdn") pod "45a79e16-1b19-4d21-9edb-ff35eb78d357" (UID: "45a79e16-1b19-4d21-9edb-ff35eb78d357"). InnerVolumeSpecName "kube-api-access-nwwdn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:17:00 crc kubenswrapper[4556]: I0218 09:17:00.694534 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b43cd61b-beae-4ef8-8648-d014d045e56c-kube-api-access-fxvxx" (OuterVolumeSpecName: "kube-api-access-fxvxx") pod "b43cd61b-beae-4ef8-8648-d014d045e56c" (UID: "b43cd61b-beae-4ef8-8648-d014d045e56c"). InnerVolumeSpecName "kube-api-access-fxvxx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:17:00 crc kubenswrapper[4556]: I0218 09:17:00.793554 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nwwdn\" (UniqueName: \"kubernetes.io/projected/45a79e16-1b19-4d21-9edb-ff35eb78d357-kube-api-access-nwwdn\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:00 crc kubenswrapper[4556]: I0218 09:17:00.793584 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fxvxx\" (UniqueName: \"kubernetes.io/projected/b43cd61b-beae-4ef8-8648-d014d045e56c-kube-api-access-fxvxx\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:00 crc kubenswrapper[4556]: I0218 09:17:00.898421 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-8n5wp" event={"ID":"b43cd61b-beae-4ef8-8648-d014d045e56c","Type":"ContainerDied","Data":"b2356cd316d50d613e71ff38b506a9cab8f2cf3ed885b981a1ca9c0d40fc0db9"} Feb 18 09:17:00 crc kubenswrapper[4556]: I0218 09:17:00.898495 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b2356cd316d50d613e71ff38b506a9cab8f2cf3ed885b981a1ca9c0d40fc0db9" Feb 18 09:17:00 crc kubenswrapper[4556]: I0218 09:17:00.898740 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-8n5wp" Feb 18 09:17:00 crc kubenswrapper[4556]: I0218 09:17:00.900210 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-94t8h" event={"ID":"ce838404-6d81-41ca-bdbf-e9af260a0d55","Type":"ContainerDied","Data":"361d9fdf8bdd4a91bb40ad117a5c7804993509173db06e574161f3fd6c1f3fa1"} Feb 18 09:17:00 crc kubenswrapper[4556]: I0218 09:17:00.900237 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-94t8h" Feb 18 09:17:00 crc kubenswrapper[4556]: I0218 09:17:00.900246 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="361d9fdf8bdd4a91bb40ad117a5c7804993509173db06e574161f3fd6c1f3fa1" Feb 18 09:17:00 crc kubenswrapper[4556]: I0218 09:17:00.901940 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-101a-account-create-update-t5fnf" event={"ID":"45a79e16-1b19-4d21-9edb-ff35eb78d357","Type":"ContainerDied","Data":"c2da0e2ea0fefcfa566beefaf3543e90b019f021cdb6cb303ea4248202f60453"} Feb 18 09:17:00 crc kubenswrapper[4556]: I0218 09:17:00.902001 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-101a-account-create-update-t5fnf" Feb 18 09:17:00 crc kubenswrapper[4556]: I0218 09:17:00.901966 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c2da0e2ea0fefcfa566beefaf3543e90b019f021cdb6cb303ea4248202f60453" Feb 18 09:17:00 crc kubenswrapper[4556]: I0218 09:17:00.904437 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-fqkrx" event={"ID":"0f905f0b-32d9-40bc-8c19-9cfb9e4f739b","Type":"ContainerDied","Data":"ea9b6cacd688ce4c821724b217147a0f2a424c58c2a97c255888c41644dee5c4"} Feb 18 09:17:00 crc kubenswrapper[4556]: I0218 09:17:00.904504 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ea9b6cacd688ce4c821724b217147a0f2a424c58c2a97c255888c41644dee5c4" Feb 18 09:17:00 crc kubenswrapper[4556]: I0218 09:17:00.904669 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-fqkrx" Feb 18 09:17:00 crc kubenswrapper[4556]: I0218 09:17:00.906992 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-de68-account-create-update-f26q6" Feb 18 09:17:00 crc kubenswrapper[4556]: I0218 09:17:00.907040 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-de68-account-create-update-f26q6" event={"ID":"8f7157d7-e589-4d41-955e-73ad86618b83","Type":"ContainerDied","Data":"e1dfee45090ef16c79c151d549b948ab4c93a8acd2703ac697f66380644c5cb0"} Feb 18 09:17:00 crc kubenswrapper[4556]: I0218 09:17:00.907095 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e1dfee45090ef16c79c151d549b948ab4c93a8acd2703ac697f66380644c5cb0" Feb 18 09:17:00 crc kubenswrapper[4556]: I0218 09:17:00.908806 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-be94-account-create-update-rqlh4" event={"ID":"03d46588-9781-48c2-9ce0-a56681c223c9","Type":"ContainerDied","Data":"e367860a749be3bea330f67d7236cbfba9304084578fc79875b5f3d3454eb27a"} Feb 18 09:17:00 crc kubenswrapper[4556]: I0218 09:17:00.909003 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e367860a749be3bea330f67d7236cbfba9304084578fc79875b5f3d3454eb27a" Feb 18 09:17:00 crc kubenswrapper[4556]: I0218 09:17:00.908963 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-be94-account-create-update-rqlh4" Feb 18 09:17:02 crc kubenswrapper[4556]: I0218 09:17:02.272381 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-lbxfb"] Feb 18 09:17:02 crc kubenswrapper[4556]: I0218 09:17:02.277345 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-lbxfb"] Feb 18 09:17:02 crc kubenswrapper[4556]: I0218 09:17:02.348587 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-blbjd"] Feb 18 09:17:02 crc kubenswrapper[4556]: E0218 09:17:02.348849 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0537334d-14e5-4ff1-ab08-4c4e7a882c9a" containerName="mariadb-account-create-update" Feb 18 09:17:02 crc kubenswrapper[4556]: I0218 09:17:02.348864 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="0537334d-14e5-4ff1-ab08-4c4e7a882c9a" containerName="mariadb-account-create-update" Feb 18 09:17:02 crc kubenswrapper[4556]: E0218 09:17:02.348887 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45a79e16-1b19-4d21-9edb-ff35eb78d357" containerName="mariadb-account-create-update" Feb 18 09:17:02 crc kubenswrapper[4556]: I0218 09:17:02.348893 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="45a79e16-1b19-4d21-9edb-ff35eb78d357" containerName="mariadb-account-create-update" Feb 18 09:17:02 crc kubenswrapper[4556]: E0218 09:17:02.348899 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f7157d7-e589-4d41-955e-73ad86618b83" containerName="mariadb-account-create-update" Feb 18 09:17:02 crc kubenswrapper[4556]: I0218 09:17:02.348905 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f7157d7-e589-4d41-955e-73ad86618b83" containerName="mariadb-account-create-update" Feb 18 09:17:02 crc kubenswrapper[4556]: E0218 09:17:02.348915 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce838404-6d81-41ca-bdbf-e9af260a0d55" containerName="mariadb-database-create" Feb 18 09:17:02 crc kubenswrapper[4556]: I0218 09:17:02.348920 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce838404-6d81-41ca-bdbf-e9af260a0d55" containerName="mariadb-database-create" Feb 18 09:17:02 crc kubenswrapper[4556]: E0218 09:17:02.348930 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03d46588-9781-48c2-9ce0-a56681c223c9" containerName="mariadb-account-create-update" Feb 18 09:17:02 crc kubenswrapper[4556]: I0218 09:17:02.348935 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="03d46588-9781-48c2-9ce0-a56681c223c9" containerName="mariadb-account-create-update" Feb 18 09:17:02 crc kubenswrapper[4556]: E0218 09:17:02.348956 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f905f0b-32d9-40bc-8c19-9cfb9e4f739b" containerName="mariadb-database-create" Feb 18 09:17:02 crc kubenswrapper[4556]: I0218 09:17:02.348963 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f905f0b-32d9-40bc-8c19-9cfb9e4f739b" containerName="mariadb-database-create" Feb 18 09:17:02 crc kubenswrapper[4556]: E0218 09:17:02.348978 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b43cd61b-beae-4ef8-8648-d014d045e56c" containerName="mariadb-database-create" Feb 18 09:17:02 crc kubenswrapper[4556]: I0218 09:17:02.348983 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="b43cd61b-beae-4ef8-8648-d014d045e56c" containerName="mariadb-database-create" Feb 18 09:17:02 crc kubenswrapper[4556]: I0218 09:17:02.349108 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="45a79e16-1b19-4d21-9edb-ff35eb78d357" containerName="mariadb-account-create-update" Feb 18 09:17:02 crc kubenswrapper[4556]: I0218 09:17:02.349119 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f7157d7-e589-4d41-955e-73ad86618b83" containerName="mariadb-account-create-update" Feb 18 09:17:02 crc kubenswrapper[4556]: I0218 09:17:02.349129 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="0537334d-14e5-4ff1-ab08-4c4e7a882c9a" containerName="mariadb-account-create-update" Feb 18 09:17:02 crc kubenswrapper[4556]: I0218 09:17:02.349140 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="03d46588-9781-48c2-9ce0-a56681c223c9" containerName="mariadb-account-create-update" Feb 18 09:17:02 crc kubenswrapper[4556]: I0218 09:17:02.349167 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="b43cd61b-beae-4ef8-8648-d014d045e56c" containerName="mariadb-database-create" Feb 18 09:17:02 crc kubenswrapper[4556]: I0218 09:17:02.349174 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f905f0b-32d9-40bc-8c19-9cfb9e4f739b" containerName="mariadb-database-create" Feb 18 09:17:02 crc kubenswrapper[4556]: I0218 09:17:02.349186 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce838404-6d81-41ca-bdbf-e9af260a0d55" containerName="mariadb-database-create" Feb 18 09:17:02 crc kubenswrapper[4556]: I0218 09:17:02.349574 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-blbjd" Feb 18 09:17:02 crc kubenswrapper[4556]: I0218 09:17:02.350974 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Feb 18 09:17:02 crc kubenswrapper[4556]: I0218 09:17:02.361514 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-blbjd"] Feb 18 09:17:02 crc kubenswrapper[4556]: I0218 09:17:02.422548 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dszlk\" (UniqueName: \"kubernetes.io/projected/7a185e40-8a02-44a1-ab00-2575bebde1d5-kube-api-access-dszlk\") pod \"root-account-create-update-blbjd\" (UID: \"7a185e40-8a02-44a1-ab00-2575bebde1d5\") " pod="openstack/root-account-create-update-blbjd" Feb 18 09:17:02 crc kubenswrapper[4556]: I0218 09:17:02.422606 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7a185e40-8a02-44a1-ab00-2575bebde1d5-operator-scripts\") pod \"root-account-create-update-blbjd\" (UID: \"7a185e40-8a02-44a1-ab00-2575bebde1d5\") " pod="openstack/root-account-create-update-blbjd" Feb 18 09:17:02 crc kubenswrapper[4556]: I0218 09:17:02.523786 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dszlk\" (UniqueName: \"kubernetes.io/projected/7a185e40-8a02-44a1-ab00-2575bebde1d5-kube-api-access-dszlk\") pod \"root-account-create-update-blbjd\" (UID: \"7a185e40-8a02-44a1-ab00-2575bebde1d5\") " pod="openstack/root-account-create-update-blbjd" Feb 18 09:17:02 crc kubenswrapper[4556]: I0218 09:17:02.523843 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7a185e40-8a02-44a1-ab00-2575bebde1d5-operator-scripts\") pod \"root-account-create-update-blbjd\" (UID: \"7a185e40-8a02-44a1-ab00-2575bebde1d5\") " pod="openstack/root-account-create-update-blbjd" Feb 18 09:17:02 crc kubenswrapper[4556]: I0218 09:17:02.524769 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7a185e40-8a02-44a1-ab00-2575bebde1d5-operator-scripts\") pod \"root-account-create-update-blbjd\" (UID: \"7a185e40-8a02-44a1-ab00-2575bebde1d5\") " pod="openstack/root-account-create-update-blbjd" Feb 18 09:17:02 crc kubenswrapper[4556]: I0218 09:17:02.541319 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dszlk\" (UniqueName: \"kubernetes.io/projected/7a185e40-8a02-44a1-ab00-2575bebde1d5-kube-api-access-dszlk\") pod \"root-account-create-update-blbjd\" (UID: \"7a185e40-8a02-44a1-ab00-2575bebde1d5\") " pod="openstack/root-account-create-update-blbjd" Feb 18 09:17:02 crc kubenswrapper[4556]: I0218 09:17:02.665647 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-blbjd" Feb 18 09:17:02 crc kubenswrapper[4556]: I0218 09:17:02.929763 4556 generic.go:334] "Generic (PLEG): container finished" podID="03ccffdf-f11c-4788-9e22-0b0661665c5d" containerID="dba6fea7c4013cc379aaaea6e6615b0125f4f7197cb5831bee26f4a54ef5bd6c" exitCode=0 Feb 18 09:17:02 crc kubenswrapper[4556]: I0218 09:17:02.929855 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-9z5cc" event={"ID":"03ccffdf-f11c-4788-9e22-0b0661665c5d","Type":"ContainerDied","Data":"dba6fea7c4013cc379aaaea6e6615b0125f4f7197cb5831bee26f4a54ef5bd6c"} Feb 18 09:17:02 crc kubenswrapper[4556]: I0218 09:17:02.977312 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-66b577f8c-9dkbr" Feb 18 09:17:03 crc kubenswrapper[4556]: I0218 09:17:03.019375 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64f7f48db9-mjw87"] Feb 18 09:17:03 crc kubenswrapper[4556]: I0218 09:17:03.019654 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-64f7f48db9-mjw87" podUID="30a1b29b-530b-4f64-acb7-a8471fc98fcb" containerName="dnsmasq-dns" containerID="cri-o://833d74497abfc08cdfa0f0c32d15a296083d4ae387fbc6542d6c8810ddc63fa8" gracePeriod=10 Feb 18 09:17:03 crc kubenswrapper[4556]: I0218 09:17:03.069039 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-blbjd"] Feb 18 09:17:03 crc kubenswrapper[4556]: W0218 09:17:03.092938 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7a185e40_8a02_44a1_ab00_2575bebde1d5.slice/crio-2103832d63f95c8337e0896c4056897296f1646d0f7359e33f887a2c83431eb2 WatchSource:0}: Error finding container 2103832d63f95c8337e0896c4056897296f1646d0f7359e33f887a2c83431eb2: Status 404 returned error can't find the container with id 2103832d63f95c8337e0896c4056897296f1646d0f7359e33f887a2c83431eb2 Feb 18 09:17:03 crc kubenswrapper[4556]: I0218 09:17:03.298850 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0537334d-14e5-4ff1-ab08-4c4e7a882c9a" path="/var/lib/kubelet/pods/0537334d-14e5-4ff1-ab08-4c4e7a882c9a/volumes" Feb 18 09:17:03 crc kubenswrapper[4556]: I0218 09:17:03.461413 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64f7f48db9-mjw87" Feb 18 09:17:03 crc kubenswrapper[4556]: I0218 09:17:03.646443 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/30a1b29b-530b-4f64-acb7-a8471fc98fcb-ovsdbserver-nb\") pod \"30a1b29b-530b-4f64-acb7-a8471fc98fcb\" (UID: \"30a1b29b-530b-4f64-acb7-a8471fc98fcb\") " Feb 18 09:17:03 crc kubenswrapper[4556]: I0218 09:17:03.647275 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/30a1b29b-530b-4f64-acb7-a8471fc98fcb-dns-svc\") pod \"30a1b29b-530b-4f64-acb7-a8471fc98fcb\" (UID: \"30a1b29b-530b-4f64-acb7-a8471fc98fcb\") " Feb 18 09:17:03 crc kubenswrapper[4556]: I0218 09:17:03.647403 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30a1b29b-530b-4f64-acb7-a8471fc98fcb-config\") pod \"30a1b29b-530b-4f64-acb7-a8471fc98fcb\" (UID: \"30a1b29b-530b-4f64-acb7-a8471fc98fcb\") " Feb 18 09:17:03 crc kubenswrapper[4556]: I0218 09:17:03.647446 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jntbs\" (UniqueName: \"kubernetes.io/projected/30a1b29b-530b-4f64-acb7-a8471fc98fcb-kube-api-access-jntbs\") pod \"30a1b29b-530b-4f64-acb7-a8471fc98fcb\" (UID: \"30a1b29b-530b-4f64-acb7-a8471fc98fcb\") " Feb 18 09:17:03 crc kubenswrapper[4556]: I0218 09:17:03.652131 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30a1b29b-530b-4f64-acb7-a8471fc98fcb-kube-api-access-jntbs" (OuterVolumeSpecName: "kube-api-access-jntbs") pod "30a1b29b-530b-4f64-acb7-a8471fc98fcb" (UID: "30a1b29b-530b-4f64-acb7-a8471fc98fcb"). InnerVolumeSpecName "kube-api-access-jntbs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:17:03 crc kubenswrapper[4556]: I0218 09:17:03.679175 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30a1b29b-530b-4f64-acb7-a8471fc98fcb-config" (OuterVolumeSpecName: "config") pod "30a1b29b-530b-4f64-acb7-a8471fc98fcb" (UID: "30a1b29b-530b-4f64-acb7-a8471fc98fcb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:17:03 crc kubenswrapper[4556]: I0218 09:17:03.679507 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30a1b29b-530b-4f64-acb7-a8471fc98fcb-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "30a1b29b-530b-4f64-acb7-a8471fc98fcb" (UID: "30a1b29b-530b-4f64-acb7-a8471fc98fcb"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:17:03 crc kubenswrapper[4556]: I0218 09:17:03.681696 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30a1b29b-530b-4f64-acb7-a8471fc98fcb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "30a1b29b-530b-4f64-acb7-a8471fc98fcb" (UID: "30a1b29b-530b-4f64-acb7-a8471fc98fcb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:17:03 crc kubenswrapper[4556]: I0218 09:17:03.749710 4556 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/30a1b29b-530b-4f64-acb7-a8471fc98fcb-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:03 crc kubenswrapper[4556]: I0218 09:17:03.749741 4556 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30a1b29b-530b-4f64-acb7-a8471fc98fcb-config\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:03 crc kubenswrapper[4556]: I0218 09:17:03.749754 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jntbs\" (UniqueName: \"kubernetes.io/projected/30a1b29b-530b-4f64-acb7-a8471fc98fcb-kube-api-access-jntbs\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:03 crc kubenswrapper[4556]: I0218 09:17:03.749766 4556 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/30a1b29b-530b-4f64-acb7-a8471fc98fcb-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:03 crc kubenswrapper[4556]: I0218 09:17:03.944534 4556 generic.go:334] "Generic (PLEG): container finished" podID="7a185e40-8a02-44a1-ab00-2575bebde1d5" containerID="acb3bfeafbc1f1a549401dc430a37eb857f13ba1b502240d540b8a9b1f9225f4" exitCode=0 Feb 18 09:17:03 crc kubenswrapper[4556]: I0218 09:17:03.944628 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-blbjd" event={"ID":"7a185e40-8a02-44a1-ab00-2575bebde1d5","Type":"ContainerDied","Data":"acb3bfeafbc1f1a549401dc430a37eb857f13ba1b502240d540b8a9b1f9225f4"} Feb 18 09:17:03 crc kubenswrapper[4556]: I0218 09:17:03.944663 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-blbjd" event={"ID":"7a185e40-8a02-44a1-ab00-2575bebde1d5","Type":"ContainerStarted","Data":"2103832d63f95c8337e0896c4056897296f1646d0f7359e33f887a2c83431eb2"} Feb 18 09:17:03 crc kubenswrapper[4556]: I0218 09:17:03.948456 4556 generic.go:334] "Generic (PLEG): container finished" podID="30a1b29b-530b-4f64-acb7-a8471fc98fcb" containerID="833d74497abfc08cdfa0f0c32d15a296083d4ae387fbc6542d6c8810ddc63fa8" exitCode=0 Feb 18 09:17:03 crc kubenswrapper[4556]: I0218 09:17:03.948640 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64f7f48db9-mjw87" Feb 18 09:17:03 crc kubenswrapper[4556]: I0218 09:17:03.949432 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64f7f48db9-mjw87" event={"ID":"30a1b29b-530b-4f64-acb7-a8471fc98fcb","Type":"ContainerDied","Data":"833d74497abfc08cdfa0f0c32d15a296083d4ae387fbc6542d6c8810ddc63fa8"} Feb 18 09:17:03 crc kubenswrapper[4556]: I0218 09:17:03.949506 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64f7f48db9-mjw87" event={"ID":"30a1b29b-530b-4f64-acb7-a8471fc98fcb","Type":"ContainerDied","Data":"5bfbcf946c5e97e1f925409a0045b84af89e75a10d25cdad68731e7561fcb4c4"} Feb 18 09:17:03 crc kubenswrapper[4556]: I0218 09:17:03.949535 4556 scope.go:117] "RemoveContainer" containerID="833d74497abfc08cdfa0f0c32d15a296083d4ae387fbc6542d6c8810ddc63fa8" Feb 18 09:17:03 crc kubenswrapper[4556]: I0218 09:17:03.977871 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64f7f48db9-mjw87"] Feb 18 09:17:03 crc kubenswrapper[4556]: I0218 09:17:03.978657 4556 scope.go:117] "RemoveContainer" containerID="0ed084463a3e474d7452081234fd8a927fb6fb2988840005710c317a2d837f74" Feb 18 09:17:03 crc kubenswrapper[4556]: I0218 09:17:03.983912 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-64f7f48db9-mjw87"] Feb 18 09:17:04 crc kubenswrapper[4556]: I0218 09:17:04.004005 4556 scope.go:117] "RemoveContainer" containerID="833d74497abfc08cdfa0f0c32d15a296083d4ae387fbc6542d6c8810ddc63fa8" Feb 18 09:17:04 crc kubenswrapper[4556]: E0218 09:17:04.004405 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"833d74497abfc08cdfa0f0c32d15a296083d4ae387fbc6542d6c8810ddc63fa8\": container with ID starting with 833d74497abfc08cdfa0f0c32d15a296083d4ae387fbc6542d6c8810ddc63fa8 not found: ID does not exist" containerID="833d74497abfc08cdfa0f0c32d15a296083d4ae387fbc6542d6c8810ddc63fa8" Feb 18 09:17:04 crc kubenswrapper[4556]: I0218 09:17:04.004428 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"833d74497abfc08cdfa0f0c32d15a296083d4ae387fbc6542d6c8810ddc63fa8"} err="failed to get container status \"833d74497abfc08cdfa0f0c32d15a296083d4ae387fbc6542d6c8810ddc63fa8\": rpc error: code = NotFound desc = could not find container \"833d74497abfc08cdfa0f0c32d15a296083d4ae387fbc6542d6c8810ddc63fa8\": container with ID starting with 833d74497abfc08cdfa0f0c32d15a296083d4ae387fbc6542d6c8810ddc63fa8 not found: ID does not exist" Feb 18 09:17:04 crc kubenswrapper[4556]: I0218 09:17:04.004447 4556 scope.go:117] "RemoveContainer" containerID="0ed084463a3e474d7452081234fd8a927fb6fb2988840005710c317a2d837f74" Feb 18 09:17:04 crc kubenswrapper[4556]: E0218 09:17:04.004701 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ed084463a3e474d7452081234fd8a927fb6fb2988840005710c317a2d837f74\": container with ID starting with 0ed084463a3e474d7452081234fd8a927fb6fb2988840005710c317a2d837f74 not found: ID does not exist" containerID="0ed084463a3e474d7452081234fd8a927fb6fb2988840005710c317a2d837f74" Feb 18 09:17:04 crc kubenswrapper[4556]: I0218 09:17:04.004718 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ed084463a3e474d7452081234fd8a927fb6fb2988840005710c317a2d837f74"} err="failed to get container status \"0ed084463a3e474d7452081234fd8a927fb6fb2988840005710c317a2d837f74\": rpc error: code = NotFound desc = could not find container \"0ed084463a3e474d7452081234fd8a927fb6fb2988840005710c317a2d837f74\": container with ID starting with 0ed084463a3e474d7452081234fd8a927fb6fb2988840005710c317a2d837f74 not found: ID does not exist" Feb 18 09:17:04 crc kubenswrapper[4556]: I0218 09:17:04.191581 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-9z5cc" Feb 18 09:17:04 crc kubenswrapper[4556]: I0218 09:17:04.357617 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/03ccffdf-f11c-4788-9e22-0b0661665c5d-ring-data-devices\") pod \"03ccffdf-f11c-4788-9e22-0b0661665c5d\" (UID: \"03ccffdf-f11c-4788-9e22-0b0661665c5d\") " Feb 18 09:17:04 crc kubenswrapper[4556]: I0218 09:17:04.357709 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/03ccffdf-f11c-4788-9e22-0b0661665c5d-dispersionconf\") pod \"03ccffdf-f11c-4788-9e22-0b0661665c5d\" (UID: \"03ccffdf-f11c-4788-9e22-0b0661665c5d\") " Feb 18 09:17:04 crc kubenswrapper[4556]: I0218 09:17:04.357776 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zhwx8\" (UniqueName: \"kubernetes.io/projected/03ccffdf-f11c-4788-9e22-0b0661665c5d-kube-api-access-zhwx8\") pod \"03ccffdf-f11c-4788-9e22-0b0661665c5d\" (UID: \"03ccffdf-f11c-4788-9e22-0b0661665c5d\") " Feb 18 09:17:04 crc kubenswrapper[4556]: I0218 09:17:04.357904 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03ccffdf-f11c-4788-9e22-0b0661665c5d-combined-ca-bundle\") pod \"03ccffdf-f11c-4788-9e22-0b0661665c5d\" (UID: \"03ccffdf-f11c-4788-9e22-0b0661665c5d\") " Feb 18 09:17:04 crc kubenswrapper[4556]: I0218 09:17:04.357943 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/03ccffdf-f11c-4788-9e22-0b0661665c5d-swiftconf\") pod \"03ccffdf-f11c-4788-9e22-0b0661665c5d\" (UID: \"03ccffdf-f11c-4788-9e22-0b0661665c5d\") " Feb 18 09:17:04 crc kubenswrapper[4556]: I0218 09:17:04.357995 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/03ccffdf-f11c-4788-9e22-0b0661665c5d-etc-swift\") pod \"03ccffdf-f11c-4788-9e22-0b0661665c5d\" (UID: \"03ccffdf-f11c-4788-9e22-0b0661665c5d\") " Feb 18 09:17:04 crc kubenswrapper[4556]: I0218 09:17:04.360074 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03ccffdf-f11c-4788-9e22-0b0661665c5d-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "03ccffdf-f11c-4788-9e22-0b0661665c5d" (UID: "03ccffdf-f11c-4788-9e22-0b0661665c5d"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:17:04 crc kubenswrapper[4556]: I0218 09:17:04.364848 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03ccffdf-f11c-4788-9e22-0b0661665c5d-kube-api-access-zhwx8" (OuterVolumeSpecName: "kube-api-access-zhwx8") pod "03ccffdf-f11c-4788-9e22-0b0661665c5d" (UID: "03ccffdf-f11c-4788-9e22-0b0661665c5d"). InnerVolumeSpecName "kube-api-access-zhwx8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:17:04 crc kubenswrapper[4556]: I0218 09:17:04.367300 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03ccffdf-f11c-4788-9e22-0b0661665c5d-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "03ccffdf-f11c-4788-9e22-0b0661665c5d" (UID: "03ccffdf-f11c-4788-9e22-0b0661665c5d"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:17:04 crc kubenswrapper[4556]: I0218 09:17:04.367448 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/03ccffdf-f11c-4788-9e22-0b0661665c5d-scripts\") pod \"03ccffdf-f11c-4788-9e22-0b0661665c5d\" (UID: \"03ccffdf-f11c-4788-9e22-0b0661665c5d\") " Feb 18 09:17:04 crc kubenswrapper[4556]: I0218 09:17:04.368932 4556 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/03ccffdf-f11c-4788-9e22-0b0661665c5d-ring-data-devices\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:04 crc kubenswrapper[4556]: I0218 09:17:04.368968 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zhwx8\" (UniqueName: \"kubernetes.io/projected/03ccffdf-f11c-4788-9e22-0b0661665c5d-kube-api-access-zhwx8\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:04 crc kubenswrapper[4556]: I0218 09:17:04.368984 4556 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/03ccffdf-f11c-4788-9e22-0b0661665c5d-etc-swift\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:04 crc kubenswrapper[4556]: I0218 09:17:04.371194 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03ccffdf-f11c-4788-9e22-0b0661665c5d-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "03ccffdf-f11c-4788-9e22-0b0661665c5d" (UID: "03ccffdf-f11c-4788-9e22-0b0661665c5d"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:17:04 crc kubenswrapper[4556]: I0218 09:17:04.383776 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03ccffdf-f11c-4788-9e22-0b0661665c5d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "03ccffdf-f11c-4788-9e22-0b0661665c5d" (UID: "03ccffdf-f11c-4788-9e22-0b0661665c5d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:17:04 crc kubenswrapper[4556]: I0218 09:17:04.389928 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03ccffdf-f11c-4788-9e22-0b0661665c5d-scripts" (OuterVolumeSpecName: "scripts") pod "03ccffdf-f11c-4788-9e22-0b0661665c5d" (UID: "03ccffdf-f11c-4788-9e22-0b0661665c5d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:17:04 crc kubenswrapper[4556]: I0218 09:17:04.390246 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03ccffdf-f11c-4788-9e22-0b0661665c5d-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "03ccffdf-f11c-4788-9e22-0b0661665c5d" (UID: "03ccffdf-f11c-4788-9e22-0b0661665c5d"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:17:04 crc kubenswrapper[4556]: I0218 09:17:04.471962 4556 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/03ccffdf-f11c-4788-9e22-0b0661665c5d-scripts\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:04 crc kubenswrapper[4556]: I0218 09:17:04.472006 4556 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/03ccffdf-f11c-4788-9e22-0b0661665c5d-dispersionconf\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:04 crc kubenswrapper[4556]: I0218 09:17:04.472028 4556 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03ccffdf-f11c-4788-9e22-0b0661665c5d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:04 crc kubenswrapper[4556]: I0218 09:17:04.472041 4556 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/03ccffdf-f11c-4788-9e22-0b0661665c5d-swiftconf\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:04 crc kubenswrapper[4556]: I0218 09:17:04.573932 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/da163339-653b-4057-9c2f-332eb6957f40-etc-swift\") pod \"swift-storage-0\" (UID: \"da163339-653b-4057-9c2f-332eb6957f40\") " pod="openstack/swift-storage-0" Feb 18 09:17:04 crc kubenswrapper[4556]: I0218 09:17:04.578273 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/da163339-653b-4057-9c2f-332eb6957f40-etc-swift\") pod \"swift-storage-0\" (UID: \"da163339-653b-4057-9c2f-332eb6957f40\") " pod="openstack/swift-storage-0" Feb 18 09:17:04 crc kubenswrapper[4556]: I0218 09:17:04.652062 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Feb 18 09:17:04 crc kubenswrapper[4556]: I0218 09:17:04.957093 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-9z5cc" event={"ID":"03ccffdf-f11c-4788-9e22-0b0661665c5d","Type":"ContainerDied","Data":"0233d450ed2b7409752b7b01e9727cb23c5995dfa206a754bf8045f48287dabd"} Feb 18 09:17:04 crc kubenswrapper[4556]: I0218 09:17:04.957411 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0233d450ed2b7409752b7b01e9727cb23c5995dfa206a754bf8045f48287dabd" Feb 18 09:17:04 crc kubenswrapper[4556]: I0218 09:17:04.957261 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-9z5cc" Feb 18 09:17:05 crc kubenswrapper[4556]: I0218 09:17:05.137826 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Feb 18 09:17:05 crc kubenswrapper[4556]: I0218 09:17:05.195165 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-blbjd" Feb 18 09:17:05 crc kubenswrapper[4556]: I0218 09:17:05.283693 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7a185e40-8a02-44a1-ab00-2575bebde1d5-operator-scripts\") pod \"7a185e40-8a02-44a1-ab00-2575bebde1d5\" (UID: \"7a185e40-8a02-44a1-ab00-2575bebde1d5\") " Feb 18 09:17:05 crc kubenswrapper[4556]: I0218 09:17:05.283748 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dszlk\" (UniqueName: \"kubernetes.io/projected/7a185e40-8a02-44a1-ab00-2575bebde1d5-kube-api-access-dszlk\") pod \"7a185e40-8a02-44a1-ab00-2575bebde1d5\" (UID: \"7a185e40-8a02-44a1-ab00-2575bebde1d5\") " Feb 18 09:17:05 crc kubenswrapper[4556]: I0218 09:17:05.284355 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a185e40-8a02-44a1-ab00-2575bebde1d5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7a185e40-8a02-44a1-ab00-2575bebde1d5" (UID: "7a185e40-8a02-44a1-ab00-2575bebde1d5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:17:05 crc kubenswrapper[4556]: I0218 09:17:05.291375 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a185e40-8a02-44a1-ab00-2575bebde1d5-kube-api-access-dszlk" (OuterVolumeSpecName: "kube-api-access-dszlk") pod "7a185e40-8a02-44a1-ab00-2575bebde1d5" (UID: "7a185e40-8a02-44a1-ab00-2575bebde1d5"). InnerVolumeSpecName "kube-api-access-dszlk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:17:05 crc kubenswrapper[4556]: I0218 09:17:05.292968 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30a1b29b-530b-4f64-acb7-a8471fc98fcb" path="/var/lib/kubelet/pods/30a1b29b-530b-4f64-acb7-a8471fc98fcb/volumes" Feb 18 09:17:05 crc kubenswrapper[4556]: I0218 09:17:05.386424 4556 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7a185e40-8a02-44a1-ab00-2575bebde1d5-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:05 crc kubenswrapper[4556]: I0218 09:17:05.386455 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dszlk\" (UniqueName: \"kubernetes.io/projected/7a185e40-8a02-44a1-ab00-2575bebde1d5-kube-api-access-dszlk\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:05 crc kubenswrapper[4556]: I0218 09:17:05.620911 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-fxcc2"] Feb 18 09:17:05 crc kubenswrapper[4556]: E0218 09:17:05.621226 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30a1b29b-530b-4f64-acb7-a8471fc98fcb" containerName="init" Feb 18 09:17:05 crc kubenswrapper[4556]: I0218 09:17:05.621244 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="30a1b29b-530b-4f64-acb7-a8471fc98fcb" containerName="init" Feb 18 09:17:05 crc kubenswrapper[4556]: E0218 09:17:05.621266 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30a1b29b-530b-4f64-acb7-a8471fc98fcb" containerName="dnsmasq-dns" Feb 18 09:17:05 crc kubenswrapper[4556]: I0218 09:17:05.621274 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="30a1b29b-530b-4f64-acb7-a8471fc98fcb" containerName="dnsmasq-dns" Feb 18 09:17:05 crc kubenswrapper[4556]: E0218 09:17:05.621296 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a185e40-8a02-44a1-ab00-2575bebde1d5" containerName="mariadb-account-create-update" Feb 18 09:17:05 crc kubenswrapper[4556]: I0218 09:17:05.621303 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a185e40-8a02-44a1-ab00-2575bebde1d5" containerName="mariadb-account-create-update" Feb 18 09:17:05 crc kubenswrapper[4556]: E0218 09:17:05.621321 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03ccffdf-f11c-4788-9e22-0b0661665c5d" containerName="swift-ring-rebalance" Feb 18 09:17:05 crc kubenswrapper[4556]: I0218 09:17:05.621327 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="03ccffdf-f11c-4788-9e22-0b0661665c5d" containerName="swift-ring-rebalance" Feb 18 09:17:05 crc kubenswrapper[4556]: I0218 09:17:05.621475 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="03ccffdf-f11c-4788-9e22-0b0661665c5d" containerName="swift-ring-rebalance" Feb 18 09:17:05 crc kubenswrapper[4556]: I0218 09:17:05.621496 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a185e40-8a02-44a1-ab00-2575bebde1d5" containerName="mariadb-account-create-update" Feb 18 09:17:05 crc kubenswrapper[4556]: I0218 09:17:05.621505 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="30a1b29b-530b-4f64-acb7-a8471fc98fcb" containerName="dnsmasq-dns" Feb 18 09:17:05 crc kubenswrapper[4556]: I0218 09:17:05.621980 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-fxcc2" Feb 18 09:17:05 crc kubenswrapper[4556]: I0218 09:17:05.623392 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-7flzx" Feb 18 09:17:05 crc kubenswrapper[4556]: I0218 09:17:05.629054 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Feb 18 09:17:05 crc kubenswrapper[4556]: I0218 09:17:05.631524 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-fxcc2"] Feb 18 09:17:05 crc kubenswrapper[4556]: I0218 09:17:05.690414 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4731e0af-7aff-41c9-bc33-ad533ee750ce-combined-ca-bundle\") pod \"glance-db-sync-fxcc2\" (UID: \"4731e0af-7aff-41c9-bc33-ad533ee750ce\") " pod="openstack/glance-db-sync-fxcc2" Feb 18 09:17:05 crc kubenswrapper[4556]: I0218 09:17:05.690517 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4731e0af-7aff-41c9-bc33-ad533ee750ce-config-data\") pod \"glance-db-sync-fxcc2\" (UID: \"4731e0af-7aff-41c9-bc33-ad533ee750ce\") " pod="openstack/glance-db-sync-fxcc2" Feb 18 09:17:05 crc kubenswrapper[4556]: I0218 09:17:05.690600 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4731e0af-7aff-41c9-bc33-ad533ee750ce-db-sync-config-data\") pod \"glance-db-sync-fxcc2\" (UID: \"4731e0af-7aff-41c9-bc33-ad533ee750ce\") " pod="openstack/glance-db-sync-fxcc2" Feb 18 09:17:05 crc kubenswrapper[4556]: I0218 09:17:05.690658 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fj5m8\" (UniqueName: \"kubernetes.io/projected/4731e0af-7aff-41c9-bc33-ad533ee750ce-kube-api-access-fj5m8\") pod \"glance-db-sync-fxcc2\" (UID: \"4731e0af-7aff-41c9-bc33-ad533ee750ce\") " pod="openstack/glance-db-sync-fxcc2" Feb 18 09:17:05 crc kubenswrapper[4556]: I0218 09:17:05.791593 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4731e0af-7aff-41c9-bc33-ad533ee750ce-combined-ca-bundle\") pod \"glance-db-sync-fxcc2\" (UID: \"4731e0af-7aff-41c9-bc33-ad533ee750ce\") " pod="openstack/glance-db-sync-fxcc2" Feb 18 09:17:05 crc kubenswrapper[4556]: I0218 09:17:05.791743 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4731e0af-7aff-41c9-bc33-ad533ee750ce-config-data\") pod \"glance-db-sync-fxcc2\" (UID: \"4731e0af-7aff-41c9-bc33-ad533ee750ce\") " pod="openstack/glance-db-sync-fxcc2" Feb 18 09:17:05 crc kubenswrapper[4556]: I0218 09:17:05.792466 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4731e0af-7aff-41c9-bc33-ad533ee750ce-db-sync-config-data\") pod \"glance-db-sync-fxcc2\" (UID: \"4731e0af-7aff-41c9-bc33-ad533ee750ce\") " pod="openstack/glance-db-sync-fxcc2" Feb 18 09:17:05 crc kubenswrapper[4556]: I0218 09:17:05.792609 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fj5m8\" (UniqueName: \"kubernetes.io/projected/4731e0af-7aff-41c9-bc33-ad533ee750ce-kube-api-access-fj5m8\") pod \"glance-db-sync-fxcc2\" (UID: \"4731e0af-7aff-41c9-bc33-ad533ee750ce\") " pod="openstack/glance-db-sync-fxcc2" Feb 18 09:17:05 crc kubenswrapper[4556]: I0218 09:17:05.796854 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4731e0af-7aff-41c9-bc33-ad533ee750ce-combined-ca-bundle\") pod \"glance-db-sync-fxcc2\" (UID: \"4731e0af-7aff-41c9-bc33-ad533ee750ce\") " pod="openstack/glance-db-sync-fxcc2" Feb 18 09:17:05 crc kubenswrapper[4556]: I0218 09:17:05.796878 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4731e0af-7aff-41c9-bc33-ad533ee750ce-db-sync-config-data\") pod \"glance-db-sync-fxcc2\" (UID: \"4731e0af-7aff-41c9-bc33-ad533ee750ce\") " pod="openstack/glance-db-sync-fxcc2" Feb 18 09:17:05 crc kubenswrapper[4556]: I0218 09:17:05.796992 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4731e0af-7aff-41c9-bc33-ad533ee750ce-config-data\") pod \"glance-db-sync-fxcc2\" (UID: \"4731e0af-7aff-41c9-bc33-ad533ee750ce\") " pod="openstack/glance-db-sync-fxcc2" Feb 18 09:17:05 crc kubenswrapper[4556]: I0218 09:17:05.806794 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fj5m8\" (UniqueName: \"kubernetes.io/projected/4731e0af-7aff-41c9-bc33-ad533ee750ce-kube-api-access-fj5m8\") pod \"glance-db-sync-fxcc2\" (UID: \"4731e0af-7aff-41c9-bc33-ad533ee750ce\") " pod="openstack/glance-db-sync-fxcc2" Feb 18 09:17:05 crc kubenswrapper[4556]: I0218 09:17:05.957580 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-fxcc2" Feb 18 09:17:05 crc kubenswrapper[4556]: I0218 09:17:05.967335 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"da163339-653b-4057-9c2f-332eb6957f40","Type":"ContainerStarted","Data":"d96bb63d2c0f3b81f88e28ecaa89e6023a41682c014ebd024e09348851a61617"} Feb 18 09:17:05 crc kubenswrapper[4556]: I0218 09:17:05.969356 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-blbjd" event={"ID":"7a185e40-8a02-44a1-ab00-2575bebde1d5","Type":"ContainerDied","Data":"2103832d63f95c8337e0896c4056897296f1646d0f7359e33f887a2c83431eb2"} Feb 18 09:17:05 crc kubenswrapper[4556]: I0218 09:17:05.969390 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2103832d63f95c8337e0896c4056897296f1646d0f7359e33f887a2c83431eb2" Feb 18 09:17:05 crc kubenswrapper[4556]: I0218 09:17:05.969410 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-blbjd" Feb 18 09:17:06 crc kubenswrapper[4556]: I0218 09:17:06.384881 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-fxcc2"] Feb 18 09:17:06 crc kubenswrapper[4556]: W0218 09:17:06.549812 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4731e0af_7aff_41c9_bc33_ad533ee750ce.slice/crio-906ff112dbba860d29503d7bae98de2325f4607da2efd57c1e572efb86861079 WatchSource:0}: Error finding container 906ff112dbba860d29503d7bae98de2325f4607da2efd57c1e572efb86861079: Status 404 returned error can't find the container with id 906ff112dbba860d29503d7bae98de2325f4607da2efd57c1e572efb86861079 Feb 18 09:17:06 crc kubenswrapper[4556]: I0218 09:17:06.981577 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-fxcc2" event={"ID":"4731e0af-7aff-41c9-bc33-ad533ee750ce","Type":"ContainerStarted","Data":"906ff112dbba860d29503d7bae98de2325f4607da2efd57c1e572efb86861079"} Feb 18 09:17:06 crc kubenswrapper[4556]: I0218 09:17:06.985393 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"da163339-653b-4057-9c2f-332eb6957f40","Type":"ContainerStarted","Data":"23f5c04a9b22c56cf56fdc1ffde93f44c502901076a46c402748e6fc52ae98ae"} Feb 18 09:17:06 crc kubenswrapper[4556]: I0218 09:17:06.985452 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"da163339-653b-4057-9c2f-332eb6957f40","Type":"ContainerStarted","Data":"b8dec1d314adb0801458b05a68858100c77a0765821295069df4b92f27e20739"} Feb 18 09:17:06 crc kubenswrapper[4556]: I0218 09:17:06.985467 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"da163339-653b-4057-9c2f-332eb6957f40","Type":"ContainerStarted","Data":"fd498498313a4a413fcf564d371f6c157f2426f0e67308a8b5ce111d39b967d2"} Feb 18 09:17:07 crc kubenswrapper[4556]: I0218 09:17:07.996836 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"da163339-653b-4057-9c2f-332eb6957f40","Type":"ContainerStarted","Data":"36dbb54d81da2a65a5099094ff7f83203d8a459f695f93fb571364a546a69683"} Feb 18 09:17:09 crc kubenswrapper[4556]: I0218 09:17:09.006568 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"da163339-653b-4057-9c2f-332eb6957f40","Type":"ContainerStarted","Data":"68f3dc642f9e3080940bf1f79142592fbf02d03c422235d4483d9a64aa113d02"} Feb 18 09:17:09 crc kubenswrapper[4556]: I0218 09:17:09.007096 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"da163339-653b-4057-9c2f-332eb6957f40","Type":"ContainerStarted","Data":"919c360f97b97b37d09ddb15f642573ebcbefcfffe4eb69d08af12cfddca5b08"} Feb 18 09:17:10 crc kubenswrapper[4556]: I0218 09:17:10.019921 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"da163339-653b-4057-9c2f-332eb6957f40","Type":"ContainerStarted","Data":"329d34b962179379db0104bf5f7c38c750dc4ae14d008326891b96ceda22af67"} Feb 18 09:17:10 crc kubenswrapper[4556]: I0218 09:17:10.019968 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"da163339-653b-4057-9c2f-332eb6957f40","Type":"ContainerStarted","Data":"dfd7b29a6e230ad22cf008f0d0406714468f795ed6c38182537da074a823da58"} Feb 18 09:17:10 crc kubenswrapper[4556]: I0218 09:17:10.656528 4556 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-6kkrw" podUID="4906afef-c7e6-4597-a4a5-e9d758917e11" containerName="ovn-controller" probeResult="failure" output=< Feb 18 09:17:10 crc kubenswrapper[4556]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Feb 18 09:17:10 crc kubenswrapper[4556]: > Feb 18 09:17:10 crc kubenswrapper[4556]: I0218 09:17:10.705466 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Feb 18 09:17:12 crc kubenswrapper[4556]: I0218 09:17:12.036829 4556 generic.go:334] "Generic (PLEG): container finished" podID="03023501-772f-4798-a953-f56dee586cb1" containerID="7bf62d08939139f2182730d0a4d9b37f729f09b7b5649c1d303400e1b02489b7" exitCode=0 Feb 18 09:17:12 crc kubenswrapper[4556]: I0218 09:17:12.036892 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"03023501-772f-4798-a953-f56dee586cb1","Type":"ContainerDied","Data":"7bf62d08939139f2182730d0a4d9b37f729f09b7b5649c1d303400e1b02489b7"} Feb 18 09:17:13 crc kubenswrapper[4556]: I0218 09:17:13.045739 4556 generic.go:334] "Generic (PLEG): container finished" podID="1d7961e1-6f96-4707-907d-1fa0c30641e2" containerID="98251fd58b2a29ac981c57fb4437b319daf966b2b07731b525037e5dccf051cf" exitCode=0 Feb 18 09:17:13 crc kubenswrapper[4556]: I0218 09:17:13.045784 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1d7961e1-6f96-4707-907d-1fa0c30641e2","Type":"ContainerDied","Data":"98251fd58b2a29ac981c57fb4437b319daf966b2b07731b525037e5dccf051cf"} Feb 18 09:17:15 crc kubenswrapper[4556]: I0218 09:17:15.651223 4556 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-6kkrw" podUID="4906afef-c7e6-4597-a4a5-e9d758917e11" containerName="ovn-controller" probeResult="failure" output=< Feb 18 09:17:15 crc kubenswrapper[4556]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Feb 18 09:17:15 crc kubenswrapper[4556]: > Feb 18 09:17:15 crc kubenswrapper[4556]: I0218 09:17:15.686793 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-srnvn" Feb 18 09:17:15 crc kubenswrapper[4556]: I0218 09:17:15.687482 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-srnvn" Feb 18 09:17:15 crc kubenswrapper[4556]: I0218 09:17:15.915658 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-6kkrw-config-q9lcv"] Feb 18 09:17:15 crc kubenswrapper[4556]: I0218 09:17:15.916989 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-6kkrw-config-q9lcv" Feb 18 09:17:15 crc kubenswrapper[4556]: I0218 09:17:15.921314 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Feb 18 09:17:15 crc kubenswrapper[4556]: I0218 09:17:15.922055 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-6kkrw-config-q9lcv"] Feb 18 09:17:15 crc kubenswrapper[4556]: I0218 09:17:15.990981 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/75a278f7-6511-40c1-ab7e-82644755b10b-additional-scripts\") pod \"ovn-controller-6kkrw-config-q9lcv\" (UID: \"75a278f7-6511-40c1-ab7e-82644755b10b\") " pod="openstack/ovn-controller-6kkrw-config-q9lcv" Feb 18 09:17:15 crc kubenswrapper[4556]: I0218 09:17:15.991104 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/75a278f7-6511-40c1-ab7e-82644755b10b-var-run-ovn\") pod \"ovn-controller-6kkrw-config-q9lcv\" (UID: \"75a278f7-6511-40c1-ab7e-82644755b10b\") " pod="openstack/ovn-controller-6kkrw-config-q9lcv" Feb 18 09:17:15 crc kubenswrapper[4556]: I0218 09:17:15.991127 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/75a278f7-6511-40c1-ab7e-82644755b10b-var-run\") pod \"ovn-controller-6kkrw-config-q9lcv\" (UID: \"75a278f7-6511-40c1-ab7e-82644755b10b\") " pod="openstack/ovn-controller-6kkrw-config-q9lcv" Feb 18 09:17:15 crc kubenswrapper[4556]: I0218 09:17:15.991182 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/75a278f7-6511-40c1-ab7e-82644755b10b-var-log-ovn\") pod \"ovn-controller-6kkrw-config-q9lcv\" (UID: \"75a278f7-6511-40c1-ab7e-82644755b10b\") " pod="openstack/ovn-controller-6kkrw-config-q9lcv" Feb 18 09:17:15 crc kubenswrapper[4556]: I0218 09:17:15.991260 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/75a278f7-6511-40c1-ab7e-82644755b10b-scripts\") pod \"ovn-controller-6kkrw-config-q9lcv\" (UID: \"75a278f7-6511-40c1-ab7e-82644755b10b\") " pod="openstack/ovn-controller-6kkrw-config-q9lcv" Feb 18 09:17:15 crc kubenswrapper[4556]: I0218 09:17:15.991296 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrg2b\" (UniqueName: \"kubernetes.io/projected/75a278f7-6511-40c1-ab7e-82644755b10b-kube-api-access-lrg2b\") pod \"ovn-controller-6kkrw-config-q9lcv\" (UID: \"75a278f7-6511-40c1-ab7e-82644755b10b\") " pod="openstack/ovn-controller-6kkrw-config-q9lcv" Feb 18 09:17:16 crc kubenswrapper[4556]: I0218 09:17:16.069458 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"03023501-772f-4798-a953-f56dee586cb1","Type":"ContainerStarted","Data":"b6d70d7c016d0deed03db8dc523b6b6d3d816ebdd2eb672bcd3ebfca6f608a64"} Feb 18 09:17:16 crc kubenswrapper[4556]: I0218 09:17:16.069690 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:17:16 crc kubenswrapper[4556]: I0218 09:17:16.072853 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1d7961e1-6f96-4707-907d-1fa0c30641e2","Type":"ContainerStarted","Data":"845e8aa21dc1c9ff225cb48767ac3a68bdf567f6d2263ac90aef2e165de13da8"} Feb 18 09:17:16 crc kubenswrapper[4556]: I0218 09:17:16.073064 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Feb 18 09:17:16 crc kubenswrapper[4556]: I0218 09:17:16.092396 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrg2b\" (UniqueName: \"kubernetes.io/projected/75a278f7-6511-40c1-ab7e-82644755b10b-kube-api-access-lrg2b\") pod \"ovn-controller-6kkrw-config-q9lcv\" (UID: \"75a278f7-6511-40c1-ab7e-82644755b10b\") " pod="openstack/ovn-controller-6kkrw-config-q9lcv" Feb 18 09:17:16 crc kubenswrapper[4556]: I0218 09:17:16.092489 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/75a278f7-6511-40c1-ab7e-82644755b10b-additional-scripts\") pod \"ovn-controller-6kkrw-config-q9lcv\" (UID: \"75a278f7-6511-40c1-ab7e-82644755b10b\") " pod="openstack/ovn-controller-6kkrw-config-q9lcv" Feb 18 09:17:16 crc kubenswrapper[4556]: I0218 09:17:16.092597 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/75a278f7-6511-40c1-ab7e-82644755b10b-var-run-ovn\") pod \"ovn-controller-6kkrw-config-q9lcv\" (UID: \"75a278f7-6511-40c1-ab7e-82644755b10b\") " pod="openstack/ovn-controller-6kkrw-config-q9lcv" Feb 18 09:17:16 crc kubenswrapper[4556]: I0218 09:17:16.092619 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/75a278f7-6511-40c1-ab7e-82644755b10b-var-run\") pod \"ovn-controller-6kkrw-config-q9lcv\" (UID: \"75a278f7-6511-40c1-ab7e-82644755b10b\") " pod="openstack/ovn-controller-6kkrw-config-q9lcv" Feb 18 09:17:16 crc kubenswrapper[4556]: I0218 09:17:16.092662 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/75a278f7-6511-40c1-ab7e-82644755b10b-var-log-ovn\") pod \"ovn-controller-6kkrw-config-q9lcv\" (UID: \"75a278f7-6511-40c1-ab7e-82644755b10b\") " pod="openstack/ovn-controller-6kkrw-config-q9lcv" Feb 18 09:17:16 crc kubenswrapper[4556]: I0218 09:17:16.092763 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/75a278f7-6511-40c1-ab7e-82644755b10b-scripts\") pod \"ovn-controller-6kkrw-config-q9lcv\" (UID: \"75a278f7-6511-40c1-ab7e-82644755b10b\") " pod="openstack/ovn-controller-6kkrw-config-q9lcv" Feb 18 09:17:16 crc kubenswrapper[4556]: I0218 09:17:16.093322 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/75a278f7-6511-40c1-ab7e-82644755b10b-var-run\") pod \"ovn-controller-6kkrw-config-q9lcv\" (UID: \"75a278f7-6511-40c1-ab7e-82644755b10b\") " pod="openstack/ovn-controller-6kkrw-config-q9lcv" Feb 18 09:17:16 crc kubenswrapper[4556]: I0218 09:17:16.093409 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/75a278f7-6511-40c1-ab7e-82644755b10b-var-run-ovn\") pod \"ovn-controller-6kkrw-config-q9lcv\" (UID: \"75a278f7-6511-40c1-ab7e-82644755b10b\") " pod="openstack/ovn-controller-6kkrw-config-q9lcv" Feb 18 09:17:16 crc kubenswrapper[4556]: I0218 09:17:16.093486 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/75a278f7-6511-40c1-ab7e-82644755b10b-var-log-ovn\") pod \"ovn-controller-6kkrw-config-q9lcv\" (UID: \"75a278f7-6511-40c1-ab7e-82644755b10b\") " pod="openstack/ovn-controller-6kkrw-config-q9lcv" Feb 18 09:17:16 crc kubenswrapper[4556]: I0218 09:17:16.093635 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/75a278f7-6511-40c1-ab7e-82644755b10b-additional-scripts\") pod \"ovn-controller-6kkrw-config-q9lcv\" (UID: \"75a278f7-6511-40c1-ab7e-82644755b10b\") " pod="openstack/ovn-controller-6kkrw-config-q9lcv" Feb 18 09:17:16 crc kubenswrapper[4556]: I0218 09:17:16.096873 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/75a278f7-6511-40c1-ab7e-82644755b10b-scripts\") pod \"ovn-controller-6kkrw-config-q9lcv\" (UID: \"75a278f7-6511-40c1-ab7e-82644755b10b\") " pod="openstack/ovn-controller-6kkrw-config-q9lcv" Feb 18 09:17:16 crc kubenswrapper[4556]: I0218 09:17:16.102303 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=49.029623155 podStartE2EDuration="55.102292787s" podCreationTimestamp="2026-02-18 09:16:21 +0000 UTC" firstStartedPulling="2026-02-18 09:16:31.343475735 +0000 UTC m=+748.360436715" lastFinishedPulling="2026-02-18 09:16:37.416145367 +0000 UTC m=+754.433106347" observedRunningTime="2026-02-18 09:17:16.09962427 +0000 UTC m=+793.116585251" watchObservedRunningTime="2026-02-18 09:17:16.102292787 +0000 UTC m=+793.119253767" Feb 18 09:17:16 crc kubenswrapper[4556]: I0218 09:17:16.103640 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"da163339-653b-4057-9c2f-332eb6957f40","Type":"ContainerStarted","Data":"17e6d89c89918ea911cce49576aa280654138363f4d12687f780f371506f5ba2"} Feb 18 09:17:16 crc kubenswrapper[4556]: I0218 09:17:16.103681 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"da163339-653b-4057-9c2f-332eb6957f40","Type":"ContainerStarted","Data":"ddae6ac7f1fc2714405dcb55d790ed5ef711c27dde0dc1e4b23c66bb7bbe8d68"} Feb 18 09:17:16 crc kubenswrapper[4556]: I0218 09:17:16.114650 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrg2b\" (UniqueName: \"kubernetes.io/projected/75a278f7-6511-40c1-ab7e-82644755b10b-kube-api-access-lrg2b\") pod \"ovn-controller-6kkrw-config-q9lcv\" (UID: \"75a278f7-6511-40c1-ab7e-82644755b10b\") " pod="openstack/ovn-controller-6kkrw-config-q9lcv" Feb 18 09:17:16 crc kubenswrapper[4556]: I0218 09:17:16.128121 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=50.240758672 podStartE2EDuration="56.128111307s" podCreationTimestamp="2026-02-18 09:16:20 +0000 UTC" firstStartedPulling="2026-02-18 09:16:31.529626347 +0000 UTC m=+748.546587327" lastFinishedPulling="2026-02-18 09:16:37.416978981 +0000 UTC m=+754.433939962" observedRunningTime="2026-02-18 09:17:16.123987253 +0000 UTC m=+793.140948233" watchObservedRunningTime="2026-02-18 09:17:16.128111307 +0000 UTC m=+793.145072286" Feb 18 09:17:16 crc kubenswrapper[4556]: I0218 09:17:16.262928 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-6kkrw-config-q9lcv" Feb 18 09:17:16 crc kubenswrapper[4556]: I0218 09:17:16.816668 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-6kkrw-config-q9lcv"] Feb 18 09:17:16 crc kubenswrapper[4556]: W0218 09:17:16.823187 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod75a278f7_6511_40c1_ab7e_82644755b10b.slice/crio-258273d4269ec5fcd95e058724c2c2d29a9d6ea5218988e0f3e5862c01dcecd5 WatchSource:0}: Error finding container 258273d4269ec5fcd95e058724c2c2d29a9d6ea5218988e0f3e5862c01dcecd5: Status 404 returned error can't find the container with id 258273d4269ec5fcd95e058724c2c2d29a9d6ea5218988e0f3e5862c01dcecd5 Feb 18 09:17:17 crc kubenswrapper[4556]: I0218 09:17:17.114200 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-fxcc2" event={"ID":"4731e0af-7aff-41c9-bc33-ad533ee750ce","Type":"ContainerStarted","Data":"c429fb238171b24a587bcafa3d8a82c21f54094c5652e03e587e69b41a0106c0"} Feb 18 09:17:17 crc kubenswrapper[4556]: I0218 09:17:17.116918 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-6kkrw-config-q9lcv" event={"ID":"75a278f7-6511-40c1-ab7e-82644755b10b","Type":"ContainerStarted","Data":"ca0d831b26e0653479e1280cd1625ab6207b9f101f1bbc574d2caad852610ea1"} Feb 18 09:17:17 crc kubenswrapper[4556]: I0218 09:17:17.116958 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-6kkrw-config-q9lcv" event={"ID":"75a278f7-6511-40c1-ab7e-82644755b10b","Type":"ContainerStarted","Data":"258273d4269ec5fcd95e058724c2c2d29a9d6ea5218988e0f3e5862c01dcecd5"} Feb 18 09:17:17 crc kubenswrapper[4556]: I0218 09:17:17.124889 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"da163339-653b-4057-9c2f-332eb6957f40","Type":"ContainerStarted","Data":"8582ceb1e61cd65e131d93d5ffe9adf9624268e212b587eeef73a5dd6d98d9cd"} Feb 18 09:17:17 crc kubenswrapper[4556]: I0218 09:17:17.124922 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"da163339-653b-4057-9c2f-332eb6957f40","Type":"ContainerStarted","Data":"ad408942071a376a37b6a8b7ac7ae8f8566022a0e65d00ea3eacb2a6897a7188"} Feb 18 09:17:17 crc kubenswrapper[4556]: I0218 09:17:17.124933 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"da163339-653b-4057-9c2f-332eb6957f40","Type":"ContainerStarted","Data":"f6009d6853d0e0123d4b26df289d323fa05cd3542f99a511dd01d087f7b4e7f0"} Feb 18 09:17:17 crc kubenswrapper[4556]: I0218 09:17:17.124942 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"da163339-653b-4057-9c2f-332eb6957f40","Type":"ContainerStarted","Data":"597192aaccaae5042d2a22d6490ad7fe26c3cefa3516c516d9c344f580dd0c3e"} Feb 18 09:17:17 crc kubenswrapper[4556]: I0218 09:17:17.124949 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"da163339-653b-4057-9c2f-332eb6957f40","Type":"ContainerStarted","Data":"cf805dab6a64bc47eeadafa03f6ddb172f3474ee737eda67956f5772a788bf82"} Feb 18 09:17:17 crc kubenswrapper[4556]: I0218 09:17:17.137986 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-fxcc2" podStartSLOduration=3.029584109 podStartE2EDuration="12.137967324s" podCreationTimestamp="2026-02-18 09:17:05 +0000 UTC" firstStartedPulling="2026-02-18 09:17:06.561968647 +0000 UTC m=+783.578929637" lastFinishedPulling="2026-02-18 09:17:15.670351872 +0000 UTC m=+792.687312852" observedRunningTime="2026-02-18 09:17:17.134791099 +0000 UTC m=+794.151752079" watchObservedRunningTime="2026-02-18 09:17:17.137967324 +0000 UTC m=+794.154928304" Feb 18 09:17:17 crc kubenswrapper[4556]: I0218 09:17:17.171247 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=19.66011744 podStartE2EDuration="30.171234481s" podCreationTimestamp="2026-02-18 09:16:47 +0000 UTC" firstStartedPulling="2026-02-18 09:17:05.147248337 +0000 UTC m=+782.164209318" lastFinishedPulling="2026-02-18 09:17:15.658365379 +0000 UTC m=+792.675326359" observedRunningTime="2026-02-18 09:17:17.162898398 +0000 UTC m=+794.179859378" watchObservedRunningTime="2026-02-18 09:17:17.171234481 +0000 UTC m=+794.188195462" Feb 18 09:17:17 crc kubenswrapper[4556]: I0218 09:17:17.519322 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-6kkrw-config-q9lcv" podStartSLOduration=2.519117177 podStartE2EDuration="2.519117177s" podCreationTimestamp="2026-02-18 09:17:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:17:17.19069996 +0000 UTC m=+794.207660940" watchObservedRunningTime="2026-02-18 09:17:17.519117177 +0000 UTC m=+794.536078147" Feb 18 09:17:17 crc kubenswrapper[4556]: I0218 09:17:17.524774 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5799dddbf-lfp9w"] Feb 18 09:17:17 crc kubenswrapper[4556]: I0218 09:17:17.538281 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5799dddbf-lfp9w" Feb 18 09:17:17 crc kubenswrapper[4556]: I0218 09:17:17.543292 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Feb 18 09:17:17 crc kubenswrapper[4556]: I0218 09:17:17.544260 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5799dddbf-lfp9w"] Feb 18 09:17:17 crc kubenswrapper[4556]: I0218 09:17:17.618357 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/abced7fe-5537-4e16-bed1-1e74789868f6-dns-swift-storage-0\") pod \"dnsmasq-dns-5799dddbf-lfp9w\" (UID: \"abced7fe-5537-4e16-bed1-1e74789868f6\") " pod="openstack/dnsmasq-dns-5799dddbf-lfp9w" Feb 18 09:17:17 crc kubenswrapper[4556]: I0218 09:17:17.618533 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/abced7fe-5537-4e16-bed1-1e74789868f6-config\") pod \"dnsmasq-dns-5799dddbf-lfp9w\" (UID: \"abced7fe-5537-4e16-bed1-1e74789868f6\") " pod="openstack/dnsmasq-dns-5799dddbf-lfp9w" Feb 18 09:17:17 crc kubenswrapper[4556]: I0218 09:17:17.618656 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/abced7fe-5537-4e16-bed1-1e74789868f6-ovsdbserver-sb\") pod \"dnsmasq-dns-5799dddbf-lfp9w\" (UID: \"abced7fe-5537-4e16-bed1-1e74789868f6\") " pod="openstack/dnsmasq-dns-5799dddbf-lfp9w" Feb 18 09:17:17 crc kubenswrapper[4556]: I0218 09:17:17.618771 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/abced7fe-5537-4e16-bed1-1e74789868f6-ovsdbserver-nb\") pod \"dnsmasq-dns-5799dddbf-lfp9w\" (UID: \"abced7fe-5537-4e16-bed1-1e74789868f6\") " pod="openstack/dnsmasq-dns-5799dddbf-lfp9w" Feb 18 09:17:17 crc kubenswrapper[4556]: I0218 09:17:17.618846 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmb22\" (UniqueName: \"kubernetes.io/projected/abced7fe-5537-4e16-bed1-1e74789868f6-kube-api-access-pmb22\") pod \"dnsmasq-dns-5799dddbf-lfp9w\" (UID: \"abced7fe-5537-4e16-bed1-1e74789868f6\") " pod="openstack/dnsmasq-dns-5799dddbf-lfp9w" Feb 18 09:17:17 crc kubenswrapper[4556]: I0218 09:17:17.618931 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/abced7fe-5537-4e16-bed1-1e74789868f6-dns-svc\") pod \"dnsmasq-dns-5799dddbf-lfp9w\" (UID: \"abced7fe-5537-4e16-bed1-1e74789868f6\") " pod="openstack/dnsmasq-dns-5799dddbf-lfp9w" Feb 18 09:17:17 crc kubenswrapper[4556]: I0218 09:17:17.720417 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmb22\" (UniqueName: \"kubernetes.io/projected/abced7fe-5537-4e16-bed1-1e74789868f6-kube-api-access-pmb22\") pod \"dnsmasq-dns-5799dddbf-lfp9w\" (UID: \"abced7fe-5537-4e16-bed1-1e74789868f6\") " pod="openstack/dnsmasq-dns-5799dddbf-lfp9w" Feb 18 09:17:17 crc kubenswrapper[4556]: I0218 09:17:17.720521 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/abced7fe-5537-4e16-bed1-1e74789868f6-dns-svc\") pod \"dnsmasq-dns-5799dddbf-lfp9w\" (UID: \"abced7fe-5537-4e16-bed1-1e74789868f6\") " pod="openstack/dnsmasq-dns-5799dddbf-lfp9w" Feb 18 09:17:17 crc kubenswrapper[4556]: I0218 09:17:17.720551 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/abced7fe-5537-4e16-bed1-1e74789868f6-dns-swift-storage-0\") pod \"dnsmasq-dns-5799dddbf-lfp9w\" (UID: \"abced7fe-5537-4e16-bed1-1e74789868f6\") " pod="openstack/dnsmasq-dns-5799dddbf-lfp9w" Feb 18 09:17:17 crc kubenswrapper[4556]: I0218 09:17:17.720636 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/abced7fe-5537-4e16-bed1-1e74789868f6-config\") pod \"dnsmasq-dns-5799dddbf-lfp9w\" (UID: \"abced7fe-5537-4e16-bed1-1e74789868f6\") " pod="openstack/dnsmasq-dns-5799dddbf-lfp9w" Feb 18 09:17:17 crc kubenswrapper[4556]: I0218 09:17:17.720717 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/abced7fe-5537-4e16-bed1-1e74789868f6-ovsdbserver-sb\") pod \"dnsmasq-dns-5799dddbf-lfp9w\" (UID: \"abced7fe-5537-4e16-bed1-1e74789868f6\") " pod="openstack/dnsmasq-dns-5799dddbf-lfp9w" Feb 18 09:17:17 crc kubenswrapper[4556]: I0218 09:17:17.720790 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/abced7fe-5537-4e16-bed1-1e74789868f6-ovsdbserver-nb\") pod \"dnsmasq-dns-5799dddbf-lfp9w\" (UID: \"abced7fe-5537-4e16-bed1-1e74789868f6\") " pod="openstack/dnsmasq-dns-5799dddbf-lfp9w" Feb 18 09:17:17 crc kubenswrapper[4556]: I0218 09:17:17.721463 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/abced7fe-5537-4e16-bed1-1e74789868f6-dns-svc\") pod \"dnsmasq-dns-5799dddbf-lfp9w\" (UID: \"abced7fe-5537-4e16-bed1-1e74789868f6\") " pod="openstack/dnsmasq-dns-5799dddbf-lfp9w" Feb 18 09:17:17 crc kubenswrapper[4556]: I0218 09:17:17.721566 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/abced7fe-5537-4e16-bed1-1e74789868f6-dns-swift-storage-0\") pod \"dnsmasq-dns-5799dddbf-lfp9w\" (UID: \"abced7fe-5537-4e16-bed1-1e74789868f6\") " pod="openstack/dnsmasq-dns-5799dddbf-lfp9w" Feb 18 09:17:17 crc kubenswrapper[4556]: I0218 09:17:17.721623 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/abced7fe-5537-4e16-bed1-1e74789868f6-ovsdbserver-sb\") pod \"dnsmasq-dns-5799dddbf-lfp9w\" (UID: \"abced7fe-5537-4e16-bed1-1e74789868f6\") " pod="openstack/dnsmasq-dns-5799dddbf-lfp9w" Feb 18 09:17:17 crc kubenswrapper[4556]: I0218 09:17:17.721899 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/abced7fe-5537-4e16-bed1-1e74789868f6-ovsdbserver-nb\") pod \"dnsmasq-dns-5799dddbf-lfp9w\" (UID: \"abced7fe-5537-4e16-bed1-1e74789868f6\") " pod="openstack/dnsmasq-dns-5799dddbf-lfp9w" Feb 18 09:17:17 crc kubenswrapper[4556]: I0218 09:17:17.721911 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/abced7fe-5537-4e16-bed1-1e74789868f6-config\") pod \"dnsmasq-dns-5799dddbf-lfp9w\" (UID: \"abced7fe-5537-4e16-bed1-1e74789868f6\") " pod="openstack/dnsmasq-dns-5799dddbf-lfp9w" Feb 18 09:17:17 crc kubenswrapper[4556]: I0218 09:17:17.752427 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmb22\" (UniqueName: \"kubernetes.io/projected/abced7fe-5537-4e16-bed1-1e74789868f6-kube-api-access-pmb22\") pod \"dnsmasq-dns-5799dddbf-lfp9w\" (UID: \"abced7fe-5537-4e16-bed1-1e74789868f6\") " pod="openstack/dnsmasq-dns-5799dddbf-lfp9w" Feb 18 09:17:17 crc kubenswrapper[4556]: I0218 09:17:17.857948 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5799dddbf-lfp9w" Feb 18 09:17:18 crc kubenswrapper[4556]: I0218 09:17:18.142475 4556 generic.go:334] "Generic (PLEG): container finished" podID="75a278f7-6511-40c1-ab7e-82644755b10b" containerID="ca0d831b26e0653479e1280cd1625ab6207b9f101f1bbc574d2caad852610ea1" exitCode=0 Feb 18 09:17:18 crc kubenswrapper[4556]: I0218 09:17:18.145490 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-6kkrw-config-q9lcv" event={"ID":"75a278f7-6511-40c1-ab7e-82644755b10b","Type":"ContainerDied","Data":"ca0d831b26e0653479e1280cd1625ab6207b9f101f1bbc574d2caad852610ea1"} Feb 18 09:17:18 crc kubenswrapper[4556]: I0218 09:17:18.315238 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5799dddbf-lfp9w"] Feb 18 09:17:19 crc kubenswrapper[4556]: I0218 09:17:19.154049 4556 generic.go:334] "Generic (PLEG): container finished" podID="abced7fe-5537-4e16-bed1-1e74789868f6" containerID="1ffa701fb8c920213ce1d746e4e846096f1a02278da39521d16a888929099a9e" exitCode=0 Feb 18 09:17:19 crc kubenswrapper[4556]: I0218 09:17:19.154217 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5799dddbf-lfp9w" event={"ID":"abced7fe-5537-4e16-bed1-1e74789868f6","Type":"ContainerDied","Data":"1ffa701fb8c920213ce1d746e4e846096f1a02278da39521d16a888929099a9e"} Feb 18 09:17:19 crc kubenswrapper[4556]: I0218 09:17:19.154460 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5799dddbf-lfp9w" event={"ID":"abced7fe-5537-4e16-bed1-1e74789868f6","Type":"ContainerStarted","Data":"a3eacd5c4c0cc133579c5dec151d5353633a7818ac6f0a11cc90219e61719a99"} Feb 18 09:17:19 crc kubenswrapper[4556]: I0218 09:17:19.386848 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-6kkrw-config-q9lcv" Feb 18 09:17:19 crc kubenswrapper[4556]: I0218 09:17:19.550322 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/75a278f7-6511-40c1-ab7e-82644755b10b-var-run-ovn\") pod \"75a278f7-6511-40c1-ab7e-82644755b10b\" (UID: \"75a278f7-6511-40c1-ab7e-82644755b10b\") " Feb 18 09:17:19 crc kubenswrapper[4556]: I0218 09:17:19.550389 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/75a278f7-6511-40c1-ab7e-82644755b10b-additional-scripts\") pod \"75a278f7-6511-40c1-ab7e-82644755b10b\" (UID: \"75a278f7-6511-40c1-ab7e-82644755b10b\") " Feb 18 09:17:19 crc kubenswrapper[4556]: I0218 09:17:19.550411 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lrg2b\" (UniqueName: \"kubernetes.io/projected/75a278f7-6511-40c1-ab7e-82644755b10b-kube-api-access-lrg2b\") pod \"75a278f7-6511-40c1-ab7e-82644755b10b\" (UID: \"75a278f7-6511-40c1-ab7e-82644755b10b\") " Feb 18 09:17:19 crc kubenswrapper[4556]: I0218 09:17:19.550422 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/75a278f7-6511-40c1-ab7e-82644755b10b-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "75a278f7-6511-40c1-ab7e-82644755b10b" (UID: "75a278f7-6511-40c1-ab7e-82644755b10b"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 09:17:19 crc kubenswrapper[4556]: I0218 09:17:19.550442 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/75a278f7-6511-40c1-ab7e-82644755b10b-scripts\") pod \"75a278f7-6511-40c1-ab7e-82644755b10b\" (UID: \"75a278f7-6511-40c1-ab7e-82644755b10b\") " Feb 18 09:17:19 crc kubenswrapper[4556]: I0218 09:17:19.550532 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/75a278f7-6511-40c1-ab7e-82644755b10b-var-log-ovn\") pod \"75a278f7-6511-40c1-ab7e-82644755b10b\" (UID: \"75a278f7-6511-40c1-ab7e-82644755b10b\") " Feb 18 09:17:19 crc kubenswrapper[4556]: I0218 09:17:19.550574 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/75a278f7-6511-40c1-ab7e-82644755b10b-var-run\") pod \"75a278f7-6511-40c1-ab7e-82644755b10b\" (UID: \"75a278f7-6511-40c1-ab7e-82644755b10b\") " Feb 18 09:17:19 crc kubenswrapper[4556]: I0218 09:17:19.550617 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/75a278f7-6511-40c1-ab7e-82644755b10b-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "75a278f7-6511-40c1-ab7e-82644755b10b" (UID: "75a278f7-6511-40c1-ab7e-82644755b10b"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 09:17:19 crc kubenswrapper[4556]: I0218 09:17:19.550732 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/75a278f7-6511-40c1-ab7e-82644755b10b-var-run" (OuterVolumeSpecName: "var-run") pod "75a278f7-6511-40c1-ab7e-82644755b10b" (UID: "75a278f7-6511-40c1-ab7e-82644755b10b"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 09:17:19 crc kubenswrapper[4556]: I0218 09:17:19.551101 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75a278f7-6511-40c1-ab7e-82644755b10b-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "75a278f7-6511-40c1-ab7e-82644755b10b" (UID: "75a278f7-6511-40c1-ab7e-82644755b10b"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:17:19 crc kubenswrapper[4556]: I0218 09:17:19.551209 4556 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/75a278f7-6511-40c1-ab7e-82644755b10b-var-log-ovn\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:19 crc kubenswrapper[4556]: I0218 09:17:19.551226 4556 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/75a278f7-6511-40c1-ab7e-82644755b10b-var-run\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:19 crc kubenswrapper[4556]: I0218 09:17:19.551235 4556 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/75a278f7-6511-40c1-ab7e-82644755b10b-var-run-ovn\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:19 crc kubenswrapper[4556]: I0218 09:17:19.551244 4556 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/75a278f7-6511-40c1-ab7e-82644755b10b-additional-scripts\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:19 crc kubenswrapper[4556]: I0218 09:17:19.551358 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75a278f7-6511-40c1-ab7e-82644755b10b-scripts" (OuterVolumeSpecName: "scripts") pod "75a278f7-6511-40c1-ab7e-82644755b10b" (UID: "75a278f7-6511-40c1-ab7e-82644755b10b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:17:19 crc kubenswrapper[4556]: I0218 09:17:19.555707 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75a278f7-6511-40c1-ab7e-82644755b10b-kube-api-access-lrg2b" (OuterVolumeSpecName: "kube-api-access-lrg2b") pod "75a278f7-6511-40c1-ab7e-82644755b10b" (UID: "75a278f7-6511-40c1-ab7e-82644755b10b"). InnerVolumeSpecName "kube-api-access-lrg2b". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:17:19 crc kubenswrapper[4556]: I0218 09:17:19.653449 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lrg2b\" (UniqueName: \"kubernetes.io/projected/75a278f7-6511-40c1-ab7e-82644755b10b-kube-api-access-lrg2b\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:19 crc kubenswrapper[4556]: I0218 09:17:19.653474 4556 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/75a278f7-6511-40c1-ab7e-82644755b10b-scripts\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:20 crc kubenswrapper[4556]: I0218 09:17:20.164080 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5799dddbf-lfp9w" event={"ID":"abced7fe-5537-4e16-bed1-1e74789868f6","Type":"ContainerStarted","Data":"b112f4d4ad233f7191de76ede77999cee186510f61d8d1143e824fc1e00f7470"} Feb 18 09:17:20 crc kubenswrapper[4556]: I0218 09:17:20.164266 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5799dddbf-lfp9w" Feb 18 09:17:20 crc kubenswrapper[4556]: I0218 09:17:20.165997 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-6kkrw-config-q9lcv" event={"ID":"75a278f7-6511-40c1-ab7e-82644755b10b","Type":"ContainerDied","Data":"258273d4269ec5fcd95e058724c2c2d29a9d6ea5218988e0f3e5862c01dcecd5"} Feb 18 09:17:20 crc kubenswrapper[4556]: I0218 09:17:20.166030 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="258273d4269ec5fcd95e058724c2c2d29a9d6ea5218988e0f3e5862c01dcecd5" Feb 18 09:17:20 crc kubenswrapper[4556]: I0218 09:17:20.166054 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-6kkrw-config-q9lcv" Feb 18 09:17:20 crc kubenswrapper[4556]: I0218 09:17:20.194092 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5799dddbf-lfp9w" podStartSLOduration=3.194076822 podStartE2EDuration="3.194076822s" podCreationTimestamp="2026-02-18 09:17:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:17:20.183870971 +0000 UTC m=+797.200831950" watchObservedRunningTime="2026-02-18 09:17:20.194076822 +0000 UTC m=+797.211037802" Feb 18 09:17:20 crc kubenswrapper[4556]: I0218 09:17:20.471280 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-6kkrw-config-q9lcv"] Feb 18 09:17:20 crc kubenswrapper[4556]: I0218 09:17:20.480505 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-6kkrw-config-q9lcv"] Feb 18 09:17:20 crc kubenswrapper[4556]: I0218 09:17:20.657629 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-6kkrw" Feb 18 09:17:21 crc kubenswrapper[4556]: I0218 09:17:21.176367 4556 generic.go:334] "Generic (PLEG): container finished" podID="4731e0af-7aff-41c9-bc33-ad533ee750ce" containerID="c429fb238171b24a587bcafa3d8a82c21f54094c5652e03e587e69b41a0106c0" exitCode=0 Feb 18 09:17:21 crc kubenswrapper[4556]: I0218 09:17:21.176473 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-fxcc2" event={"ID":"4731e0af-7aff-41c9-bc33-ad533ee750ce","Type":"ContainerDied","Data":"c429fb238171b24a587bcafa3d8a82c21f54094c5652e03e587e69b41a0106c0"} Feb 18 09:17:21 crc kubenswrapper[4556]: I0218 09:17:21.292350 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75a278f7-6511-40c1-ab7e-82644755b10b" path="/var/lib/kubelet/pods/75a278f7-6511-40c1-ab7e-82644755b10b/volumes" Feb 18 09:17:22 crc kubenswrapper[4556]: I0218 09:17:22.497594 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-fxcc2" Feb 18 09:17:22 crc kubenswrapper[4556]: I0218 09:17:22.616352 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4731e0af-7aff-41c9-bc33-ad533ee750ce-db-sync-config-data\") pod \"4731e0af-7aff-41c9-bc33-ad533ee750ce\" (UID: \"4731e0af-7aff-41c9-bc33-ad533ee750ce\") " Feb 18 09:17:22 crc kubenswrapper[4556]: I0218 09:17:22.616477 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fj5m8\" (UniqueName: \"kubernetes.io/projected/4731e0af-7aff-41c9-bc33-ad533ee750ce-kube-api-access-fj5m8\") pod \"4731e0af-7aff-41c9-bc33-ad533ee750ce\" (UID: \"4731e0af-7aff-41c9-bc33-ad533ee750ce\") " Feb 18 09:17:22 crc kubenswrapper[4556]: I0218 09:17:22.616659 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4731e0af-7aff-41c9-bc33-ad533ee750ce-config-data\") pod \"4731e0af-7aff-41c9-bc33-ad533ee750ce\" (UID: \"4731e0af-7aff-41c9-bc33-ad533ee750ce\") " Feb 18 09:17:22 crc kubenswrapper[4556]: I0218 09:17:22.616708 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4731e0af-7aff-41c9-bc33-ad533ee750ce-combined-ca-bundle\") pod \"4731e0af-7aff-41c9-bc33-ad533ee750ce\" (UID: \"4731e0af-7aff-41c9-bc33-ad533ee750ce\") " Feb 18 09:17:22 crc kubenswrapper[4556]: I0218 09:17:22.621966 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4731e0af-7aff-41c9-bc33-ad533ee750ce-kube-api-access-fj5m8" (OuterVolumeSpecName: "kube-api-access-fj5m8") pod "4731e0af-7aff-41c9-bc33-ad533ee750ce" (UID: "4731e0af-7aff-41c9-bc33-ad533ee750ce"). InnerVolumeSpecName "kube-api-access-fj5m8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:17:22 crc kubenswrapper[4556]: I0218 09:17:22.624603 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4731e0af-7aff-41c9-bc33-ad533ee750ce-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "4731e0af-7aff-41c9-bc33-ad533ee750ce" (UID: "4731e0af-7aff-41c9-bc33-ad533ee750ce"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:17:22 crc kubenswrapper[4556]: I0218 09:17:22.638901 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4731e0af-7aff-41c9-bc33-ad533ee750ce-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4731e0af-7aff-41c9-bc33-ad533ee750ce" (UID: "4731e0af-7aff-41c9-bc33-ad533ee750ce"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:17:22 crc kubenswrapper[4556]: I0218 09:17:22.651308 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4731e0af-7aff-41c9-bc33-ad533ee750ce-config-data" (OuterVolumeSpecName: "config-data") pod "4731e0af-7aff-41c9-bc33-ad533ee750ce" (UID: "4731e0af-7aff-41c9-bc33-ad533ee750ce"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:17:22 crc kubenswrapper[4556]: I0218 09:17:22.719806 4556 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4731e0af-7aff-41c9-bc33-ad533ee750ce-config-data\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:22 crc kubenswrapper[4556]: I0218 09:17:22.719832 4556 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4731e0af-7aff-41c9-bc33-ad533ee750ce-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:22 crc kubenswrapper[4556]: I0218 09:17:22.719845 4556 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4731e0af-7aff-41c9-bc33-ad533ee750ce-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:22 crc kubenswrapper[4556]: I0218 09:17:22.719855 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fj5m8\" (UniqueName: \"kubernetes.io/projected/4731e0af-7aff-41c9-bc33-ad533ee750ce-kube-api-access-fj5m8\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:23 crc kubenswrapper[4556]: I0218 09:17:23.192707 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-fxcc2" event={"ID":"4731e0af-7aff-41c9-bc33-ad533ee750ce","Type":"ContainerDied","Data":"906ff112dbba860d29503d7bae98de2325f4607da2efd57c1e572efb86861079"} Feb 18 09:17:23 crc kubenswrapper[4556]: I0218 09:17:23.192957 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-fxcc2" Feb 18 09:17:23 crc kubenswrapper[4556]: I0218 09:17:23.192988 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="906ff112dbba860d29503d7bae98de2325f4607da2efd57c1e572efb86861079" Feb 18 09:17:23 crc kubenswrapper[4556]: I0218 09:17:23.653231 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5799dddbf-lfp9w"] Feb 18 09:17:23 crc kubenswrapper[4556]: I0218 09:17:23.653416 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5799dddbf-lfp9w" podUID="abced7fe-5537-4e16-bed1-1e74789868f6" containerName="dnsmasq-dns" containerID="cri-o://b112f4d4ad233f7191de76ede77999cee186510f61d8d1143e824fc1e00f7470" gracePeriod=10 Feb 18 09:17:23 crc kubenswrapper[4556]: I0218 09:17:23.686178 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-54cc5c87cc-ck95n"] Feb 18 09:17:23 crc kubenswrapper[4556]: E0218 09:17:23.686671 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75a278f7-6511-40c1-ab7e-82644755b10b" containerName="ovn-config" Feb 18 09:17:23 crc kubenswrapper[4556]: I0218 09:17:23.686708 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="75a278f7-6511-40c1-ab7e-82644755b10b" containerName="ovn-config" Feb 18 09:17:23 crc kubenswrapper[4556]: E0218 09:17:23.686743 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4731e0af-7aff-41c9-bc33-ad533ee750ce" containerName="glance-db-sync" Feb 18 09:17:23 crc kubenswrapper[4556]: I0218 09:17:23.686750 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="4731e0af-7aff-41c9-bc33-ad533ee750ce" containerName="glance-db-sync" Feb 18 09:17:23 crc kubenswrapper[4556]: I0218 09:17:23.687038 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="75a278f7-6511-40c1-ab7e-82644755b10b" containerName="ovn-config" Feb 18 09:17:23 crc kubenswrapper[4556]: I0218 09:17:23.687067 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="4731e0af-7aff-41c9-bc33-ad533ee750ce" containerName="glance-db-sync" Feb 18 09:17:23 crc kubenswrapper[4556]: I0218 09:17:23.691635 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54cc5c87cc-ck95n" Feb 18 09:17:23 crc kubenswrapper[4556]: I0218 09:17:23.698472 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54cc5c87cc-ck95n"] Feb 18 09:17:23 crc kubenswrapper[4556]: I0218 09:17:23.843314 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/269247cf-79a0-49a1-8fc9-da82411da8e5-ovsdbserver-nb\") pod \"dnsmasq-dns-54cc5c87cc-ck95n\" (UID: \"269247cf-79a0-49a1-8fc9-da82411da8e5\") " pod="openstack/dnsmasq-dns-54cc5c87cc-ck95n" Feb 18 09:17:23 crc kubenswrapper[4556]: I0218 09:17:23.843648 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/269247cf-79a0-49a1-8fc9-da82411da8e5-dns-svc\") pod \"dnsmasq-dns-54cc5c87cc-ck95n\" (UID: \"269247cf-79a0-49a1-8fc9-da82411da8e5\") " pod="openstack/dnsmasq-dns-54cc5c87cc-ck95n" Feb 18 09:17:23 crc kubenswrapper[4556]: I0218 09:17:23.843698 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/269247cf-79a0-49a1-8fc9-da82411da8e5-dns-swift-storage-0\") pod \"dnsmasq-dns-54cc5c87cc-ck95n\" (UID: \"269247cf-79a0-49a1-8fc9-da82411da8e5\") " pod="openstack/dnsmasq-dns-54cc5c87cc-ck95n" Feb 18 09:17:23 crc kubenswrapper[4556]: I0218 09:17:23.843732 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7cwx\" (UniqueName: \"kubernetes.io/projected/269247cf-79a0-49a1-8fc9-da82411da8e5-kube-api-access-t7cwx\") pod \"dnsmasq-dns-54cc5c87cc-ck95n\" (UID: \"269247cf-79a0-49a1-8fc9-da82411da8e5\") " pod="openstack/dnsmasq-dns-54cc5c87cc-ck95n" Feb 18 09:17:23 crc kubenswrapper[4556]: I0218 09:17:23.843773 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/269247cf-79a0-49a1-8fc9-da82411da8e5-ovsdbserver-sb\") pod \"dnsmasq-dns-54cc5c87cc-ck95n\" (UID: \"269247cf-79a0-49a1-8fc9-da82411da8e5\") " pod="openstack/dnsmasq-dns-54cc5c87cc-ck95n" Feb 18 09:17:23 crc kubenswrapper[4556]: I0218 09:17:23.843797 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/269247cf-79a0-49a1-8fc9-da82411da8e5-config\") pod \"dnsmasq-dns-54cc5c87cc-ck95n\" (UID: \"269247cf-79a0-49a1-8fc9-da82411da8e5\") " pod="openstack/dnsmasq-dns-54cc5c87cc-ck95n" Feb 18 09:17:23 crc kubenswrapper[4556]: I0218 09:17:23.945317 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/269247cf-79a0-49a1-8fc9-da82411da8e5-ovsdbserver-sb\") pod \"dnsmasq-dns-54cc5c87cc-ck95n\" (UID: \"269247cf-79a0-49a1-8fc9-da82411da8e5\") " pod="openstack/dnsmasq-dns-54cc5c87cc-ck95n" Feb 18 09:17:23 crc kubenswrapper[4556]: I0218 09:17:23.945395 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/269247cf-79a0-49a1-8fc9-da82411da8e5-config\") pod \"dnsmasq-dns-54cc5c87cc-ck95n\" (UID: \"269247cf-79a0-49a1-8fc9-da82411da8e5\") " pod="openstack/dnsmasq-dns-54cc5c87cc-ck95n" Feb 18 09:17:23 crc kubenswrapper[4556]: I0218 09:17:23.945472 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/269247cf-79a0-49a1-8fc9-da82411da8e5-ovsdbserver-nb\") pod \"dnsmasq-dns-54cc5c87cc-ck95n\" (UID: \"269247cf-79a0-49a1-8fc9-da82411da8e5\") " pod="openstack/dnsmasq-dns-54cc5c87cc-ck95n" Feb 18 09:17:23 crc kubenswrapper[4556]: I0218 09:17:23.945510 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/269247cf-79a0-49a1-8fc9-da82411da8e5-dns-svc\") pod \"dnsmasq-dns-54cc5c87cc-ck95n\" (UID: \"269247cf-79a0-49a1-8fc9-da82411da8e5\") " pod="openstack/dnsmasq-dns-54cc5c87cc-ck95n" Feb 18 09:17:23 crc kubenswrapper[4556]: I0218 09:17:23.945585 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/269247cf-79a0-49a1-8fc9-da82411da8e5-dns-swift-storage-0\") pod \"dnsmasq-dns-54cc5c87cc-ck95n\" (UID: \"269247cf-79a0-49a1-8fc9-da82411da8e5\") " pod="openstack/dnsmasq-dns-54cc5c87cc-ck95n" Feb 18 09:17:23 crc kubenswrapper[4556]: I0218 09:17:23.945626 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7cwx\" (UniqueName: \"kubernetes.io/projected/269247cf-79a0-49a1-8fc9-da82411da8e5-kube-api-access-t7cwx\") pod \"dnsmasq-dns-54cc5c87cc-ck95n\" (UID: \"269247cf-79a0-49a1-8fc9-da82411da8e5\") " pod="openstack/dnsmasq-dns-54cc5c87cc-ck95n" Feb 18 09:17:23 crc kubenswrapper[4556]: I0218 09:17:23.946593 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/269247cf-79a0-49a1-8fc9-da82411da8e5-dns-svc\") pod \"dnsmasq-dns-54cc5c87cc-ck95n\" (UID: \"269247cf-79a0-49a1-8fc9-da82411da8e5\") " pod="openstack/dnsmasq-dns-54cc5c87cc-ck95n" Feb 18 09:17:23 crc kubenswrapper[4556]: I0218 09:17:23.946765 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/269247cf-79a0-49a1-8fc9-da82411da8e5-ovsdbserver-nb\") pod \"dnsmasq-dns-54cc5c87cc-ck95n\" (UID: \"269247cf-79a0-49a1-8fc9-da82411da8e5\") " pod="openstack/dnsmasq-dns-54cc5c87cc-ck95n" Feb 18 09:17:23 crc kubenswrapper[4556]: I0218 09:17:23.946969 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/269247cf-79a0-49a1-8fc9-da82411da8e5-config\") pod \"dnsmasq-dns-54cc5c87cc-ck95n\" (UID: \"269247cf-79a0-49a1-8fc9-da82411da8e5\") " pod="openstack/dnsmasq-dns-54cc5c87cc-ck95n" Feb 18 09:17:23 crc kubenswrapper[4556]: I0218 09:17:23.947960 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/269247cf-79a0-49a1-8fc9-da82411da8e5-ovsdbserver-sb\") pod \"dnsmasq-dns-54cc5c87cc-ck95n\" (UID: \"269247cf-79a0-49a1-8fc9-da82411da8e5\") " pod="openstack/dnsmasq-dns-54cc5c87cc-ck95n" Feb 18 09:17:23 crc kubenswrapper[4556]: I0218 09:17:23.948531 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/269247cf-79a0-49a1-8fc9-da82411da8e5-dns-swift-storage-0\") pod \"dnsmasq-dns-54cc5c87cc-ck95n\" (UID: \"269247cf-79a0-49a1-8fc9-da82411da8e5\") " pod="openstack/dnsmasq-dns-54cc5c87cc-ck95n" Feb 18 09:17:23 crc kubenswrapper[4556]: I0218 09:17:23.964858 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7cwx\" (UniqueName: \"kubernetes.io/projected/269247cf-79a0-49a1-8fc9-da82411da8e5-kube-api-access-t7cwx\") pod \"dnsmasq-dns-54cc5c87cc-ck95n\" (UID: \"269247cf-79a0-49a1-8fc9-da82411da8e5\") " pod="openstack/dnsmasq-dns-54cc5c87cc-ck95n" Feb 18 09:17:24 crc kubenswrapper[4556]: I0218 09:17:24.027279 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5799dddbf-lfp9w" Feb 18 09:17:24 crc kubenswrapper[4556]: I0218 09:17:24.057283 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54cc5c87cc-ck95n" Feb 18 09:17:24 crc kubenswrapper[4556]: I0218 09:17:24.150968 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pmb22\" (UniqueName: \"kubernetes.io/projected/abced7fe-5537-4e16-bed1-1e74789868f6-kube-api-access-pmb22\") pod \"abced7fe-5537-4e16-bed1-1e74789868f6\" (UID: \"abced7fe-5537-4e16-bed1-1e74789868f6\") " Feb 18 09:17:24 crc kubenswrapper[4556]: I0218 09:17:24.151513 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/abced7fe-5537-4e16-bed1-1e74789868f6-dns-svc\") pod \"abced7fe-5537-4e16-bed1-1e74789868f6\" (UID: \"abced7fe-5537-4e16-bed1-1e74789868f6\") " Feb 18 09:17:24 crc kubenswrapper[4556]: I0218 09:17:24.151629 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/abced7fe-5537-4e16-bed1-1e74789868f6-ovsdbserver-nb\") pod \"abced7fe-5537-4e16-bed1-1e74789868f6\" (UID: \"abced7fe-5537-4e16-bed1-1e74789868f6\") " Feb 18 09:17:24 crc kubenswrapper[4556]: I0218 09:17:24.151649 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/abced7fe-5537-4e16-bed1-1e74789868f6-config\") pod \"abced7fe-5537-4e16-bed1-1e74789868f6\" (UID: \"abced7fe-5537-4e16-bed1-1e74789868f6\") " Feb 18 09:17:24 crc kubenswrapper[4556]: I0218 09:17:24.151755 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/abced7fe-5537-4e16-bed1-1e74789868f6-dns-swift-storage-0\") pod \"abced7fe-5537-4e16-bed1-1e74789868f6\" (UID: \"abced7fe-5537-4e16-bed1-1e74789868f6\") " Feb 18 09:17:24 crc kubenswrapper[4556]: I0218 09:17:24.151791 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/abced7fe-5537-4e16-bed1-1e74789868f6-ovsdbserver-sb\") pod \"abced7fe-5537-4e16-bed1-1e74789868f6\" (UID: \"abced7fe-5537-4e16-bed1-1e74789868f6\") " Feb 18 09:17:24 crc kubenswrapper[4556]: I0218 09:17:24.155002 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abced7fe-5537-4e16-bed1-1e74789868f6-kube-api-access-pmb22" (OuterVolumeSpecName: "kube-api-access-pmb22") pod "abced7fe-5537-4e16-bed1-1e74789868f6" (UID: "abced7fe-5537-4e16-bed1-1e74789868f6"). InnerVolumeSpecName "kube-api-access-pmb22". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:17:24 crc kubenswrapper[4556]: I0218 09:17:24.190473 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/abced7fe-5537-4e16-bed1-1e74789868f6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "abced7fe-5537-4e16-bed1-1e74789868f6" (UID: "abced7fe-5537-4e16-bed1-1e74789868f6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:17:24 crc kubenswrapper[4556]: I0218 09:17:24.197353 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/abced7fe-5537-4e16-bed1-1e74789868f6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "abced7fe-5537-4e16-bed1-1e74789868f6" (UID: "abced7fe-5537-4e16-bed1-1e74789868f6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:17:24 crc kubenswrapper[4556]: I0218 09:17:24.206116 4556 generic.go:334] "Generic (PLEG): container finished" podID="abced7fe-5537-4e16-bed1-1e74789868f6" containerID="b112f4d4ad233f7191de76ede77999cee186510f61d8d1143e824fc1e00f7470" exitCode=0 Feb 18 09:17:24 crc kubenswrapper[4556]: I0218 09:17:24.206209 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5799dddbf-lfp9w" Feb 18 09:17:24 crc kubenswrapper[4556]: I0218 09:17:24.206204 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5799dddbf-lfp9w" event={"ID":"abced7fe-5537-4e16-bed1-1e74789868f6","Type":"ContainerDied","Data":"b112f4d4ad233f7191de76ede77999cee186510f61d8d1143e824fc1e00f7470"} Feb 18 09:17:24 crc kubenswrapper[4556]: I0218 09:17:24.206359 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5799dddbf-lfp9w" event={"ID":"abced7fe-5537-4e16-bed1-1e74789868f6","Type":"ContainerDied","Data":"a3eacd5c4c0cc133579c5dec151d5353633a7818ac6f0a11cc90219e61719a99"} Feb 18 09:17:24 crc kubenswrapper[4556]: I0218 09:17:24.206390 4556 scope.go:117] "RemoveContainer" containerID="b112f4d4ad233f7191de76ede77999cee186510f61d8d1143e824fc1e00f7470" Feb 18 09:17:24 crc kubenswrapper[4556]: I0218 09:17:24.208395 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/abced7fe-5537-4e16-bed1-1e74789868f6-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "abced7fe-5537-4e16-bed1-1e74789868f6" (UID: "abced7fe-5537-4e16-bed1-1e74789868f6"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:17:24 crc kubenswrapper[4556]: I0218 09:17:24.209136 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/abced7fe-5537-4e16-bed1-1e74789868f6-config" (OuterVolumeSpecName: "config") pod "abced7fe-5537-4e16-bed1-1e74789868f6" (UID: "abced7fe-5537-4e16-bed1-1e74789868f6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:17:24 crc kubenswrapper[4556]: I0218 09:17:24.216370 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/abced7fe-5537-4e16-bed1-1e74789868f6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "abced7fe-5537-4e16-bed1-1e74789868f6" (UID: "abced7fe-5537-4e16-bed1-1e74789868f6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:17:24 crc kubenswrapper[4556]: I0218 09:17:24.233710 4556 scope.go:117] "RemoveContainer" containerID="1ffa701fb8c920213ce1d746e4e846096f1a02278da39521d16a888929099a9e" Feb 18 09:17:24 crc kubenswrapper[4556]: I0218 09:17:24.253318 4556 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/abced7fe-5537-4e16-bed1-1e74789868f6-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:24 crc kubenswrapper[4556]: I0218 09:17:24.253345 4556 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/abced7fe-5537-4e16-bed1-1e74789868f6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:24 crc kubenswrapper[4556]: I0218 09:17:24.253356 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pmb22\" (UniqueName: \"kubernetes.io/projected/abced7fe-5537-4e16-bed1-1e74789868f6-kube-api-access-pmb22\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:24 crc kubenswrapper[4556]: I0218 09:17:24.253380 4556 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/abced7fe-5537-4e16-bed1-1e74789868f6-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:24 crc kubenswrapper[4556]: I0218 09:17:24.253390 4556 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/abced7fe-5537-4e16-bed1-1e74789868f6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:24 crc kubenswrapper[4556]: I0218 09:17:24.253400 4556 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/abced7fe-5537-4e16-bed1-1e74789868f6-config\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:24 crc kubenswrapper[4556]: I0218 09:17:24.258871 4556 scope.go:117] "RemoveContainer" containerID="b112f4d4ad233f7191de76ede77999cee186510f61d8d1143e824fc1e00f7470" Feb 18 09:17:24 crc kubenswrapper[4556]: E0218 09:17:24.259544 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b112f4d4ad233f7191de76ede77999cee186510f61d8d1143e824fc1e00f7470\": container with ID starting with b112f4d4ad233f7191de76ede77999cee186510f61d8d1143e824fc1e00f7470 not found: ID does not exist" containerID="b112f4d4ad233f7191de76ede77999cee186510f61d8d1143e824fc1e00f7470" Feb 18 09:17:24 crc kubenswrapper[4556]: I0218 09:17:24.259576 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b112f4d4ad233f7191de76ede77999cee186510f61d8d1143e824fc1e00f7470"} err="failed to get container status \"b112f4d4ad233f7191de76ede77999cee186510f61d8d1143e824fc1e00f7470\": rpc error: code = NotFound desc = could not find container \"b112f4d4ad233f7191de76ede77999cee186510f61d8d1143e824fc1e00f7470\": container with ID starting with b112f4d4ad233f7191de76ede77999cee186510f61d8d1143e824fc1e00f7470 not found: ID does not exist" Feb 18 09:17:24 crc kubenswrapper[4556]: I0218 09:17:24.259597 4556 scope.go:117] "RemoveContainer" containerID="1ffa701fb8c920213ce1d746e4e846096f1a02278da39521d16a888929099a9e" Feb 18 09:17:24 crc kubenswrapper[4556]: E0218 09:17:24.260386 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ffa701fb8c920213ce1d746e4e846096f1a02278da39521d16a888929099a9e\": container with ID starting with 1ffa701fb8c920213ce1d746e4e846096f1a02278da39521d16a888929099a9e not found: ID does not exist" containerID="1ffa701fb8c920213ce1d746e4e846096f1a02278da39521d16a888929099a9e" Feb 18 09:17:24 crc kubenswrapper[4556]: I0218 09:17:24.260422 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ffa701fb8c920213ce1d746e4e846096f1a02278da39521d16a888929099a9e"} err="failed to get container status \"1ffa701fb8c920213ce1d746e4e846096f1a02278da39521d16a888929099a9e\": rpc error: code = NotFound desc = could not find container \"1ffa701fb8c920213ce1d746e4e846096f1a02278da39521d16a888929099a9e\": container with ID starting with 1ffa701fb8c920213ce1d746e4e846096f1a02278da39521d16a888929099a9e not found: ID does not exist" Feb 18 09:17:24 crc kubenswrapper[4556]: I0218 09:17:24.439697 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54cc5c87cc-ck95n"] Feb 18 09:17:24 crc kubenswrapper[4556]: W0218 09:17:24.446651 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod269247cf_79a0_49a1_8fc9_da82411da8e5.slice/crio-a23111e48eb9ed7e01c5600fc97cdec0e8c7e2fdd91d09e5fbafbb3f2e281d74 WatchSource:0}: Error finding container a23111e48eb9ed7e01c5600fc97cdec0e8c7e2fdd91d09e5fbafbb3f2e281d74: Status 404 returned error can't find the container with id a23111e48eb9ed7e01c5600fc97cdec0e8c7e2fdd91d09e5fbafbb3f2e281d74 Feb 18 09:17:24 crc kubenswrapper[4556]: I0218 09:17:24.551073 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5799dddbf-lfp9w"] Feb 18 09:17:24 crc kubenswrapper[4556]: I0218 09:17:24.556837 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5799dddbf-lfp9w"] Feb 18 09:17:25 crc kubenswrapper[4556]: I0218 09:17:25.218524 4556 generic.go:334] "Generic (PLEG): container finished" podID="269247cf-79a0-49a1-8fc9-da82411da8e5" containerID="6b3748e795909615f30929c23ebb9f851741f83ac5a40113175b1337bf80b42f" exitCode=0 Feb 18 09:17:25 crc kubenswrapper[4556]: I0218 09:17:25.218591 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54cc5c87cc-ck95n" event={"ID":"269247cf-79a0-49a1-8fc9-da82411da8e5","Type":"ContainerDied","Data":"6b3748e795909615f30929c23ebb9f851741f83ac5a40113175b1337bf80b42f"} Feb 18 09:17:25 crc kubenswrapper[4556]: I0218 09:17:25.218946 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54cc5c87cc-ck95n" event={"ID":"269247cf-79a0-49a1-8fc9-da82411da8e5","Type":"ContainerStarted","Data":"a23111e48eb9ed7e01c5600fc97cdec0e8c7e2fdd91d09e5fbafbb3f2e281d74"} Feb 18 09:17:25 crc kubenswrapper[4556]: I0218 09:17:25.292055 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="abced7fe-5537-4e16-bed1-1e74789868f6" path="/var/lib/kubelet/pods/abced7fe-5537-4e16-bed1-1e74789868f6/volumes" Feb 18 09:17:28 crc kubenswrapper[4556]: I0218 09:17:28.240984 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54cc5c87cc-ck95n" event={"ID":"269247cf-79a0-49a1-8fc9-da82411da8e5","Type":"ContainerStarted","Data":"a62979af6d8d1b2e9b80ec0dfe2c502d3c4c4c1c4f03145c6c6e9d54c955a58b"} Feb 18 09:17:28 crc kubenswrapper[4556]: I0218 09:17:28.241626 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-54cc5c87cc-ck95n" Feb 18 09:17:28 crc kubenswrapper[4556]: I0218 09:17:28.259028 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-54cc5c87cc-ck95n" podStartSLOduration=5.259016848 podStartE2EDuration="5.259016848s" podCreationTimestamp="2026-02-18 09:17:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:17:28.255275627 +0000 UTC m=+805.272236606" watchObservedRunningTime="2026-02-18 09:17:28.259016848 +0000 UTC m=+805.275977827" Feb 18 09:17:32 crc kubenswrapper[4556]: I0218 09:17:32.499063 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Feb 18 09:17:32 crc kubenswrapper[4556]: I0218 09:17:32.538336 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:17:32 crc kubenswrapper[4556]: I0218 09:17:32.751487 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-v7c5w"] Feb 18 09:17:32 crc kubenswrapper[4556]: E0218 09:17:32.751870 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abced7fe-5537-4e16-bed1-1e74789868f6" containerName="dnsmasq-dns" Feb 18 09:17:32 crc kubenswrapper[4556]: I0218 09:17:32.751888 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="abced7fe-5537-4e16-bed1-1e74789868f6" containerName="dnsmasq-dns" Feb 18 09:17:32 crc kubenswrapper[4556]: E0218 09:17:32.751899 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abced7fe-5537-4e16-bed1-1e74789868f6" containerName="init" Feb 18 09:17:32 crc kubenswrapper[4556]: I0218 09:17:32.751906 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="abced7fe-5537-4e16-bed1-1e74789868f6" containerName="init" Feb 18 09:17:32 crc kubenswrapper[4556]: I0218 09:17:32.752111 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="abced7fe-5537-4e16-bed1-1e74789868f6" containerName="dnsmasq-dns" Feb 18 09:17:32 crc kubenswrapper[4556]: I0218 09:17:32.752731 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-v7c5w" Feb 18 09:17:32 crc kubenswrapper[4556]: I0218 09:17:32.776341 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-v7c5w"] Feb 18 09:17:32 crc kubenswrapper[4556]: I0218 09:17:32.911578 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a037e8b8-1769-4728-a30c-dcb8671d2e1b-operator-scripts\") pod \"cinder-db-create-v7c5w\" (UID: \"a037e8b8-1769-4728-a30c-dcb8671d2e1b\") " pod="openstack/cinder-db-create-v7c5w" Feb 18 09:17:32 crc kubenswrapper[4556]: I0218 09:17:32.911876 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8mh8\" (UniqueName: \"kubernetes.io/projected/a037e8b8-1769-4728-a30c-dcb8671d2e1b-kube-api-access-s8mh8\") pod \"cinder-db-create-v7c5w\" (UID: \"a037e8b8-1769-4728-a30c-dcb8671d2e1b\") " pod="openstack/cinder-db-create-v7c5w" Feb 18 09:17:32 crc kubenswrapper[4556]: I0218 09:17:32.941345 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-2d25r"] Feb 18 09:17:32 crc kubenswrapper[4556]: I0218 09:17:32.942245 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-2d25r" Feb 18 09:17:32 crc kubenswrapper[4556]: I0218 09:17:32.947427 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-8f0c-account-create-update-l4dkr"] Feb 18 09:17:32 crc kubenswrapper[4556]: I0218 09:17:32.948140 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-8f0c-account-create-update-l4dkr" Feb 18 09:17:32 crc kubenswrapper[4556]: I0218 09:17:32.949564 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Feb 18 09:17:32 crc kubenswrapper[4556]: I0218 09:17:32.952620 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-2d25r"] Feb 18 09:17:32 crc kubenswrapper[4556]: I0218 09:17:32.958632 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-8f0c-account-create-update-l4dkr"] Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.014146 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a037e8b8-1769-4728-a30c-dcb8671d2e1b-operator-scripts\") pod \"cinder-db-create-v7c5w\" (UID: \"a037e8b8-1769-4728-a30c-dcb8671d2e1b\") " pod="openstack/cinder-db-create-v7c5w" Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.014291 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8mh8\" (UniqueName: \"kubernetes.io/projected/a037e8b8-1769-4728-a30c-dcb8671d2e1b-kube-api-access-s8mh8\") pod \"cinder-db-create-v7c5w\" (UID: \"a037e8b8-1769-4728-a30c-dcb8671d2e1b\") " pod="openstack/cinder-db-create-v7c5w" Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.016974 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a037e8b8-1769-4728-a30c-dcb8671d2e1b-operator-scripts\") pod \"cinder-db-create-v7c5w\" (UID: \"a037e8b8-1769-4728-a30c-dcb8671d2e1b\") " pod="openstack/cinder-db-create-v7c5w" Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.036027 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8mh8\" (UniqueName: \"kubernetes.io/projected/a037e8b8-1769-4728-a30c-dcb8671d2e1b-kube-api-access-s8mh8\") pod \"cinder-db-create-v7c5w\" (UID: \"a037e8b8-1769-4728-a30c-dcb8671d2e1b\") " pod="openstack/cinder-db-create-v7c5w" Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.065800 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-fbv4q"] Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.067034 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-fbv4q" Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.071375 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-7946-account-create-update-7t9xc"] Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.072094 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-7946-account-create-update-7t9xc" Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.073926 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.075395 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-v7c5w" Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.077000 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-fbv4q"] Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.088339 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-7946-account-create-update-7t9xc"] Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.116371 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a6bb6b9c-b2a6-4908-8da1-c519224d8ef2-operator-scripts\") pod \"cinder-8f0c-account-create-update-l4dkr\" (UID: \"a6bb6b9c-b2a6-4908-8da1-c519224d8ef2\") " pod="openstack/cinder-8f0c-account-create-update-l4dkr" Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.116494 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9f46d1fa-3786-4dde-ba6f-d6775782faeb-operator-scripts\") pod \"barbican-db-create-2d25r\" (UID: \"9f46d1fa-3786-4dde-ba6f-d6775782faeb\") " pod="openstack/barbican-db-create-2d25r" Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.116622 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5t44\" (UniqueName: \"kubernetes.io/projected/a6bb6b9c-b2a6-4908-8da1-c519224d8ef2-kube-api-access-g5t44\") pod \"cinder-8f0c-account-create-update-l4dkr\" (UID: \"a6bb6b9c-b2a6-4908-8da1-c519224d8ef2\") " pod="openstack/cinder-8f0c-account-create-update-l4dkr" Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.116645 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmg5l\" (UniqueName: \"kubernetes.io/projected/9f46d1fa-3786-4dde-ba6f-d6775782faeb-kube-api-access-fmg5l\") pod \"barbican-db-create-2d25r\" (UID: \"9f46d1fa-3786-4dde-ba6f-d6775782faeb\") " pod="openstack/barbican-db-create-2d25r" Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.218326 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zjgp\" (UniqueName: \"kubernetes.io/projected/2856804d-f222-49c2-abc3-153227545c0a-kube-api-access-5zjgp\") pod \"neutron-db-create-fbv4q\" (UID: \"2856804d-f222-49c2-abc3-153227545c0a\") " pod="openstack/neutron-db-create-fbv4q" Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.218752 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a6bb6b9c-b2a6-4908-8da1-c519224d8ef2-operator-scripts\") pod \"cinder-8f0c-account-create-update-l4dkr\" (UID: \"a6bb6b9c-b2a6-4908-8da1-c519224d8ef2\") " pod="openstack/cinder-8f0c-account-create-update-l4dkr" Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.218934 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9f46d1fa-3786-4dde-ba6f-d6775782faeb-operator-scripts\") pod \"barbican-db-create-2d25r\" (UID: \"9f46d1fa-3786-4dde-ba6f-d6775782faeb\") " pod="openstack/barbican-db-create-2d25r" Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.219024 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddtnw\" (UniqueName: \"kubernetes.io/projected/45043340-ef17-4b6c-a6e0-a87eac1942ec-kube-api-access-ddtnw\") pod \"barbican-7946-account-create-update-7t9xc\" (UID: \"45043340-ef17-4b6c-a6e0-a87eac1942ec\") " pod="openstack/barbican-7946-account-create-update-7t9xc" Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.219097 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2856804d-f222-49c2-abc3-153227545c0a-operator-scripts\") pod \"neutron-db-create-fbv4q\" (UID: \"2856804d-f222-49c2-abc3-153227545c0a\") " pod="openstack/neutron-db-create-fbv4q" Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.219132 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/45043340-ef17-4b6c-a6e0-a87eac1942ec-operator-scripts\") pod \"barbican-7946-account-create-update-7t9xc\" (UID: \"45043340-ef17-4b6c-a6e0-a87eac1942ec\") " pod="openstack/barbican-7946-account-create-update-7t9xc" Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.219222 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5t44\" (UniqueName: \"kubernetes.io/projected/a6bb6b9c-b2a6-4908-8da1-c519224d8ef2-kube-api-access-g5t44\") pod \"cinder-8f0c-account-create-update-l4dkr\" (UID: \"a6bb6b9c-b2a6-4908-8da1-c519224d8ef2\") " pod="openstack/cinder-8f0c-account-create-update-l4dkr" Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.219250 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmg5l\" (UniqueName: \"kubernetes.io/projected/9f46d1fa-3786-4dde-ba6f-d6775782faeb-kube-api-access-fmg5l\") pod \"barbican-db-create-2d25r\" (UID: \"9f46d1fa-3786-4dde-ba6f-d6775782faeb\") " pod="openstack/barbican-db-create-2d25r" Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.219942 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9f46d1fa-3786-4dde-ba6f-d6775782faeb-operator-scripts\") pod \"barbican-db-create-2d25r\" (UID: \"9f46d1fa-3786-4dde-ba6f-d6775782faeb\") " pod="openstack/barbican-db-create-2d25r" Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.220443 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a6bb6b9c-b2a6-4908-8da1-c519224d8ef2-operator-scripts\") pod \"cinder-8f0c-account-create-update-l4dkr\" (UID: \"a6bb6b9c-b2a6-4908-8da1-c519224d8ef2\") " pod="openstack/cinder-8f0c-account-create-update-l4dkr" Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.226228 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-hzc26"] Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.227465 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-hzc26" Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.235714 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.235937 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-dpsn5" Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.236051 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.236181 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.252759 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-hzc26"] Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.264810 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5t44\" (UniqueName: \"kubernetes.io/projected/a6bb6b9c-b2a6-4908-8da1-c519224d8ef2-kube-api-access-g5t44\") pod \"cinder-8f0c-account-create-update-l4dkr\" (UID: \"a6bb6b9c-b2a6-4908-8da1-c519224d8ef2\") " pod="openstack/cinder-8f0c-account-create-update-l4dkr" Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.265253 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-8f0c-account-create-update-l4dkr" Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.265561 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmg5l\" (UniqueName: \"kubernetes.io/projected/9f46d1fa-3786-4dde-ba6f-d6775782faeb-kube-api-access-fmg5l\") pod \"barbican-db-create-2d25r\" (UID: \"9f46d1fa-3786-4dde-ba6f-d6775782faeb\") " pod="openstack/barbican-db-create-2d25r" Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.320864 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ddtnw\" (UniqueName: \"kubernetes.io/projected/45043340-ef17-4b6c-a6e0-a87eac1942ec-kube-api-access-ddtnw\") pod \"barbican-7946-account-create-update-7t9xc\" (UID: \"45043340-ef17-4b6c-a6e0-a87eac1942ec\") " pod="openstack/barbican-7946-account-create-update-7t9xc" Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.320935 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2856804d-f222-49c2-abc3-153227545c0a-operator-scripts\") pod \"neutron-db-create-fbv4q\" (UID: \"2856804d-f222-49c2-abc3-153227545c0a\") " pod="openstack/neutron-db-create-fbv4q" Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.320966 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/45043340-ef17-4b6c-a6e0-a87eac1942ec-operator-scripts\") pod \"barbican-7946-account-create-update-7t9xc\" (UID: \"45043340-ef17-4b6c-a6e0-a87eac1942ec\") " pod="openstack/barbican-7946-account-create-update-7t9xc" Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.321169 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zjgp\" (UniqueName: \"kubernetes.io/projected/2856804d-f222-49c2-abc3-153227545c0a-kube-api-access-5zjgp\") pod \"neutron-db-create-fbv4q\" (UID: \"2856804d-f222-49c2-abc3-153227545c0a\") " pod="openstack/neutron-db-create-fbv4q" Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.321960 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2856804d-f222-49c2-abc3-153227545c0a-operator-scripts\") pod \"neutron-db-create-fbv4q\" (UID: \"2856804d-f222-49c2-abc3-153227545c0a\") " pod="openstack/neutron-db-create-fbv4q" Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.322501 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/45043340-ef17-4b6c-a6e0-a87eac1942ec-operator-scripts\") pod \"barbican-7946-account-create-update-7t9xc\" (UID: \"45043340-ef17-4b6c-a6e0-a87eac1942ec\") " pod="openstack/barbican-7946-account-create-update-7t9xc" Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.339618 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zjgp\" (UniqueName: \"kubernetes.io/projected/2856804d-f222-49c2-abc3-153227545c0a-kube-api-access-5zjgp\") pod \"neutron-db-create-fbv4q\" (UID: \"2856804d-f222-49c2-abc3-153227545c0a\") " pod="openstack/neutron-db-create-fbv4q" Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.350444 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ddtnw\" (UniqueName: \"kubernetes.io/projected/45043340-ef17-4b6c-a6e0-a87eac1942ec-kube-api-access-ddtnw\") pod \"barbican-7946-account-create-update-7t9xc\" (UID: \"45043340-ef17-4b6c-a6e0-a87eac1942ec\") " pod="openstack/barbican-7946-account-create-update-7t9xc" Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.375563 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-ecdb-account-create-update-52wl7"] Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.376621 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-ecdb-account-create-update-52wl7" Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.379607 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.381836 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-ecdb-account-create-update-52wl7"] Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.384072 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-fbv4q" Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.393007 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-7946-account-create-update-7t9xc" Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.422595 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4q2k\" (UniqueName: \"kubernetes.io/projected/43bc6089-0cfa-478c-8edd-77e667cbb137-kube-api-access-l4q2k\") pod \"keystone-db-sync-hzc26\" (UID: \"43bc6089-0cfa-478c-8edd-77e667cbb137\") " pod="openstack/keystone-db-sync-hzc26" Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.422697 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43bc6089-0cfa-478c-8edd-77e667cbb137-combined-ca-bundle\") pod \"keystone-db-sync-hzc26\" (UID: \"43bc6089-0cfa-478c-8edd-77e667cbb137\") " pod="openstack/keystone-db-sync-hzc26" Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.422896 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43bc6089-0cfa-478c-8edd-77e667cbb137-config-data\") pod \"keystone-db-sync-hzc26\" (UID: \"43bc6089-0cfa-478c-8edd-77e667cbb137\") " pod="openstack/keystone-db-sync-hzc26" Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.525775 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f1b45040-901f-4eab-a55d-46a9842686a8-operator-scripts\") pod \"neutron-ecdb-account-create-update-52wl7\" (UID: \"f1b45040-901f-4eab-a55d-46a9842686a8\") " pod="openstack/neutron-ecdb-account-create-update-52wl7" Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.526082 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43bc6089-0cfa-478c-8edd-77e667cbb137-config-data\") pod \"keystone-db-sync-hzc26\" (UID: \"43bc6089-0cfa-478c-8edd-77e667cbb137\") " pod="openstack/keystone-db-sync-hzc26" Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.526202 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8hld\" (UniqueName: \"kubernetes.io/projected/f1b45040-901f-4eab-a55d-46a9842686a8-kube-api-access-h8hld\") pod \"neutron-ecdb-account-create-update-52wl7\" (UID: \"f1b45040-901f-4eab-a55d-46a9842686a8\") " pod="openstack/neutron-ecdb-account-create-update-52wl7" Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.526241 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4q2k\" (UniqueName: \"kubernetes.io/projected/43bc6089-0cfa-478c-8edd-77e667cbb137-kube-api-access-l4q2k\") pod \"keystone-db-sync-hzc26\" (UID: \"43bc6089-0cfa-478c-8edd-77e667cbb137\") " pod="openstack/keystone-db-sync-hzc26" Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.526313 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43bc6089-0cfa-478c-8edd-77e667cbb137-combined-ca-bundle\") pod \"keystone-db-sync-hzc26\" (UID: \"43bc6089-0cfa-478c-8edd-77e667cbb137\") " pod="openstack/keystone-db-sync-hzc26" Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.530105 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43bc6089-0cfa-478c-8edd-77e667cbb137-config-data\") pod \"keystone-db-sync-hzc26\" (UID: \"43bc6089-0cfa-478c-8edd-77e667cbb137\") " pod="openstack/keystone-db-sync-hzc26" Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.535357 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43bc6089-0cfa-478c-8edd-77e667cbb137-combined-ca-bundle\") pod \"keystone-db-sync-hzc26\" (UID: \"43bc6089-0cfa-478c-8edd-77e667cbb137\") " pod="openstack/keystone-db-sync-hzc26" Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.541881 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-v7c5w"] Feb 18 09:17:33 crc kubenswrapper[4556]: W0218 09:17:33.546905 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda037e8b8_1769_4728_a30c_dcb8671d2e1b.slice/crio-89f6c6ec5230303b0133546ad867d9ac9d26287241c3865ef3597b1e0d47d826 WatchSource:0}: Error finding container 89f6c6ec5230303b0133546ad867d9ac9d26287241c3865ef3597b1e0d47d826: Status 404 returned error can't find the container with id 89f6c6ec5230303b0133546ad867d9ac9d26287241c3865ef3597b1e0d47d826 Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.551572 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4q2k\" (UniqueName: \"kubernetes.io/projected/43bc6089-0cfa-478c-8edd-77e667cbb137-kube-api-access-l4q2k\") pod \"keystone-db-sync-hzc26\" (UID: \"43bc6089-0cfa-478c-8edd-77e667cbb137\") " pod="openstack/keystone-db-sync-hzc26" Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.555882 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-2d25r" Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.607171 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-hzc26" Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.637285 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8hld\" (UniqueName: \"kubernetes.io/projected/f1b45040-901f-4eab-a55d-46a9842686a8-kube-api-access-h8hld\") pod \"neutron-ecdb-account-create-update-52wl7\" (UID: \"f1b45040-901f-4eab-a55d-46a9842686a8\") " pod="openstack/neutron-ecdb-account-create-update-52wl7" Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.637585 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f1b45040-901f-4eab-a55d-46a9842686a8-operator-scripts\") pod \"neutron-ecdb-account-create-update-52wl7\" (UID: \"f1b45040-901f-4eab-a55d-46a9842686a8\") " pod="openstack/neutron-ecdb-account-create-update-52wl7" Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.638867 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f1b45040-901f-4eab-a55d-46a9842686a8-operator-scripts\") pod \"neutron-ecdb-account-create-update-52wl7\" (UID: \"f1b45040-901f-4eab-a55d-46a9842686a8\") " pod="openstack/neutron-ecdb-account-create-update-52wl7" Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.663559 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8hld\" (UniqueName: \"kubernetes.io/projected/f1b45040-901f-4eab-a55d-46a9842686a8-kube-api-access-h8hld\") pod \"neutron-ecdb-account-create-update-52wl7\" (UID: \"f1b45040-901f-4eab-a55d-46a9842686a8\") " pod="openstack/neutron-ecdb-account-create-update-52wl7" Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.694782 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-ecdb-account-create-update-52wl7" Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.721083 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-8f0c-account-create-update-l4dkr"] Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.855327 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-fbv4q"] Feb 18 09:17:33 crc kubenswrapper[4556]: I0218 09:17:33.929692 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-7946-account-create-update-7t9xc"] Feb 18 09:17:34 crc kubenswrapper[4556]: I0218 09:17:34.012198 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-2d25r"] Feb 18 09:17:34 crc kubenswrapper[4556]: W0218 09:17:34.015977 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9f46d1fa_3786_4dde_ba6f_d6775782faeb.slice/crio-d2871b80439715efc68ccc0366756b40d9670943c0babaa0d834fa0086f27b60 WatchSource:0}: Error finding container d2871b80439715efc68ccc0366756b40d9670943c0babaa0d834fa0086f27b60: Status 404 returned error can't find the container with id d2871b80439715efc68ccc0366756b40d9670943c0babaa0d834fa0086f27b60 Feb 18 09:17:34 crc kubenswrapper[4556]: I0218 09:17:34.058303 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-54cc5c87cc-ck95n" Feb 18 09:17:34 crc kubenswrapper[4556]: I0218 09:17:34.146864 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-66b577f8c-9dkbr"] Feb 18 09:17:34 crc kubenswrapper[4556]: I0218 09:17:34.147057 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-66b577f8c-9dkbr" podUID="b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f" containerName="dnsmasq-dns" containerID="cri-o://a8dadc6867286da49a84752ef6856ada290207ac296cb9cfefb9d2498343692e" gracePeriod=10 Feb 18 09:17:34 crc kubenswrapper[4556]: I0218 09:17:34.166797 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-hzc26"] Feb 18 09:17:34 crc kubenswrapper[4556]: I0218 09:17:34.225020 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-ecdb-account-create-update-52wl7"] Feb 18 09:17:34 crc kubenswrapper[4556]: I0218 09:17:34.321715 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-fbv4q" event={"ID":"2856804d-f222-49c2-abc3-153227545c0a","Type":"ContainerStarted","Data":"0b7821217191a992f6f9836f655f75444413ac1a3fcb61632d8577074da62496"} Feb 18 09:17:34 crc kubenswrapper[4556]: I0218 09:17:34.321758 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-fbv4q" event={"ID":"2856804d-f222-49c2-abc3-153227545c0a","Type":"ContainerStarted","Data":"dd7fd6eb2cf63fdb2bd10635d50c8bb5d30f4871c0bb66ea33041e862a2aa6fe"} Feb 18 09:17:34 crc kubenswrapper[4556]: I0218 09:17:34.323603 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-hzc26" event={"ID":"43bc6089-0cfa-478c-8edd-77e667cbb137","Type":"ContainerStarted","Data":"76cb8d213ea0c6c5ae7e7ebfa37e904cda9933377e7c4d7a8f2ff6015b595fe7"} Feb 18 09:17:34 crc kubenswrapper[4556]: I0218 09:17:34.324800 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-8f0c-account-create-update-l4dkr" event={"ID":"a6bb6b9c-b2a6-4908-8da1-c519224d8ef2","Type":"ContainerStarted","Data":"60fc6854fa8cb8ee9bf35bd5ed24e067c93a7990408bc10a44f23a6a41ec457f"} Feb 18 09:17:34 crc kubenswrapper[4556]: I0218 09:17:34.324820 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-8f0c-account-create-update-l4dkr" event={"ID":"a6bb6b9c-b2a6-4908-8da1-c519224d8ef2","Type":"ContainerStarted","Data":"b3c7fc4a02d466c6f126d6c81d9584e2f39c28f9806164504733cb4a4de9ae11"} Feb 18 09:17:34 crc kubenswrapper[4556]: I0218 09:17:34.326033 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-ecdb-account-create-update-52wl7" event={"ID":"f1b45040-901f-4eab-a55d-46a9842686a8","Type":"ContainerStarted","Data":"9b8c124784ea719fc32b7abb78016fa395b0bf5c44f6668440d3fa2dadeef2f1"} Feb 18 09:17:34 crc kubenswrapper[4556]: I0218 09:17:34.338461 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-7946-account-create-update-7t9xc" event={"ID":"45043340-ef17-4b6c-a6e0-a87eac1942ec","Type":"ContainerStarted","Data":"4d34e95a0625ce58e4c834cf1c350945e7d68e2b836629b5f4fbd36bb21af0c5"} Feb 18 09:17:34 crc kubenswrapper[4556]: I0218 09:17:34.338481 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-7946-account-create-update-7t9xc" event={"ID":"45043340-ef17-4b6c-a6e0-a87eac1942ec","Type":"ContainerStarted","Data":"a6a775a7d5bc1f8c61d2e715e19cf27023004e7d09677edd265c6ec28bd68f44"} Feb 18 09:17:34 crc kubenswrapper[4556]: I0218 09:17:34.348621 4556 generic.go:334] "Generic (PLEG): container finished" podID="a037e8b8-1769-4728-a30c-dcb8671d2e1b" containerID="90c9dc724cf9d95bf753513453202e2c030fad7cd2d4ef3b508497249f9a4a7e" exitCode=0 Feb 18 09:17:34 crc kubenswrapper[4556]: I0218 09:17:34.348671 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-v7c5w" event={"ID":"a037e8b8-1769-4728-a30c-dcb8671d2e1b","Type":"ContainerDied","Data":"90c9dc724cf9d95bf753513453202e2c030fad7cd2d4ef3b508497249f9a4a7e"} Feb 18 09:17:34 crc kubenswrapper[4556]: I0218 09:17:34.348686 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-v7c5w" event={"ID":"a037e8b8-1769-4728-a30c-dcb8671d2e1b","Type":"ContainerStarted","Data":"89f6c6ec5230303b0133546ad867d9ac9d26287241c3865ef3597b1e0d47d826"} Feb 18 09:17:34 crc kubenswrapper[4556]: I0218 09:17:34.354762 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-2d25r" event={"ID":"9f46d1fa-3786-4dde-ba6f-d6775782faeb","Type":"ContainerStarted","Data":"d2871b80439715efc68ccc0366756b40d9670943c0babaa0d834fa0086f27b60"} Feb 18 09:17:34 crc kubenswrapper[4556]: I0218 09:17:34.355924 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-fbv4q" podStartSLOduration=1.355913674 podStartE2EDuration="1.355913674s" podCreationTimestamp="2026-02-18 09:17:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:17:34.33887695 +0000 UTC m=+811.355837930" watchObservedRunningTime="2026-02-18 09:17:34.355913674 +0000 UTC m=+811.372874655" Feb 18 09:17:34 crc kubenswrapper[4556]: I0218 09:17:34.368449 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-7946-account-create-update-7t9xc" podStartSLOduration=1.368439275 podStartE2EDuration="1.368439275s" podCreationTimestamp="2026-02-18 09:17:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:17:34.361115412 +0000 UTC m=+811.378076392" watchObservedRunningTime="2026-02-18 09:17:34.368439275 +0000 UTC m=+811.385400255" Feb 18 09:17:34 crc kubenswrapper[4556]: I0218 09:17:34.418200 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-8f0c-account-create-update-l4dkr" podStartSLOduration=2.41818262 podStartE2EDuration="2.41818262s" podCreationTimestamp="2026-02-18 09:17:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:17:34.392308686 +0000 UTC m=+811.409269666" watchObservedRunningTime="2026-02-18 09:17:34.41818262 +0000 UTC m=+811.435143600" Feb 18 09:17:34 crc kubenswrapper[4556]: I0218 09:17:34.802931 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66b577f8c-9dkbr" Feb 18 09:17:34 crc kubenswrapper[4556]: I0218 09:17:34.882718 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f-config\") pod \"b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f\" (UID: \"b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f\") " Feb 18 09:17:34 crc kubenswrapper[4556]: I0218 09:17:34.882780 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6kpfc\" (UniqueName: \"kubernetes.io/projected/b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f-kube-api-access-6kpfc\") pod \"b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f\" (UID: \"b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f\") " Feb 18 09:17:34 crc kubenswrapper[4556]: I0218 09:17:34.882824 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f-dns-svc\") pod \"b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f\" (UID: \"b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f\") " Feb 18 09:17:34 crc kubenswrapper[4556]: I0218 09:17:34.882905 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f-ovsdbserver-sb\") pod \"b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f\" (UID: \"b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f\") " Feb 18 09:17:34 crc kubenswrapper[4556]: I0218 09:17:34.882952 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f-ovsdbserver-nb\") pod \"b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f\" (UID: \"b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f\") " Feb 18 09:17:34 crc kubenswrapper[4556]: I0218 09:17:34.889265 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f-kube-api-access-6kpfc" (OuterVolumeSpecName: "kube-api-access-6kpfc") pod "b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f" (UID: "b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f"). InnerVolumeSpecName "kube-api-access-6kpfc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:17:34 crc kubenswrapper[4556]: I0218 09:17:34.923183 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f" (UID: "b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:17:34 crc kubenswrapper[4556]: I0218 09:17:34.924219 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f-config" (OuterVolumeSpecName: "config") pod "b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f" (UID: "b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:17:34 crc kubenswrapper[4556]: I0218 09:17:34.927881 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f" (UID: "b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:17:34 crc kubenswrapper[4556]: I0218 09:17:34.932627 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f" (UID: "b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:17:34 crc kubenswrapper[4556]: I0218 09:17:34.985251 4556 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f-config\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:34 crc kubenswrapper[4556]: I0218 09:17:34.985273 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6kpfc\" (UniqueName: \"kubernetes.io/projected/b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f-kube-api-access-6kpfc\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:34 crc kubenswrapper[4556]: I0218 09:17:34.985285 4556 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:34 crc kubenswrapper[4556]: I0218 09:17:34.985294 4556 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:34 crc kubenswrapper[4556]: I0218 09:17:34.985314 4556 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:35 crc kubenswrapper[4556]: I0218 09:17:35.363717 4556 generic.go:334] "Generic (PLEG): container finished" podID="2856804d-f222-49c2-abc3-153227545c0a" containerID="0b7821217191a992f6f9836f655f75444413ac1a3fcb61632d8577074da62496" exitCode=0 Feb 18 09:17:35 crc kubenswrapper[4556]: I0218 09:17:35.363802 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-fbv4q" event={"ID":"2856804d-f222-49c2-abc3-153227545c0a","Type":"ContainerDied","Data":"0b7821217191a992f6f9836f655f75444413ac1a3fcb61632d8577074da62496"} Feb 18 09:17:35 crc kubenswrapper[4556]: I0218 09:17:35.365614 4556 generic.go:334] "Generic (PLEG): container finished" podID="a6bb6b9c-b2a6-4908-8da1-c519224d8ef2" containerID="60fc6854fa8cb8ee9bf35bd5ed24e067c93a7990408bc10a44f23a6a41ec457f" exitCode=0 Feb 18 09:17:35 crc kubenswrapper[4556]: I0218 09:17:35.365690 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-8f0c-account-create-update-l4dkr" event={"ID":"a6bb6b9c-b2a6-4908-8da1-c519224d8ef2","Type":"ContainerDied","Data":"60fc6854fa8cb8ee9bf35bd5ed24e067c93a7990408bc10a44f23a6a41ec457f"} Feb 18 09:17:35 crc kubenswrapper[4556]: I0218 09:17:35.367111 4556 generic.go:334] "Generic (PLEG): container finished" podID="f1b45040-901f-4eab-a55d-46a9842686a8" containerID="98c5176b13bae0c1183c67a8105d8142bcb611f8fd37c453ef15b184633083cd" exitCode=0 Feb 18 09:17:35 crc kubenswrapper[4556]: I0218 09:17:35.367180 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-ecdb-account-create-update-52wl7" event={"ID":"f1b45040-901f-4eab-a55d-46a9842686a8","Type":"ContainerDied","Data":"98c5176b13bae0c1183c67a8105d8142bcb611f8fd37c453ef15b184633083cd"} Feb 18 09:17:35 crc kubenswrapper[4556]: I0218 09:17:35.368386 4556 generic.go:334] "Generic (PLEG): container finished" podID="45043340-ef17-4b6c-a6e0-a87eac1942ec" containerID="4d34e95a0625ce58e4c834cf1c350945e7d68e2b836629b5f4fbd36bb21af0c5" exitCode=0 Feb 18 09:17:35 crc kubenswrapper[4556]: I0218 09:17:35.368437 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-7946-account-create-update-7t9xc" event={"ID":"45043340-ef17-4b6c-a6e0-a87eac1942ec","Type":"ContainerDied","Data":"4d34e95a0625ce58e4c834cf1c350945e7d68e2b836629b5f4fbd36bb21af0c5"} Feb 18 09:17:35 crc kubenswrapper[4556]: I0218 09:17:35.370240 4556 generic.go:334] "Generic (PLEG): container finished" podID="b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f" containerID="a8dadc6867286da49a84752ef6856ada290207ac296cb9cfefb9d2498343692e" exitCode=0 Feb 18 09:17:35 crc kubenswrapper[4556]: I0218 09:17:35.370298 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66b577f8c-9dkbr" event={"ID":"b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f","Type":"ContainerDied","Data":"a8dadc6867286da49a84752ef6856ada290207ac296cb9cfefb9d2498343692e"} Feb 18 09:17:35 crc kubenswrapper[4556]: I0218 09:17:35.370316 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66b577f8c-9dkbr" event={"ID":"b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f","Type":"ContainerDied","Data":"87f39b4d6f60421cb6f3ab8da95f31310358b095971c339e77a5eff518431062"} Feb 18 09:17:35 crc kubenswrapper[4556]: I0218 09:17:35.370333 4556 scope.go:117] "RemoveContainer" containerID="a8dadc6867286da49a84752ef6856ada290207ac296cb9cfefb9d2498343692e" Feb 18 09:17:35 crc kubenswrapper[4556]: I0218 09:17:35.370301 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66b577f8c-9dkbr" Feb 18 09:17:35 crc kubenswrapper[4556]: I0218 09:17:35.371574 4556 generic.go:334] "Generic (PLEG): container finished" podID="9f46d1fa-3786-4dde-ba6f-d6775782faeb" containerID="d714dd60903d22c75de48292c1cf000bdfad19de9da8e937a7413be5a38e4e5e" exitCode=0 Feb 18 09:17:35 crc kubenswrapper[4556]: I0218 09:17:35.371675 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-2d25r" event={"ID":"9f46d1fa-3786-4dde-ba6f-d6775782faeb","Type":"ContainerDied","Data":"d714dd60903d22c75de48292c1cf000bdfad19de9da8e937a7413be5a38e4e5e"} Feb 18 09:17:35 crc kubenswrapper[4556]: I0218 09:17:35.385994 4556 scope.go:117] "RemoveContainer" containerID="8f567d29c8442ee45a00a6b7f4c2a7e49859c1fe116422ede582ac534a996100" Feb 18 09:17:35 crc kubenswrapper[4556]: I0218 09:17:35.409529 4556 scope.go:117] "RemoveContainer" containerID="a8dadc6867286da49a84752ef6856ada290207ac296cb9cfefb9d2498343692e" Feb 18 09:17:35 crc kubenswrapper[4556]: E0218 09:17:35.410027 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8dadc6867286da49a84752ef6856ada290207ac296cb9cfefb9d2498343692e\": container with ID starting with a8dadc6867286da49a84752ef6856ada290207ac296cb9cfefb9d2498343692e not found: ID does not exist" containerID="a8dadc6867286da49a84752ef6856ada290207ac296cb9cfefb9d2498343692e" Feb 18 09:17:35 crc kubenswrapper[4556]: I0218 09:17:35.410063 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8dadc6867286da49a84752ef6856ada290207ac296cb9cfefb9d2498343692e"} err="failed to get container status \"a8dadc6867286da49a84752ef6856ada290207ac296cb9cfefb9d2498343692e\": rpc error: code = NotFound desc = could not find container \"a8dadc6867286da49a84752ef6856ada290207ac296cb9cfefb9d2498343692e\": container with ID starting with a8dadc6867286da49a84752ef6856ada290207ac296cb9cfefb9d2498343692e not found: ID does not exist" Feb 18 09:17:35 crc kubenswrapper[4556]: I0218 09:17:35.410090 4556 scope.go:117] "RemoveContainer" containerID="8f567d29c8442ee45a00a6b7f4c2a7e49859c1fe116422ede582ac534a996100" Feb 18 09:17:35 crc kubenswrapper[4556]: E0218 09:17:35.412559 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f567d29c8442ee45a00a6b7f4c2a7e49859c1fe116422ede582ac534a996100\": container with ID starting with 8f567d29c8442ee45a00a6b7f4c2a7e49859c1fe116422ede582ac534a996100 not found: ID does not exist" containerID="8f567d29c8442ee45a00a6b7f4c2a7e49859c1fe116422ede582ac534a996100" Feb 18 09:17:35 crc kubenswrapper[4556]: I0218 09:17:35.412590 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f567d29c8442ee45a00a6b7f4c2a7e49859c1fe116422ede582ac534a996100"} err="failed to get container status \"8f567d29c8442ee45a00a6b7f4c2a7e49859c1fe116422ede582ac534a996100\": rpc error: code = NotFound desc = could not find container \"8f567d29c8442ee45a00a6b7f4c2a7e49859c1fe116422ede582ac534a996100\": container with ID starting with 8f567d29c8442ee45a00a6b7f4c2a7e49859c1fe116422ede582ac534a996100 not found: ID does not exist" Feb 18 09:17:35 crc kubenswrapper[4556]: I0218 09:17:35.450834 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-66b577f8c-9dkbr"] Feb 18 09:17:35 crc kubenswrapper[4556]: I0218 09:17:35.465485 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-66b577f8c-9dkbr"] Feb 18 09:17:35 crc kubenswrapper[4556]: I0218 09:17:35.656268 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-v7c5w" Feb 18 09:17:35 crc kubenswrapper[4556]: I0218 09:17:35.698094 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s8mh8\" (UniqueName: \"kubernetes.io/projected/a037e8b8-1769-4728-a30c-dcb8671d2e1b-kube-api-access-s8mh8\") pod \"a037e8b8-1769-4728-a30c-dcb8671d2e1b\" (UID: \"a037e8b8-1769-4728-a30c-dcb8671d2e1b\") " Feb 18 09:17:35 crc kubenswrapper[4556]: I0218 09:17:35.698209 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a037e8b8-1769-4728-a30c-dcb8671d2e1b-operator-scripts\") pod \"a037e8b8-1769-4728-a30c-dcb8671d2e1b\" (UID: \"a037e8b8-1769-4728-a30c-dcb8671d2e1b\") " Feb 18 09:17:35 crc kubenswrapper[4556]: I0218 09:17:35.698812 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a037e8b8-1769-4728-a30c-dcb8671d2e1b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a037e8b8-1769-4728-a30c-dcb8671d2e1b" (UID: "a037e8b8-1769-4728-a30c-dcb8671d2e1b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:17:35 crc kubenswrapper[4556]: I0218 09:17:35.703229 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a037e8b8-1769-4728-a30c-dcb8671d2e1b-kube-api-access-s8mh8" (OuterVolumeSpecName: "kube-api-access-s8mh8") pod "a037e8b8-1769-4728-a30c-dcb8671d2e1b" (UID: "a037e8b8-1769-4728-a30c-dcb8671d2e1b"). InnerVolumeSpecName "kube-api-access-s8mh8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:17:35 crc kubenswrapper[4556]: I0218 09:17:35.800789 4556 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a037e8b8-1769-4728-a30c-dcb8671d2e1b-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:35 crc kubenswrapper[4556]: I0218 09:17:35.800820 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s8mh8\" (UniqueName: \"kubernetes.io/projected/a037e8b8-1769-4728-a30c-dcb8671d2e1b-kube-api-access-s8mh8\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:36 crc kubenswrapper[4556]: I0218 09:17:36.391873 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-v7c5w" event={"ID":"a037e8b8-1769-4728-a30c-dcb8671d2e1b","Type":"ContainerDied","Data":"89f6c6ec5230303b0133546ad867d9ac9d26287241c3865ef3597b1e0d47d826"} Feb 18 09:17:36 crc kubenswrapper[4556]: I0218 09:17:36.391932 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="89f6c6ec5230303b0133546ad867d9ac9d26287241c3865ef3597b1e0d47d826" Feb 18 09:17:36 crc kubenswrapper[4556]: I0218 09:17:36.391893 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-v7c5w" Feb 18 09:17:36 crc kubenswrapper[4556]: I0218 09:17:36.648175 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-2d25r" Feb 18 09:17:36 crc kubenswrapper[4556]: I0218 09:17:36.717299 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fmg5l\" (UniqueName: \"kubernetes.io/projected/9f46d1fa-3786-4dde-ba6f-d6775782faeb-kube-api-access-fmg5l\") pod \"9f46d1fa-3786-4dde-ba6f-d6775782faeb\" (UID: \"9f46d1fa-3786-4dde-ba6f-d6775782faeb\") " Feb 18 09:17:36 crc kubenswrapper[4556]: I0218 09:17:36.717591 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9f46d1fa-3786-4dde-ba6f-d6775782faeb-operator-scripts\") pod \"9f46d1fa-3786-4dde-ba6f-d6775782faeb\" (UID: \"9f46d1fa-3786-4dde-ba6f-d6775782faeb\") " Feb 18 09:17:36 crc kubenswrapper[4556]: I0218 09:17:36.719650 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f46d1fa-3786-4dde-ba6f-d6775782faeb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9f46d1fa-3786-4dde-ba6f-d6775782faeb" (UID: "9f46d1fa-3786-4dde-ba6f-d6775782faeb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:17:36 crc kubenswrapper[4556]: I0218 09:17:36.747352 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f46d1fa-3786-4dde-ba6f-d6775782faeb-kube-api-access-fmg5l" (OuterVolumeSpecName: "kube-api-access-fmg5l") pod "9f46d1fa-3786-4dde-ba6f-d6775782faeb" (UID: "9f46d1fa-3786-4dde-ba6f-d6775782faeb"). InnerVolumeSpecName "kube-api-access-fmg5l". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:17:36 crc kubenswrapper[4556]: I0218 09:17:36.820832 4556 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9f46d1fa-3786-4dde-ba6f-d6775782faeb-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:36 crc kubenswrapper[4556]: I0218 09:17:36.820876 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fmg5l\" (UniqueName: \"kubernetes.io/projected/9f46d1fa-3786-4dde-ba6f-d6775782faeb-kube-api-access-fmg5l\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:37 crc kubenswrapper[4556]: I0218 09:17:37.292741 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f" path="/var/lib/kubelet/pods/b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f/volumes" Feb 18 09:17:37 crc kubenswrapper[4556]: I0218 09:17:37.399549 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-2d25r" Feb 18 09:17:37 crc kubenswrapper[4556]: I0218 09:17:37.399540 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-2d25r" event={"ID":"9f46d1fa-3786-4dde-ba6f-d6775782faeb","Type":"ContainerDied","Data":"d2871b80439715efc68ccc0366756b40d9670943c0babaa0d834fa0086f27b60"} Feb 18 09:17:37 crc kubenswrapper[4556]: I0218 09:17:37.399671 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2871b80439715efc68ccc0366756b40d9670943c0babaa0d834fa0086f27b60" Feb 18 09:17:39 crc kubenswrapper[4556]: I0218 09:17:39.415997 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-fbv4q" event={"ID":"2856804d-f222-49c2-abc3-153227545c0a","Type":"ContainerDied","Data":"dd7fd6eb2cf63fdb2bd10635d50c8bb5d30f4871c0bb66ea33041e862a2aa6fe"} Feb 18 09:17:39 crc kubenswrapper[4556]: I0218 09:17:39.416202 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd7fd6eb2cf63fdb2bd10635d50c8bb5d30f4871c0bb66ea33041e862a2aa6fe" Feb 18 09:17:39 crc kubenswrapper[4556]: I0218 09:17:39.418230 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-ecdb-account-create-update-52wl7" event={"ID":"f1b45040-901f-4eab-a55d-46a9842686a8","Type":"ContainerDied","Data":"9b8c124784ea719fc32b7abb78016fa395b0bf5c44f6668440d3fa2dadeef2f1"} Feb 18 09:17:39 crc kubenswrapper[4556]: I0218 09:17:39.418253 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9b8c124784ea719fc32b7abb78016fa395b0bf5c44f6668440d3fa2dadeef2f1" Feb 18 09:17:39 crc kubenswrapper[4556]: I0218 09:17:39.421001 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-8f0c-account-create-update-l4dkr" event={"ID":"a6bb6b9c-b2a6-4908-8da1-c519224d8ef2","Type":"ContainerDied","Data":"b3c7fc4a02d466c6f126d6c81d9584e2f39c28f9806164504733cb4a4de9ae11"} Feb 18 09:17:39 crc kubenswrapper[4556]: I0218 09:17:39.421020 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b3c7fc4a02d466c6f126d6c81d9584e2f39c28f9806164504733cb4a4de9ae11" Feb 18 09:17:39 crc kubenswrapper[4556]: I0218 09:17:39.422522 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-7946-account-create-update-7t9xc" event={"ID":"45043340-ef17-4b6c-a6e0-a87eac1942ec","Type":"ContainerDied","Data":"a6a775a7d5bc1f8c61d2e715e19cf27023004e7d09677edd265c6ec28bd68f44"} Feb 18 09:17:39 crc kubenswrapper[4556]: I0218 09:17:39.422545 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a6a775a7d5bc1f8c61d2e715e19cf27023004e7d09677edd265c6ec28bd68f44" Feb 18 09:17:39 crc kubenswrapper[4556]: I0218 09:17:39.460885 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-fbv4q" Feb 18 09:17:39 crc kubenswrapper[4556]: I0218 09:17:39.502493 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-8f0c-account-create-update-l4dkr" Feb 18 09:17:39 crc kubenswrapper[4556]: I0218 09:17:39.516727 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-7946-account-create-update-7t9xc" Feb 18 09:17:39 crc kubenswrapper[4556]: I0218 09:17:39.527725 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-ecdb-account-create-update-52wl7" Feb 18 09:17:39 crc kubenswrapper[4556]: I0218 09:17:39.559732 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5zjgp\" (UniqueName: \"kubernetes.io/projected/2856804d-f222-49c2-abc3-153227545c0a-kube-api-access-5zjgp\") pod \"2856804d-f222-49c2-abc3-153227545c0a\" (UID: \"2856804d-f222-49c2-abc3-153227545c0a\") " Feb 18 09:17:39 crc kubenswrapper[4556]: I0218 09:17:39.559787 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a6bb6b9c-b2a6-4908-8da1-c519224d8ef2-operator-scripts\") pod \"a6bb6b9c-b2a6-4908-8da1-c519224d8ef2\" (UID: \"a6bb6b9c-b2a6-4908-8da1-c519224d8ef2\") " Feb 18 09:17:39 crc kubenswrapper[4556]: I0218 09:17:39.559853 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/45043340-ef17-4b6c-a6e0-a87eac1942ec-operator-scripts\") pod \"45043340-ef17-4b6c-a6e0-a87eac1942ec\" (UID: \"45043340-ef17-4b6c-a6e0-a87eac1942ec\") " Feb 18 09:17:39 crc kubenswrapper[4556]: I0218 09:17:39.559873 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ddtnw\" (UniqueName: \"kubernetes.io/projected/45043340-ef17-4b6c-a6e0-a87eac1942ec-kube-api-access-ddtnw\") pod \"45043340-ef17-4b6c-a6e0-a87eac1942ec\" (UID: \"45043340-ef17-4b6c-a6e0-a87eac1942ec\") " Feb 18 09:17:39 crc kubenswrapper[4556]: I0218 09:17:39.559896 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2856804d-f222-49c2-abc3-153227545c0a-operator-scripts\") pod \"2856804d-f222-49c2-abc3-153227545c0a\" (UID: \"2856804d-f222-49c2-abc3-153227545c0a\") " Feb 18 09:17:39 crc kubenswrapper[4556]: I0218 09:17:39.559927 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g5t44\" (UniqueName: \"kubernetes.io/projected/a6bb6b9c-b2a6-4908-8da1-c519224d8ef2-kube-api-access-g5t44\") pod \"a6bb6b9c-b2a6-4908-8da1-c519224d8ef2\" (UID: \"a6bb6b9c-b2a6-4908-8da1-c519224d8ef2\") " Feb 18 09:17:39 crc kubenswrapper[4556]: I0218 09:17:39.560034 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f1b45040-901f-4eab-a55d-46a9842686a8-operator-scripts\") pod \"f1b45040-901f-4eab-a55d-46a9842686a8\" (UID: \"f1b45040-901f-4eab-a55d-46a9842686a8\") " Feb 18 09:17:39 crc kubenswrapper[4556]: I0218 09:17:39.560060 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h8hld\" (UniqueName: \"kubernetes.io/projected/f1b45040-901f-4eab-a55d-46a9842686a8-kube-api-access-h8hld\") pod \"f1b45040-901f-4eab-a55d-46a9842686a8\" (UID: \"f1b45040-901f-4eab-a55d-46a9842686a8\") " Feb 18 09:17:39 crc kubenswrapper[4556]: I0218 09:17:39.560330 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/45043340-ef17-4b6c-a6e0-a87eac1942ec-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "45043340-ef17-4b6c-a6e0-a87eac1942ec" (UID: "45043340-ef17-4b6c-a6e0-a87eac1942ec"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:17:39 crc kubenswrapper[4556]: I0218 09:17:39.560362 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2856804d-f222-49c2-abc3-153227545c0a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2856804d-f222-49c2-abc3-153227545c0a" (UID: "2856804d-f222-49c2-abc3-153227545c0a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:17:39 crc kubenswrapper[4556]: I0218 09:17:39.560444 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f1b45040-901f-4eab-a55d-46a9842686a8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f1b45040-901f-4eab-a55d-46a9842686a8" (UID: "f1b45040-901f-4eab-a55d-46a9842686a8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:17:39 crc kubenswrapper[4556]: I0218 09:17:39.560849 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6bb6b9c-b2a6-4908-8da1-c519224d8ef2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a6bb6b9c-b2a6-4908-8da1-c519224d8ef2" (UID: "a6bb6b9c-b2a6-4908-8da1-c519224d8ef2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:17:39 crc kubenswrapper[4556]: I0218 09:17:39.564617 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2856804d-f222-49c2-abc3-153227545c0a-kube-api-access-5zjgp" (OuterVolumeSpecName: "kube-api-access-5zjgp") pod "2856804d-f222-49c2-abc3-153227545c0a" (UID: "2856804d-f222-49c2-abc3-153227545c0a"). InnerVolumeSpecName "kube-api-access-5zjgp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:17:39 crc kubenswrapper[4556]: I0218 09:17:39.564653 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1b45040-901f-4eab-a55d-46a9842686a8-kube-api-access-h8hld" (OuterVolumeSpecName: "kube-api-access-h8hld") pod "f1b45040-901f-4eab-a55d-46a9842686a8" (UID: "f1b45040-901f-4eab-a55d-46a9842686a8"). InnerVolumeSpecName "kube-api-access-h8hld". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:17:39 crc kubenswrapper[4556]: I0218 09:17:39.565001 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45043340-ef17-4b6c-a6e0-a87eac1942ec-kube-api-access-ddtnw" (OuterVolumeSpecName: "kube-api-access-ddtnw") pod "45043340-ef17-4b6c-a6e0-a87eac1942ec" (UID: "45043340-ef17-4b6c-a6e0-a87eac1942ec"). InnerVolumeSpecName "kube-api-access-ddtnw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:17:39 crc kubenswrapper[4556]: I0218 09:17:39.566945 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6bb6b9c-b2a6-4908-8da1-c519224d8ef2-kube-api-access-g5t44" (OuterVolumeSpecName: "kube-api-access-g5t44") pod "a6bb6b9c-b2a6-4908-8da1-c519224d8ef2" (UID: "a6bb6b9c-b2a6-4908-8da1-c519224d8ef2"). InnerVolumeSpecName "kube-api-access-g5t44". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:17:39 crc kubenswrapper[4556]: I0218 09:17:39.661390 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5zjgp\" (UniqueName: \"kubernetes.io/projected/2856804d-f222-49c2-abc3-153227545c0a-kube-api-access-5zjgp\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:39 crc kubenswrapper[4556]: I0218 09:17:39.661420 4556 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a6bb6b9c-b2a6-4908-8da1-c519224d8ef2-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:39 crc kubenswrapper[4556]: I0218 09:17:39.661430 4556 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/45043340-ef17-4b6c-a6e0-a87eac1942ec-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:39 crc kubenswrapper[4556]: I0218 09:17:39.661439 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ddtnw\" (UniqueName: \"kubernetes.io/projected/45043340-ef17-4b6c-a6e0-a87eac1942ec-kube-api-access-ddtnw\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:39 crc kubenswrapper[4556]: I0218 09:17:39.661447 4556 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2856804d-f222-49c2-abc3-153227545c0a-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:39 crc kubenswrapper[4556]: I0218 09:17:39.661456 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g5t44\" (UniqueName: \"kubernetes.io/projected/a6bb6b9c-b2a6-4908-8da1-c519224d8ef2-kube-api-access-g5t44\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:39 crc kubenswrapper[4556]: I0218 09:17:39.661464 4556 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f1b45040-901f-4eab-a55d-46a9842686a8-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:39 crc kubenswrapper[4556]: I0218 09:17:39.661473 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h8hld\" (UniqueName: \"kubernetes.io/projected/f1b45040-901f-4eab-a55d-46a9842686a8-kube-api-access-h8hld\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:40 crc kubenswrapper[4556]: I0218 09:17:40.429841 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-7946-account-create-update-7t9xc" Feb 18 09:17:40 crc kubenswrapper[4556]: I0218 09:17:40.440384 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-hzc26" event={"ID":"43bc6089-0cfa-478c-8edd-77e667cbb137","Type":"ContainerStarted","Data":"6f8121f2008bdbb40e850334b987b119fb32bea427009c272cd281fb6b8a4735"} Feb 18 09:17:40 crc kubenswrapper[4556]: I0218 09:17:40.440498 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-fbv4q" Feb 18 09:17:40 crc kubenswrapper[4556]: I0218 09:17:40.441427 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-ecdb-account-create-update-52wl7" Feb 18 09:17:40 crc kubenswrapper[4556]: I0218 09:17:40.441440 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-8f0c-account-create-update-l4dkr" Feb 18 09:17:40 crc kubenswrapper[4556]: I0218 09:17:40.497689 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-hzc26" podStartSLOduration=2.277052378 podStartE2EDuration="7.497675257s" podCreationTimestamp="2026-02-18 09:17:33 +0000 UTC" firstStartedPulling="2026-02-18 09:17:34.186669287 +0000 UTC m=+811.203630267" lastFinishedPulling="2026-02-18 09:17:39.407292165 +0000 UTC m=+816.424253146" observedRunningTime="2026-02-18 09:17:40.454368351 +0000 UTC m=+817.471329331" watchObservedRunningTime="2026-02-18 09:17:40.497675257 +0000 UTC m=+817.514636238" Feb 18 09:17:41 crc kubenswrapper[4556]: I0218 09:17:41.437603 4556 generic.go:334] "Generic (PLEG): container finished" podID="43bc6089-0cfa-478c-8edd-77e667cbb137" containerID="6f8121f2008bdbb40e850334b987b119fb32bea427009c272cd281fb6b8a4735" exitCode=0 Feb 18 09:17:41 crc kubenswrapper[4556]: I0218 09:17:41.437638 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-hzc26" event={"ID":"43bc6089-0cfa-478c-8edd-77e667cbb137","Type":"ContainerDied","Data":"6f8121f2008bdbb40e850334b987b119fb32bea427009c272cd281fb6b8a4735"} Feb 18 09:17:42 crc kubenswrapper[4556]: I0218 09:17:42.678962 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-hzc26" Feb 18 09:17:42 crc kubenswrapper[4556]: I0218 09:17:42.715632 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l4q2k\" (UniqueName: \"kubernetes.io/projected/43bc6089-0cfa-478c-8edd-77e667cbb137-kube-api-access-l4q2k\") pod \"43bc6089-0cfa-478c-8edd-77e667cbb137\" (UID: \"43bc6089-0cfa-478c-8edd-77e667cbb137\") " Feb 18 09:17:42 crc kubenswrapper[4556]: I0218 09:17:42.715684 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43bc6089-0cfa-478c-8edd-77e667cbb137-config-data\") pod \"43bc6089-0cfa-478c-8edd-77e667cbb137\" (UID: \"43bc6089-0cfa-478c-8edd-77e667cbb137\") " Feb 18 09:17:42 crc kubenswrapper[4556]: I0218 09:17:42.715771 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43bc6089-0cfa-478c-8edd-77e667cbb137-combined-ca-bundle\") pod \"43bc6089-0cfa-478c-8edd-77e667cbb137\" (UID: \"43bc6089-0cfa-478c-8edd-77e667cbb137\") " Feb 18 09:17:42 crc kubenswrapper[4556]: I0218 09:17:42.721685 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43bc6089-0cfa-478c-8edd-77e667cbb137-kube-api-access-l4q2k" (OuterVolumeSpecName: "kube-api-access-l4q2k") pod "43bc6089-0cfa-478c-8edd-77e667cbb137" (UID: "43bc6089-0cfa-478c-8edd-77e667cbb137"). InnerVolumeSpecName "kube-api-access-l4q2k". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:17:42 crc kubenswrapper[4556]: I0218 09:17:42.733799 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43bc6089-0cfa-478c-8edd-77e667cbb137-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "43bc6089-0cfa-478c-8edd-77e667cbb137" (UID: "43bc6089-0cfa-478c-8edd-77e667cbb137"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:17:42 crc kubenswrapper[4556]: I0218 09:17:42.746293 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43bc6089-0cfa-478c-8edd-77e667cbb137-config-data" (OuterVolumeSpecName: "config-data") pod "43bc6089-0cfa-478c-8edd-77e667cbb137" (UID: "43bc6089-0cfa-478c-8edd-77e667cbb137"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:17:42 crc kubenswrapper[4556]: I0218 09:17:42.817533 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l4q2k\" (UniqueName: \"kubernetes.io/projected/43bc6089-0cfa-478c-8edd-77e667cbb137-kube-api-access-l4q2k\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:42 crc kubenswrapper[4556]: I0218 09:17:42.817681 4556 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43bc6089-0cfa-478c-8edd-77e667cbb137-config-data\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:42 crc kubenswrapper[4556]: I0218 09:17:42.817691 4556 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43bc6089-0cfa-478c-8edd-77e667cbb137-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:43 crc kubenswrapper[4556]: I0218 09:17:43.451315 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-hzc26" event={"ID":"43bc6089-0cfa-478c-8edd-77e667cbb137","Type":"ContainerDied","Data":"76cb8d213ea0c6c5ae7e7ebfa37e904cda9933377e7c4d7a8f2ff6015b595fe7"} Feb 18 09:17:43 crc kubenswrapper[4556]: I0218 09:17:43.451355 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="76cb8d213ea0c6c5ae7e7ebfa37e904cda9933377e7c4d7a8f2ff6015b595fe7" Feb 18 09:17:43 crc kubenswrapper[4556]: I0218 09:17:43.451403 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-hzc26" Feb 18 09:17:43 crc kubenswrapper[4556]: I0218 09:17:43.861973 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-866975879-zxrfh"] Feb 18 09:17:43 crc kubenswrapper[4556]: E0218 09:17:43.865367 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1b45040-901f-4eab-a55d-46a9842686a8" containerName="mariadb-account-create-update" Feb 18 09:17:43 crc kubenswrapper[4556]: I0218 09:17:43.865394 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1b45040-901f-4eab-a55d-46a9842686a8" containerName="mariadb-account-create-update" Feb 18 09:17:43 crc kubenswrapper[4556]: E0218 09:17:43.865405 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43bc6089-0cfa-478c-8edd-77e667cbb137" containerName="keystone-db-sync" Feb 18 09:17:43 crc kubenswrapper[4556]: I0218 09:17:43.865411 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="43bc6089-0cfa-478c-8edd-77e667cbb137" containerName="keystone-db-sync" Feb 18 09:17:43 crc kubenswrapper[4556]: E0218 09:17:43.865425 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f46d1fa-3786-4dde-ba6f-d6775782faeb" containerName="mariadb-database-create" Feb 18 09:17:43 crc kubenswrapper[4556]: I0218 09:17:43.865430 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f46d1fa-3786-4dde-ba6f-d6775782faeb" containerName="mariadb-database-create" Feb 18 09:17:43 crc kubenswrapper[4556]: E0218 09:17:43.865441 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45043340-ef17-4b6c-a6e0-a87eac1942ec" containerName="mariadb-account-create-update" Feb 18 09:17:43 crc kubenswrapper[4556]: I0218 09:17:43.865446 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="45043340-ef17-4b6c-a6e0-a87eac1942ec" containerName="mariadb-account-create-update" Feb 18 09:17:43 crc kubenswrapper[4556]: E0218 09:17:43.865460 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a037e8b8-1769-4728-a30c-dcb8671d2e1b" containerName="mariadb-database-create" Feb 18 09:17:43 crc kubenswrapper[4556]: I0218 09:17:43.865466 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="a037e8b8-1769-4728-a30c-dcb8671d2e1b" containerName="mariadb-database-create" Feb 18 09:17:43 crc kubenswrapper[4556]: E0218 09:17:43.865473 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6bb6b9c-b2a6-4908-8da1-c519224d8ef2" containerName="mariadb-account-create-update" Feb 18 09:17:43 crc kubenswrapper[4556]: I0218 09:17:43.865478 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6bb6b9c-b2a6-4908-8da1-c519224d8ef2" containerName="mariadb-account-create-update" Feb 18 09:17:43 crc kubenswrapper[4556]: E0218 09:17:43.865487 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2856804d-f222-49c2-abc3-153227545c0a" containerName="mariadb-database-create" Feb 18 09:17:43 crc kubenswrapper[4556]: I0218 09:17:43.865492 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="2856804d-f222-49c2-abc3-153227545c0a" containerName="mariadb-database-create" Feb 18 09:17:43 crc kubenswrapper[4556]: E0218 09:17:43.865500 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f" containerName="dnsmasq-dns" Feb 18 09:17:43 crc kubenswrapper[4556]: I0218 09:17:43.865505 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f" containerName="dnsmasq-dns" Feb 18 09:17:43 crc kubenswrapper[4556]: E0218 09:17:43.865521 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f" containerName="init" Feb 18 09:17:43 crc kubenswrapper[4556]: I0218 09:17:43.865526 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f" containerName="init" Feb 18 09:17:43 crc kubenswrapper[4556]: I0218 09:17:43.865780 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="2856804d-f222-49c2-abc3-153227545c0a" containerName="mariadb-database-create" Feb 18 09:17:43 crc kubenswrapper[4556]: I0218 09:17:43.865797 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="a037e8b8-1769-4728-a30c-dcb8671d2e1b" containerName="mariadb-database-create" Feb 18 09:17:43 crc kubenswrapper[4556]: I0218 09:17:43.865803 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f46d1fa-3786-4dde-ba6f-d6775782faeb" containerName="mariadb-database-create" Feb 18 09:17:43 crc kubenswrapper[4556]: I0218 09:17:43.865813 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1b45040-901f-4eab-a55d-46a9842686a8" containerName="mariadb-account-create-update" Feb 18 09:17:43 crc kubenswrapper[4556]: I0218 09:17:43.865823 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="45043340-ef17-4b6c-a6e0-a87eac1942ec" containerName="mariadb-account-create-update" Feb 18 09:17:43 crc kubenswrapper[4556]: I0218 09:17:43.865830 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="b58ee1ba-c499-44bd-88ff-f5f5b7a0b02f" containerName="dnsmasq-dns" Feb 18 09:17:43 crc kubenswrapper[4556]: I0218 09:17:43.865840 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6bb6b9c-b2a6-4908-8da1-c519224d8ef2" containerName="mariadb-account-create-update" Feb 18 09:17:43 crc kubenswrapper[4556]: I0218 09:17:43.865846 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="43bc6089-0cfa-478c-8edd-77e667cbb137" containerName="keystone-db-sync" Feb 18 09:17:43 crc kubenswrapper[4556]: I0218 09:17:43.866711 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-866975879-zxrfh" Feb 18 09:17:43 crc kubenswrapper[4556]: I0218 09:17:43.877861 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-866975879-zxrfh"] Feb 18 09:17:43 crc kubenswrapper[4556]: I0218 09:17:43.908594 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-hgqj7"] Feb 18 09:17:43 crc kubenswrapper[4556]: I0218 09:17:43.909596 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-hgqj7" Feb 18 09:17:43 crc kubenswrapper[4556]: I0218 09:17:43.911795 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 18 09:17:43 crc kubenswrapper[4556]: I0218 09:17:43.911873 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Feb 18 09:17:43 crc kubenswrapper[4556]: I0218 09:17:43.912202 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-dpsn5" Feb 18 09:17:43 crc kubenswrapper[4556]: I0218 09:17:43.912360 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 18 09:17:43 crc kubenswrapper[4556]: I0218 09:17:43.912522 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 18 09:17:43 crc kubenswrapper[4556]: I0218 09:17:43.927187 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-hgqj7"] Feb 18 09:17:43 crc kubenswrapper[4556]: I0218 09:17:43.935513 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77955a10-5158-461d-a0fa-702e45c6eb87-config\") pod \"dnsmasq-dns-866975879-zxrfh\" (UID: \"77955a10-5158-461d-a0fa-702e45c6eb87\") " pod="openstack/dnsmasq-dns-866975879-zxrfh" Feb 18 09:17:43 crc kubenswrapper[4556]: I0218 09:17:43.935594 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/77955a10-5158-461d-a0fa-702e45c6eb87-ovsdbserver-sb\") pod \"dnsmasq-dns-866975879-zxrfh\" (UID: \"77955a10-5158-461d-a0fa-702e45c6eb87\") " pod="openstack/dnsmasq-dns-866975879-zxrfh" Feb 18 09:17:43 crc kubenswrapper[4556]: I0218 09:17:43.935612 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/77955a10-5158-461d-a0fa-702e45c6eb87-dns-swift-storage-0\") pod \"dnsmasq-dns-866975879-zxrfh\" (UID: \"77955a10-5158-461d-a0fa-702e45c6eb87\") " pod="openstack/dnsmasq-dns-866975879-zxrfh" Feb 18 09:17:43 crc kubenswrapper[4556]: I0218 09:17:43.935628 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/77955a10-5158-461d-a0fa-702e45c6eb87-ovsdbserver-nb\") pod \"dnsmasq-dns-866975879-zxrfh\" (UID: \"77955a10-5158-461d-a0fa-702e45c6eb87\") " pod="openstack/dnsmasq-dns-866975879-zxrfh" Feb 18 09:17:43 crc kubenswrapper[4556]: I0218 09:17:43.935708 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/77955a10-5158-461d-a0fa-702e45c6eb87-dns-svc\") pod \"dnsmasq-dns-866975879-zxrfh\" (UID: \"77955a10-5158-461d-a0fa-702e45c6eb87\") " pod="openstack/dnsmasq-dns-866975879-zxrfh" Feb 18 09:17:43 crc kubenswrapper[4556]: I0218 09:17:43.935812 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ksgt5\" (UniqueName: \"kubernetes.io/projected/77955a10-5158-461d-a0fa-702e45c6eb87-kube-api-access-ksgt5\") pod \"dnsmasq-dns-866975879-zxrfh\" (UID: \"77955a10-5158-461d-a0fa-702e45c6eb87\") " pod="openstack/dnsmasq-dns-866975879-zxrfh" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.037366 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/12c97a89-6f28-4eab-add6-448279050305-fernet-keys\") pod \"keystone-bootstrap-hgqj7\" (UID: \"12c97a89-6f28-4eab-add6-448279050305\") " pod="openstack/keystone-bootstrap-hgqj7" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.037406 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12c97a89-6f28-4eab-add6-448279050305-config-data\") pod \"keystone-bootstrap-hgqj7\" (UID: \"12c97a89-6f28-4eab-add6-448279050305\") " pod="openstack/keystone-bootstrap-hgqj7" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.037454 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ksgt5\" (UniqueName: \"kubernetes.io/projected/77955a10-5158-461d-a0fa-702e45c6eb87-kube-api-access-ksgt5\") pod \"dnsmasq-dns-866975879-zxrfh\" (UID: \"77955a10-5158-461d-a0fa-702e45c6eb87\") " pod="openstack/dnsmasq-dns-866975879-zxrfh" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.037474 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rwkn\" (UniqueName: \"kubernetes.io/projected/12c97a89-6f28-4eab-add6-448279050305-kube-api-access-6rwkn\") pod \"keystone-bootstrap-hgqj7\" (UID: \"12c97a89-6f28-4eab-add6-448279050305\") " pod="openstack/keystone-bootstrap-hgqj7" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.037491 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12c97a89-6f28-4eab-add6-448279050305-scripts\") pod \"keystone-bootstrap-hgqj7\" (UID: \"12c97a89-6f28-4eab-add6-448279050305\") " pod="openstack/keystone-bootstrap-hgqj7" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.037544 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77955a10-5158-461d-a0fa-702e45c6eb87-config\") pod \"dnsmasq-dns-866975879-zxrfh\" (UID: \"77955a10-5158-461d-a0fa-702e45c6eb87\") " pod="openstack/dnsmasq-dns-866975879-zxrfh" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.037581 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/77955a10-5158-461d-a0fa-702e45c6eb87-ovsdbserver-sb\") pod \"dnsmasq-dns-866975879-zxrfh\" (UID: \"77955a10-5158-461d-a0fa-702e45c6eb87\") " pod="openstack/dnsmasq-dns-866975879-zxrfh" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.037594 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12c97a89-6f28-4eab-add6-448279050305-combined-ca-bundle\") pod \"keystone-bootstrap-hgqj7\" (UID: \"12c97a89-6f28-4eab-add6-448279050305\") " pod="openstack/keystone-bootstrap-hgqj7" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.037613 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/77955a10-5158-461d-a0fa-702e45c6eb87-dns-swift-storage-0\") pod \"dnsmasq-dns-866975879-zxrfh\" (UID: \"77955a10-5158-461d-a0fa-702e45c6eb87\") " pod="openstack/dnsmasq-dns-866975879-zxrfh" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.037631 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/77955a10-5158-461d-a0fa-702e45c6eb87-ovsdbserver-nb\") pod \"dnsmasq-dns-866975879-zxrfh\" (UID: \"77955a10-5158-461d-a0fa-702e45c6eb87\") " pod="openstack/dnsmasq-dns-866975879-zxrfh" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.037684 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/77955a10-5158-461d-a0fa-702e45c6eb87-dns-svc\") pod \"dnsmasq-dns-866975879-zxrfh\" (UID: \"77955a10-5158-461d-a0fa-702e45c6eb87\") " pod="openstack/dnsmasq-dns-866975879-zxrfh" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.037710 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/12c97a89-6f28-4eab-add6-448279050305-credential-keys\") pod \"keystone-bootstrap-hgqj7\" (UID: \"12c97a89-6f28-4eab-add6-448279050305\") " pod="openstack/keystone-bootstrap-hgqj7" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.038898 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77955a10-5158-461d-a0fa-702e45c6eb87-config\") pod \"dnsmasq-dns-866975879-zxrfh\" (UID: \"77955a10-5158-461d-a0fa-702e45c6eb87\") " pod="openstack/dnsmasq-dns-866975879-zxrfh" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.039458 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/77955a10-5158-461d-a0fa-702e45c6eb87-ovsdbserver-sb\") pod \"dnsmasq-dns-866975879-zxrfh\" (UID: \"77955a10-5158-461d-a0fa-702e45c6eb87\") " pod="openstack/dnsmasq-dns-866975879-zxrfh" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.040014 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/77955a10-5158-461d-a0fa-702e45c6eb87-dns-swift-storage-0\") pod \"dnsmasq-dns-866975879-zxrfh\" (UID: \"77955a10-5158-461d-a0fa-702e45c6eb87\") " pod="openstack/dnsmasq-dns-866975879-zxrfh" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.040515 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/77955a10-5158-461d-a0fa-702e45c6eb87-ovsdbserver-nb\") pod \"dnsmasq-dns-866975879-zxrfh\" (UID: \"77955a10-5158-461d-a0fa-702e45c6eb87\") " pod="openstack/dnsmasq-dns-866975879-zxrfh" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.040997 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/77955a10-5158-461d-a0fa-702e45c6eb87-dns-svc\") pod \"dnsmasq-dns-866975879-zxrfh\" (UID: \"77955a10-5158-461d-a0fa-702e45c6eb87\") " pod="openstack/dnsmasq-dns-866975879-zxrfh" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.053809 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-77cf5996cc-dj8s4"] Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.054995 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-77cf5996cc-dj8s4" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.059428 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.059585 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.059605 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.059617 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-5jdpm" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.070013 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ksgt5\" (UniqueName: \"kubernetes.io/projected/77955a10-5158-461d-a0fa-702e45c6eb87-kube-api-access-ksgt5\") pod \"dnsmasq-dns-866975879-zxrfh\" (UID: \"77955a10-5158-461d-a0fa-702e45c6eb87\") " pod="openstack/dnsmasq-dns-866975879-zxrfh" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.086335 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-77cf5996cc-dj8s4"] Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.105116 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.106810 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.113843 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.114038 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.130202 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-dtq7t"] Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.131209 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-dtq7t" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.136308 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.141185 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.141370 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.142428 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2plk8\" (UniqueName: \"kubernetes.io/projected/66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98-kube-api-access-2plk8\") pod \"horizon-77cf5996cc-dj8s4\" (UID: \"66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98\") " pod="openstack/horizon-77cf5996cc-dj8s4" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.142479 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98-logs\") pod \"horizon-77cf5996cc-dj8s4\" (UID: \"66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98\") " pod="openstack/horizon-77cf5996cc-dj8s4" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.142498 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98-scripts\") pod \"horizon-77cf5996cc-dj8s4\" (UID: \"66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98\") " pod="openstack/horizon-77cf5996cc-dj8s4" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.142550 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12c97a89-6f28-4eab-add6-448279050305-combined-ca-bundle\") pod \"keystone-bootstrap-hgqj7\" (UID: \"12c97a89-6f28-4eab-add6-448279050305\") " pod="openstack/keystone-bootstrap-hgqj7" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.142624 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/12c97a89-6f28-4eab-add6-448279050305-credential-keys\") pod \"keystone-bootstrap-hgqj7\" (UID: \"12c97a89-6f28-4eab-add6-448279050305\") " pod="openstack/keystone-bootstrap-hgqj7" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.142666 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/12c97a89-6f28-4eab-add6-448279050305-fernet-keys\") pod \"keystone-bootstrap-hgqj7\" (UID: \"12c97a89-6f28-4eab-add6-448279050305\") " pod="openstack/keystone-bootstrap-hgqj7" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.142685 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12c97a89-6f28-4eab-add6-448279050305-config-data\") pod \"keystone-bootstrap-hgqj7\" (UID: \"12c97a89-6f28-4eab-add6-448279050305\") " pod="openstack/keystone-bootstrap-hgqj7" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.142714 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98-config-data\") pod \"horizon-77cf5996cc-dj8s4\" (UID: \"66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98\") " pod="openstack/horizon-77cf5996cc-dj8s4" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.142733 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98-horizon-secret-key\") pod \"horizon-77cf5996cc-dj8s4\" (UID: \"66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98\") " pod="openstack/horizon-77cf5996cc-dj8s4" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.142752 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rwkn\" (UniqueName: \"kubernetes.io/projected/12c97a89-6f28-4eab-add6-448279050305-kube-api-access-6rwkn\") pod \"keystone-bootstrap-hgqj7\" (UID: \"12c97a89-6f28-4eab-add6-448279050305\") " pod="openstack/keystone-bootstrap-hgqj7" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.142776 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12c97a89-6f28-4eab-add6-448279050305-scripts\") pod \"keystone-bootstrap-hgqj7\" (UID: \"12c97a89-6f28-4eab-add6-448279050305\") " pod="openstack/keystone-bootstrap-hgqj7" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.143369 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-kz8d4" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.146968 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/12c97a89-6f28-4eab-add6-448279050305-credential-keys\") pod \"keystone-bootstrap-hgqj7\" (UID: \"12c97a89-6f28-4eab-add6-448279050305\") " pod="openstack/keystone-bootstrap-hgqj7" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.149785 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/12c97a89-6f28-4eab-add6-448279050305-fernet-keys\") pod \"keystone-bootstrap-hgqj7\" (UID: \"12c97a89-6f28-4eab-add6-448279050305\") " pod="openstack/keystone-bootstrap-hgqj7" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.150454 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12c97a89-6f28-4eab-add6-448279050305-scripts\") pod \"keystone-bootstrap-hgqj7\" (UID: \"12c97a89-6f28-4eab-add6-448279050305\") " pod="openstack/keystone-bootstrap-hgqj7" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.151028 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-dtq7t"] Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.153679 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12c97a89-6f28-4eab-add6-448279050305-config-data\") pod \"keystone-bootstrap-hgqj7\" (UID: \"12c97a89-6f28-4eab-add6-448279050305\") " pod="openstack/keystone-bootstrap-hgqj7" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.167583 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12c97a89-6f28-4eab-add6-448279050305-combined-ca-bundle\") pod \"keystone-bootstrap-hgqj7\" (UID: \"12c97a89-6f28-4eab-add6-448279050305\") " pod="openstack/keystone-bootstrap-hgqj7" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.186025 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-866975879-zxrfh" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.207307 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rwkn\" (UniqueName: \"kubernetes.io/projected/12c97a89-6f28-4eab-add6-448279050305-kube-api-access-6rwkn\") pod \"keystone-bootstrap-hgqj7\" (UID: \"12c97a89-6f28-4eab-add6-448279050305\") " pod="openstack/keystone-bootstrap-hgqj7" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.227825 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-hgqj7" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.243328 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-4k7vg"] Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.244991 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-4k7vg" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.246836 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0739f72-891b-46e1-bce5-e17c54a407ab-scripts\") pod \"ceilometer-0\" (UID: \"e0739f72-891b-46e1-bce5-e17c54a407ab\") " pod="openstack/ceilometer-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.246926 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0739f72-891b-46e1-bce5-e17c54a407ab-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e0739f72-891b-46e1-bce5-e17c54a407ab\") " pod="openstack/ceilometer-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.247016 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e0739f72-891b-46e1-bce5-e17c54a407ab-run-httpd\") pod \"ceilometer-0\" (UID: \"e0739f72-891b-46e1-bce5-e17c54a407ab\") " pod="openstack/ceilometer-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.247077 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e0739f72-891b-46e1-bce5-e17c54a407ab-log-httpd\") pod \"ceilometer-0\" (UID: \"e0739f72-891b-46e1-bce5-e17c54a407ab\") " pod="openstack/ceilometer-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.251390 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98-config-data\") pod \"horizon-77cf5996cc-dj8s4\" (UID: \"66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98\") " pod="openstack/horizon-77cf5996cc-dj8s4" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.251476 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98-horizon-secret-key\") pod \"horizon-77cf5996cc-dj8s4\" (UID: \"66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98\") " pod="openstack/horizon-77cf5996cc-dj8s4" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.251546 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb11270a-c754-44b2-b0b1-937cff515ddd-config-data\") pod \"cinder-db-sync-dtq7t\" (UID: \"cb11270a-c754-44b2-b0b1-937cff515ddd\") " pod="openstack/cinder-db-sync-dtq7t" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.251649 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2plk8\" (UniqueName: \"kubernetes.io/projected/66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98-kube-api-access-2plk8\") pod \"horizon-77cf5996cc-dj8s4\" (UID: \"66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98\") " pod="openstack/horizon-77cf5996cc-dj8s4" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.251741 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb11270a-c754-44b2-b0b1-937cff515ddd-combined-ca-bundle\") pod \"cinder-db-sync-dtq7t\" (UID: \"cb11270a-c754-44b2-b0b1-937cff515ddd\") " pod="openstack/cinder-db-sync-dtq7t" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.251821 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98-logs\") pod \"horizon-77cf5996cc-dj8s4\" (UID: \"66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98\") " pod="openstack/horizon-77cf5996cc-dj8s4" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.251885 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98-scripts\") pod \"horizon-77cf5996cc-dj8s4\" (UID: \"66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98\") " pod="openstack/horizon-77cf5996cc-dj8s4" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.251950 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb11270a-c754-44b2-b0b1-937cff515ddd-scripts\") pod \"cinder-db-sync-dtq7t\" (UID: \"cb11270a-c754-44b2-b0b1-937cff515ddd\") " pod="openstack/cinder-db-sync-dtq7t" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.252043 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e0739f72-891b-46e1-bce5-e17c54a407ab-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e0739f72-891b-46e1-bce5-e17c54a407ab\") " pod="openstack/ceilometer-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.252126 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cb11270a-c754-44b2-b0b1-937cff515ddd-etc-machine-id\") pod \"cinder-db-sync-dtq7t\" (UID: \"cb11270a-c754-44b2-b0b1-937cff515ddd\") " pod="openstack/cinder-db-sync-dtq7t" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.252208 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46vwv\" (UniqueName: \"kubernetes.io/projected/cb11270a-c754-44b2-b0b1-937cff515ddd-kube-api-access-46vwv\") pod \"cinder-db-sync-dtq7t\" (UID: \"cb11270a-c754-44b2-b0b1-937cff515ddd\") " pod="openstack/cinder-db-sync-dtq7t" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.252353 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0739f72-891b-46e1-bce5-e17c54a407ab-config-data\") pod \"ceilometer-0\" (UID: \"e0739f72-891b-46e1-bce5-e17c54a407ab\") " pod="openstack/ceilometer-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.252440 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c985q\" (UniqueName: \"kubernetes.io/projected/e0739f72-891b-46e1-bce5-e17c54a407ab-kube-api-access-c985q\") pod \"ceilometer-0\" (UID: \"e0739f72-891b-46e1-bce5-e17c54a407ab\") " pod="openstack/ceilometer-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.252502 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cb11270a-c754-44b2-b0b1-937cff515ddd-db-sync-config-data\") pod \"cinder-db-sync-dtq7t\" (UID: \"cb11270a-c754-44b2-b0b1-937cff515ddd\") " pod="openstack/cinder-db-sync-dtq7t" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.253594 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98-config-data\") pod \"horizon-77cf5996cc-dj8s4\" (UID: \"66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98\") " pod="openstack/horizon-77cf5996cc-dj8s4" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.256625 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.256815 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-bcc55" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.256945 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.258404 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98-logs\") pod \"horizon-77cf5996cc-dj8s4\" (UID: \"66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98\") " pod="openstack/horizon-77cf5996cc-dj8s4" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.258814 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98-scripts\") pod \"horizon-77cf5996cc-dj8s4\" (UID: \"66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98\") " pod="openstack/horizon-77cf5996cc-dj8s4" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.258959 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98-horizon-secret-key\") pod \"horizon-77cf5996cc-dj8s4\" (UID: \"66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98\") " pod="openstack/horizon-77cf5996cc-dj8s4" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.270342 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-xhbg8"] Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.271351 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-xhbg8" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.280920 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.281082 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.281528 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-wgbkj" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.284711 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-74c5b6b6c5-q8djs"] Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.288429 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-74c5b6b6c5-q8djs" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.295258 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-4k7vg"] Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.303697 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2plk8\" (UniqueName: \"kubernetes.io/projected/66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98-kube-api-access-2plk8\") pod \"horizon-77cf5996cc-dj8s4\" (UID: \"66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98\") " pod="openstack/horizon-77cf5996cc-dj8s4" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.309933 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-xhbg8"] Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.346922 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-74c5b6b6c5-q8djs"] Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.354177 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e0739f72-891b-46e1-bce5-e17c54a407ab-log-httpd\") pod \"ceilometer-0\" (UID: \"e0739f72-891b-46e1-bce5-e17c54a407ab\") " pod="openstack/ceilometer-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.354243 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50228011-f908-465e-a754-d4b6b626dda9-logs\") pod \"horizon-74c5b6b6c5-q8djs\" (UID: \"50228011-f908-465e-a754-d4b6b626dda9\") " pod="openstack/horizon-74c5b6b6c5-q8djs" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.354279 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b111dd2-11a6-47cc-ba9e-2886186c676b-combined-ca-bundle\") pod \"placement-db-sync-xhbg8\" (UID: \"2b111dd2-11a6-47cc-ba9e-2886186c676b\") " pod="openstack/placement-db-sync-xhbg8" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.354300 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b111dd2-11a6-47cc-ba9e-2886186c676b-logs\") pod \"placement-db-sync-xhbg8\" (UID: \"2b111dd2-11a6-47cc-ba9e-2886186c676b\") " pod="openstack/placement-db-sync-xhbg8" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.354330 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bf90aa4-605f-4b7d-abde-2a1f3236a6af-combined-ca-bundle\") pod \"neutron-db-sync-4k7vg\" (UID: \"6bf90aa4-605f-4b7d-abde-2a1f3236a6af\") " pod="openstack/neutron-db-sync-4k7vg" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.354354 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb11270a-c754-44b2-b0b1-937cff515ddd-config-data\") pod \"cinder-db-sync-dtq7t\" (UID: \"cb11270a-c754-44b2-b0b1-937cff515ddd\") " pod="openstack/cinder-db-sync-dtq7t" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.354376 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6bf90aa4-605f-4b7d-abde-2a1f3236a6af-config\") pod \"neutron-db-sync-4k7vg\" (UID: \"6bf90aa4-605f-4b7d-abde-2a1f3236a6af\") " pod="openstack/neutron-db-sync-4k7vg" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.354447 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8mr2\" (UniqueName: \"kubernetes.io/projected/50228011-f908-465e-a754-d4b6b626dda9-kube-api-access-b8mr2\") pod \"horizon-74c5b6b6c5-q8djs\" (UID: \"50228011-f908-465e-a754-d4b6b626dda9\") " pod="openstack/horizon-74c5b6b6c5-q8djs" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.354467 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb11270a-c754-44b2-b0b1-937cff515ddd-combined-ca-bundle\") pod \"cinder-db-sync-dtq7t\" (UID: \"cb11270a-c754-44b2-b0b1-937cff515ddd\") " pod="openstack/cinder-db-sync-dtq7t" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.354492 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb11270a-c754-44b2-b0b1-937cff515ddd-scripts\") pod \"cinder-db-sync-dtq7t\" (UID: \"cb11270a-c754-44b2-b0b1-937cff515ddd\") " pod="openstack/cinder-db-sync-dtq7t" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.354521 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/50228011-f908-465e-a754-d4b6b626dda9-horizon-secret-key\") pod \"horizon-74c5b6b6c5-q8djs\" (UID: \"50228011-f908-465e-a754-d4b6b626dda9\") " pod="openstack/horizon-74c5b6b6c5-q8djs" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.354548 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e0739f72-891b-46e1-bce5-e17c54a407ab-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e0739f72-891b-46e1-bce5-e17c54a407ab\") " pod="openstack/ceilometer-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.354584 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cb11270a-c754-44b2-b0b1-937cff515ddd-etc-machine-id\") pod \"cinder-db-sync-dtq7t\" (UID: \"cb11270a-c754-44b2-b0b1-937cff515ddd\") " pod="openstack/cinder-db-sync-dtq7t" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.354599 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b111dd2-11a6-47cc-ba9e-2886186c676b-scripts\") pod \"placement-db-sync-xhbg8\" (UID: \"2b111dd2-11a6-47cc-ba9e-2886186c676b\") " pod="openstack/placement-db-sync-xhbg8" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.354622 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46vwv\" (UniqueName: \"kubernetes.io/projected/cb11270a-c754-44b2-b0b1-937cff515ddd-kube-api-access-46vwv\") pod \"cinder-db-sync-dtq7t\" (UID: \"cb11270a-c754-44b2-b0b1-937cff515ddd\") " pod="openstack/cinder-db-sync-dtq7t" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.354642 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmcgv\" (UniqueName: \"kubernetes.io/projected/6bf90aa4-605f-4b7d-abde-2a1f3236a6af-kube-api-access-hmcgv\") pod \"neutron-db-sync-4k7vg\" (UID: \"6bf90aa4-605f-4b7d-abde-2a1f3236a6af\") " pod="openstack/neutron-db-sync-4k7vg" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.354666 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0739f72-891b-46e1-bce5-e17c54a407ab-config-data\") pod \"ceilometer-0\" (UID: \"e0739f72-891b-46e1-bce5-e17c54a407ab\") " pod="openstack/ceilometer-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.354695 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/50228011-f908-465e-a754-d4b6b626dda9-scripts\") pod \"horizon-74c5b6b6c5-q8djs\" (UID: \"50228011-f908-465e-a754-d4b6b626dda9\") " pod="openstack/horizon-74c5b6b6c5-q8djs" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.354712 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b111dd2-11a6-47cc-ba9e-2886186c676b-config-data\") pod \"placement-db-sync-xhbg8\" (UID: \"2b111dd2-11a6-47cc-ba9e-2886186c676b\") " pod="openstack/placement-db-sync-xhbg8" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.354734 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c985q\" (UniqueName: \"kubernetes.io/projected/e0739f72-891b-46e1-bce5-e17c54a407ab-kube-api-access-c985q\") pod \"ceilometer-0\" (UID: \"e0739f72-891b-46e1-bce5-e17c54a407ab\") " pod="openstack/ceilometer-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.354751 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cb11270a-c754-44b2-b0b1-937cff515ddd-db-sync-config-data\") pod \"cinder-db-sync-dtq7t\" (UID: \"cb11270a-c754-44b2-b0b1-937cff515ddd\") " pod="openstack/cinder-db-sync-dtq7t" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.354790 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/50228011-f908-465e-a754-d4b6b626dda9-config-data\") pod \"horizon-74c5b6b6c5-q8djs\" (UID: \"50228011-f908-465e-a754-d4b6b626dda9\") " pod="openstack/horizon-74c5b6b6c5-q8djs" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.354814 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0739f72-891b-46e1-bce5-e17c54a407ab-scripts\") pod \"ceilometer-0\" (UID: \"e0739f72-891b-46e1-bce5-e17c54a407ab\") " pod="openstack/ceilometer-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.354840 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0739f72-891b-46e1-bce5-e17c54a407ab-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e0739f72-891b-46e1-bce5-e17c54a407ab\") " pod="openstack/ceilometer-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.354878 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5dj6\" (UniqueName: \"kubernetes.io/projected/2b111dd2-11a6-47cc-ba9e-2886186c676b-kube-api-access-x5dj6\") pod \"placement-db-sync-xhbg8\" (UID: \"2b111dd2-11a6-47cc-ba9e-2886186c676b\") " pod="openstack/placement-db-sync-xhbg8" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.354902 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e0739f72-891b-46e1-bce5-e17c54a407ab-run-httpd\") pod \"ceilometer-0\" (UID: \"e0739f72-891b-46e1-bce5-e17c54a407ab\") " pod="openstack/ceilometer-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.361663 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cb11270a-c754-44b2-b0b1-937cff515ddd-etc-machine-id\") pod \"cinder-db-sync-dtq7t\" (UID: \"cb11270a-c754-44b2-b0b1-937cff515ddd\") " pod="openstack/cinder-db-sync-dtq7t" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.362209 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e0739f72-891b-46e1-bce5-e17c54a407ab-log-httpd\") pod \"ceilometer-0\" (UID: \"e0739f72-891b-46e1-bce5-e17c54a407ab\") " pod="openstack/ceilometer-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.380392 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb11270a-c754-44b2-b0b1-937cff515ddd-config-data\") pod \"cinder-db-sync-dtq7t\" (UID: \"cb11270a-c754-44b2-b0b1-937cff515ddd\") " pod="openstack/cinder-db-sync-dtq7t" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.381638 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb11270a-c754-44b2-b0b1-937cff515ddd-combined-ca-bundle\") pod \"cinder-db-sync-dtq7t\" (UID: \"cb11270a-c754-44b2-b0b1-937cff515ddd\") " pod="openstack/cinder-db-sync-dtq7t" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.382457 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0739f72-891b-46e1-bce5-e17c54a407ab-scripts\") pod \"ceilometer-0\" (UID: \"e0739f72-891b-46e1-bce5-e17c54a407ab\") " pod="openstack/ceilometer-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.383722 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e0739f72-891b-46e1-bce5-e17c54a407ab-run-httpd\") pod \"ceilometer-0\" (UID: \"e0739f72-891b-46e1-bce5-e17c54a407ab\") " pod="openstack/ceilometer-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.390227 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e0739f72-891b-46e1-bce5-e17c54a407ab-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e0739f72-891b-46e1-bce5-e17c54a407ab\") " pod="openstack/ceilometer-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.391022 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cb11270a-c754-44b2-b0b1-937cff515ddd-db-sync-config-data\") pod \"cinder-db-sync-dtq7t\" (UID: \"cb11270a-c754-44b2-b0b1-937cff515ddd\") " pod="openstack/cinder-db-sync-dtq7t" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.394568 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb11270a-c754-44b2-b0b1-937cff515ddd-scripts\") pod \"cinder-db-sync-dtq7t\" (UID: \"cb11270a-c754-44b2-b0b1-937cff515ddd\") " pod="openstack/cinder-db-sync-dtq7t" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.398195 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0739f72-891b-46e1-bce5-e17c54a407ab-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e0739f72-891b-46e1-bce5-e17c54a407ab\") " pod="openstack/ceilometer-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.414849 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-77cf5996cc-dj8s4" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.417864 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0739f72-891b-46e1-bce5-e17c54a407ab-config-data\") pod \"ceilometer-0\" (UID: \"e0739f72-891b-46e1-bce5-e17c54a407ab\") " pod="openstack/ceilometer-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.433498 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.454183 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.455144 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c985q\" (UniqueName: \"kubernetes.io/projected/e0739f72-891b-46e1-bce5-e17c54a407ab-kube-api-access-c985q\") pod \"ceilometer-0\" (UID: \"e0739f72-891b-46e1-bce5-e17c54a407ab\") " pod="openstack/ceilometer-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.457034 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6bf90aa4-605f-4b7d-abde-2a1f3236a6af-config\") pod \"neutron-db-sync-4k7vg\" (UID: \"6bf90aa4-605f-4b7d-abde-2a1f3236a6af\") " pod="openstack/neutron-db-sync-4k7vg" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.457121 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8mr2\" (UniqueName: \"kubernetes.io/projected/50228011-f908-465e-a754-d4b6b626dda9-kube-api-access-b8mr2\") pod \"horizon-74c5b6b6c5-q8djs\" (UID: \"50228011-f908-465e-a754-d4b6b626dda9\") " pod="openstack/horizon-74c5b6b6c5-q8djs" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.457217 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/50228011-f908-465e-a754-d4b6b626dda9-horizon-secret-key\") pod \"horizon-74c5b6b6c5-q8djs\" (UID: \"50228011-f908-465e-a754-d4b6b626dda9\") " pod="openstack/horizon-74c5b6b6c5-q8djs" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.457278 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b111dd2-11a6-47cc-ba9e-2886186c676b-scripts\") pod \"placement-db-sync-xhbg8\" (UID: \"2b111dd2-11a6-47cc-ba9e-2886186c676b\") " pod="openstack/placement-db-sync-xhbg8" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.457321 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmcgv\" (UniqueName: \"kubernetes.io/projected/6bf90aa4-605f-4b7d-abde-2a1f3236a6af-kube-api-access-hmcgv\") pod \"neutron-db-sync-4k7vg\" (UID: \"6bf90aa4-605f-4b7d-abde-2a1f3236a6af\") " pod="openstack/neutron-db-sync-4k7vg" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.457372 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/50228011-f908-465e-a754-d4b6b626dda9-scripts\") pod \"horizon-74c5b6b6c5-q8djs\" (UID: \"50228011-f908-465e-a754-d4b6b626dda9\") " pod="openstack/horizon-74c5b6b6c5-q8djs" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.457388 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b111dd2-11a6-47cc-ba9e-2886186c676b-config-data\") pod \"placement-db-sync-xhbg8\" (UID: \"2b111dd2-11a6-47cc-ba9e-2886186c676b\") " pod="openstack/placement-db-sync-xhbg8" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.457416 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/50228011-f908-465e-a754-d4b6b626dda9-config-data\") pod \"horizon-74c5b6b6c5-q8djs\" (UID: \"50228011-f908-465e-a754-d4b6b626dda9\") " pod="openstack/horizon-74c5b6b6c5-q8djs" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.457472 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5dj6\" (UniqueName: \"kubernetes.io/projected/2b111dd2-11a6-47cc-ba9e-2886186c676b-kube-api-access-x5dj6\") pod \"placement-db-sync-xhbg8\" (UID: \"2b111dd2-11a6-47cc-ba9e-2886186c676b\") " pod="openstack/placement-db-sync-xhbg8" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.457772 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50228011-f908-465e-a754-d4b6b626dda9-logs\") pod \"horizon-74c5b6b6c5-q8djs\" (UID: \"50228011-f908-465e-a754-d4b6b626dda9\") " pod="openstack/horizon-74c5b6b6c5-q8djs" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.457808 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b111dd2-11a6-47cc-ba9e-2886186c676b-combined-ca-bundle\") pod \"placement-db-sync-xhbg8\" (UID: \"2b111dd2-11a6-47cc-ba9e-2886186c676b\") " pod="openstack/placement-db-sync-xhbg8" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.457832 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b111dd2-11a6-47cc-ba9e-2886186c676b-logs\") pod \"placement-db-sync-xhbg8\" (UID: \"2b111dd2-11a6-47cc-ba9e-2886186c676b\") " pod="openstack/placement-db-sync-xhbg8" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.457871 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bf90aa4-605f-4b7d-abde-2a1f3236a6af-combined-ca-bundle\") pod \"neutron-db-sync-4k7vg\" (UID: \"6bf90aa4-605f-4b7d-abde-2a1f3236a6af\") " pod="openstack/neutron-db-sync-4k7vg" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.460981 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bf90aa4-605f-4b7d-abde-2a1f3236a6af-combined-ca-bundle\") pod \"neutron-db-sync-4k7vg\" (UID: \"6bf90aa4-605f-4b7d-abde-2a1f3236a6af\") " pod="openstack/neutron-db-sync-4k7vg" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.464503 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/50228011-f908-465e-a754-d4b6b626dda9-config-data\") pod \"horizon-74c5b6b6c5-q8djs\" (UID: \"50228011-f908-465e-a754-d4b6b626dda9\") " pod="openstack/horizon-74c5b6b6c5-q8djs" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.464875 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50228011-f908-465e-a754-d4b6b626dda9-logs\") pod \"horizon-74c5b6b6c5-q8djs\" (UID: \"50228011-f908-465e-a754-d4b6b626dda9\") " pod="openstack/horizon-74c5b6b6c5-q8djs" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.465132 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/50228011-f908-465e-a754-d4b6b626dda9-scripts\") pod \"horizon-74c5b6b6c5-q8djs\" (UID: \"50228011-f908-465e-a754-d4b6b626dda9\") " pod="openstack/horizon-74c5b6b6c5-q8djs" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.467326 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b111dd2-11a6-47cc-ba9e-2886186c676b-combined-ca-bundle\") pod \"placement-db-sync-xhbg8\" (UID: \"2b111dd2-11a6-47cc-ba9e-2886186c676b\") " pod="openstack/placement-db-sync-xhbg8" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.467596 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b111dd2-11a6-47cc-ba9e-2886186c676b-logs\") pod \"placement-db-sync-xhbg8\" (UID: \"2b111dd2-11a6-47cc-ba9e-2886186c676b\") " pod="openstack/placement-db-sync-xhbg8" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.468059 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b111dd2-11a6-47cc-ba9e-2886186c676b-config-data\") pod \"placement-db-sync-xhbg8\" (UID: \"2b111dd2-11a6-47cc-ba9e-2886186c676b\") " pod="openstack/placement-db-sync-xhbg8" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.471719 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/6bf90aa4-605f-4b7d-abde-2a1f3236a6af-config\") pod \"neutron-db-sync-4k7vg\" (UID: \"6bf90aa4-605f-4b7d-abde-2a1f3236a6af\") " pod="openstack/neutron-db-sync-4k7vg" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.472073 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46vwv\" (UniqueName: \"kubernetes.io/projected/cb11270a-c754-44b2-b0b1-937cff515ddd-kube-api-access-46vwv\") pod \"cinder-db-sync-dtq7t\" (UID: \"cb11270a-c754-44b2-b0b1-937cff515ddd\") " pod="openstack/cinder-db-sync-dtq7t" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.476677 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.476857 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-7flzx" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.477099 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.477567 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.480467 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/50228011-f908-465e-a754-d4b6b626dda9-horizon-secret-key\") pod \"horizon-74c5b6b6c5-q8djs\" (UID: \"50228011-f908-465e-a754-d4b6b626dda9\") " pod="openstack/horizon-74c5b6b6c5-q8djs" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.487643 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b111dd2-11a6-47cc-ba9e-2886186c676b-scripts\") pod \"placement-db-sync-xhbg8\" (UID: \"2b111dd2-11a6-47cc-ba9e-2886186c676b\") " pod="openstack/placement-db-sync-xhbg8" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.502380 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-866975879-zxrfh"] Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.520084 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5dj6\" (UniqueName: \"kubernetes.io/projected/2b111dd2-11a6-47cc-ba9e-2886186c676b-kube-api-access-x5dj6\") pod \"placement-db-sync-xhbg8\" (UID: \"2b111dd2-11a6-47cc-ba9e-2886186c676b\") " pod="openstack/placement-db-sync-xhbg8" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.522613 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8mr2\" (UniqueName: \"kubernetes.io/projected/50228011-f908-465e-a754-d4b6b626dda9-kube-api-access-b8mr2\") pod \"horizon-74c5b6b6c5-q8djs\" (UID: \"50228011-f908-465e-a754-d4b6b626dda9\") " pod="openstack/horizon-74c5b6b6c5-q8djs" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.523070 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmcgv\" (UniqueName: \"kubernetes.io/projected/6bf90aa4-605f-4b7d-abde-2a1f3236a6af-kube-api-access-hmcgv\") pod \"neutron-db-sync-4k7vg\" (UID: \"6bf90aa4-605f-4b7d-abde-2a1f3236a6af\") " pod="openstack/neutron-db-sync-4k7vg" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.560394 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.561670 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6cb6fa8-2338-4694-a50b-9f7ce0073e07-config-data\") pod \"glance-default-external-api-0\" (UID: \"d6cb6fa8-2338-4694-a50b-9f7ce0073e07\") " pod="openstack/glance-default-external-api-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.569505 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6cb6fa8-2338-4694-a50b-9f7ce0073e07-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d6cb6fa8-2338-4694-a50b-9f7ce0073e07\") " pod="openstack/glance-default-external-api-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.569648 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qw7lk\" (UniqueName: \"kubernetes.io/projected/d6cb6fa8-2338-4694-a50b-9f7ce0073e07-kube-api-access-qw7lk\") pod \"glance-default-external-api-0\" (UID: \"d6cb6fa8-2338-4694-a50b-9f7ce0073e07\") " pod="openstack/glance-default-external-api-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.569799 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6cb6fa8-2338-4694-a50b-9f7ce0073e07-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d6cb6fa8-2338-4694-a50b-9f7ce0073e07\") " pod="openstack/glance-default-external-api-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.569911 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d6cb6fa8-2338-4694-a50b-9f7ce0073e07-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d6cb6fa8-2338-4694-a50b-9f7ce0073e07\") " pod="openstack/glance-default-external-api-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.570012 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"d6cb6fa8-2338-4694-a50b-9f7ce0073e07\") " pod="openstack/glance-default-external-api-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.570096 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6cb6fa8-2338-4694-a50b-9f7ce0073e07-scripts\") pod \"glance-default-external-api-0\" (UID: \"d6cb6fa8-2338-4694-a50b-9f7ce0073e07\") " pod="openstack/glance-default-external-api-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.570262 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6cb6fa8-2338-4694-a50b-9f7ce0073e07-logs\") pod \"glance-default-external-api-0\" (UID: \"d6cb6fa8-2338-4694-a50b-9f7ce0073e07\") " pod="openstack/glance-default-external-api-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.608627 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-dtq7t" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.623194 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-f58898b5f-r69gj"] Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.630193 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-4k7vg" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.630459 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f58898b5f-r69gj" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.634973 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-r4h4k"] Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.636420 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-r4h4k" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.639788 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-xbqch" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.640003 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.642405 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f58898b5f-r69gj"] Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.650381 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-xhbg8" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.654308 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-r4h4k"] Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.671236 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.672979 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8c46304b-1b60-493b-918f-678451f13057-db-sync-config-data\") pod \"barbican-db-sync-r4h4k\" (UID: \"8c46304b-1b60-493b-918f-678451f13057\") " pod="openstack/barbican-db-sync-r4h4k" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.673021 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/840ae99a-2a91-4c3e-9fab-b05787d3b97b-ovsdbserver-nb\") pod \"dnsmasq-dns-f58898b5f-r69gj\" (UID: \"840ae99a-2a91-4c3e-9fab-b05787d3b97b\") " pod="openstack/dnsmasq-dns-f58898b5f-r69gj" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.673060 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qw7lk\" (UniqueName: \"kubernetes.io/projected/d6cb6fa8-2338-4694-a50b-9f7ce0073e07-kube-api-access-qw7lk\") pod \"glance-default-external-api-0\" (UID: \"d6cb6fa8-2338-4694-a50b-9f7ce0073e07\") " pod="openstack/glance-default-external-api-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.673121 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6cb6fa8-2338-4694-a50b-9f7ce0073e07-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d6cb6fa8-2338-4694-a50b-9f7ce0073e07\") " pod="openstack/glance-default-external-api-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.673168 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/840ae99a-2a91-4c3e-9fab-b05787d3b97b-config\") pod \"dnsmasq-dns-f58898b5f-r69gj\" (UID: \"840ae99a-2a91-4c3e-9fab-b05787d3b97b\") " pod="openstack/dnsmasq-dns-f58898b5f-r69gj" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.673186 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/840ae99a-2a91-4c3e-9fab-b05787d3b97b-ovsdbserver-sb\") pod \"dnsmasq-dns-f58898b5f-r69gj\" (UID: \"840ae99a-2a91-4c3e-9fab-b05787d3b97b\") " pod="openstack/dnsmasq-dns-f58898b5f-r69gj" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.673204 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d6cb6fa8-2338-4694-a50b-9f7ce0073e07-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d6cb6fa8-2338-4694-a50b-9f7ce0073e07\") " pod="openstack/glance-default-external-api-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.673227 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/840ae99a-2a91-4c3e-9fab-b05787d3b97b-dns-svc\") pod \"dnsmasq-dns-f58898b5f-r69gj\" (UID: \"840ae99a-2a91-4c3e-9fab-b05787d3b97b\") " pod="openstack/dnsmasq-dns-f58898b5f-r69gj" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.673251 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mss5k\" (UniqueName: \"kubernetes.io/projected/8c46304b-1b60-493b-918f-678451f13057-kube-api-access-mss5k\") pod \"barbican-db-sync-r4h4k\" (UID: \"8c46304b-1b60-493b-918f-678451f13057\") " pod="openstack/barbican-db-sync-r4h4k" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.673273 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"d6cb6fa8-2338-4694-a50b-9f7ce0073e07\") " pod="openstack/glance-default-external-api-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.673306 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6cb6fa8-2338-4694-a50b-9f7ce0073e07-scripts\") pod \"glance-default-external-api-0\" (UID: \"d6cb6fa8-2338-4694-a50b-9f7ce0073e07\") " pod="openstack/glance-default-external-api-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.673323 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c46304b-1b60-493b-918f-678451f13057-combined-ca-bundle\") pod \"barbican-db-sync-r4h4k\" (UID: \"8c46304b-1b60-493b-918f-678451f13057\") " pod="openstack/barbican-db-sync-r4h4k" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.673374 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/840ae99a-2a91-4c3e-9fab-b05787d3b97b-dns-swift-storage-0\") pod \"dnsmasq-dns-f58898b5f-r69gj\" (UID: \"840ae99a-2a91-4c3e-9fab-b05787d3b97b\") " pod="openstack/dnsmasq-dns-f58898b5f-r69gj" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.673388 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nnrt\" (UniqueName: \"kubernetes.io/projected/840ae99a-2a91-4c3e-9fab-b05787d3b97b-kube-api-access-5nnrt\") pod \"dnsmasq-dns-f58898b5f-r69gj\" (UID: \"840ae99a-2a91-4c3e-9fab-b05787d3b97b\") " pod="openstack/dnsmasq-dns-f58898b5f-r69gj" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.673403 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6cb6fa8-2338-4694-a50b-9f7ce0073e07-logs\") pod \"glance-default-external-api-0\" (UID: \"d6cb6fa8-2338-4694-a50b-9f7ce0073e07\") " pod="openstack/glance-default-external-api-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.673418 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.673455 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6cb6fa8-2338-4694-a50b-9f7ce0073e07-config-data\") pod \"glance-default-external-api-0\" (UID: \"d6cb6fa8-2338-4694-a50b-9f7ce0073e07\") " pod="openstack/glance-default-external-api-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.673489 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6cb6fa8-2338-4694-a50b-9f7ce0073e07-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d6cb6fa8-2338-4694-a50b-9f7ce0073e07\") " pod="openstack/glance-default-external-api-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.674399 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d6cb6fa8-2338-4694-a50b-9f7ce0073e07-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d6cb6fa8-2338-4694-a50b-9f7ce0073e07\") " pod="openstack/glance-default-external-api-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.674422 4556 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"d6cb6fa8-2338-4694-a50b-9f7ce0073e07\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-external-api-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.675818 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.675975 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.676490 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6cb6fa8-2338-4694-a50b-9f7ce0073e07-logs\") pod \"glance-default-external-api-0\" (UID: \"d6cb6fa8-2338-4694-a50b-9f7ce0073e07\") " pod="openstack/glance-default-external-api-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.677921 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6cb6fa8-2338-4694-a50b-9f7ce0073e07-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d6cb6fa8-2338-4694-a50b-9f7ce0073e07\") " pod="openstack/glance-default-external-api-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.682670 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6cb6fa8-2338-4694-a50b-9f7ce0073e07-config-data\") pod \"glance-default-external-api-0\" (UID: \"d6cb6fa8-2338-4694-a50b-9f7ce0073e07\") " pod="openstack/glance-default-external-api-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.684104 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6cb6fa8-2338-4694-a50b-9f7ce0073e07-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d6cb6fa8-2338-4694-a50b-9f7ce0073e07\") " pod="openstack/glance-default-external-api-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.684644 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6cb6fa8-2338-4694-a50b-9f7ce0073e07-scripts\") pod \"glance-default-external-api-0\" (UID: \"d6cb6fa8-2338-4694-a50b-9f7ce0073e07\") " pod="openstack/glance-default-external-api-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.689920 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.692694 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-74c5b6b6c5-q8djs" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.705470 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"d6cb6fa8-2338-4694-a50b-9f7ce0073e07\") " pod="openstack/glance-default-external-api-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.705800 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qw7lk\" (UniqueName: \"kubernetes.io/projected/d6cb6fa8-2338-4694-a50b-9f7ce0073e07-kube-api-access-qw7lk\") pod \"glance-default-external-api-0\" (UID: \"d6cb6fa8-2338-4694-a50b-9f7ce0073e07\") " pod="openstack/glance-default-external-api-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.722058 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-866975879-zxrfh"] Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.723520 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.776456 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ee4e9c9-306f-4499-a53e-77cd55f1d28e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5ee4e9c9-306f-4499-a53e-77cd55f1d28e\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.776810 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ee4e9c9-306f-4499-a53e-77cd55f1d28e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5ee4e9c9-306f-4499-a53e-77cd55f1d28e\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.776849 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8c46304b-1b60-493b-918f-678451f13057-db-sync-config-data\") pod \"barbican-db-sync-r4h4k\" (UID: \"8c46304b-1b60-493b-918f-678451f13057\") " pod="openstack/barbican-db-sync-r4h4k" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.776868 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5ee4e9c9-306f-4499-a53e-77cd55f1d28e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5ee4e9c9-306f-4499-a53e-77cd55f1d28e\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.776906 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/840ae99a-2a91-4c3e-9fab-b05787d3b97b-ovsdbserver-nb\") pod \"dnsmasq-dns-f58898b5f-r69gj\" (UID: \"840ae99a-2a91-4c3e-9fab-b05787d3b97b\") " pod="openstack/dnsmasq-dns-f58898b5f-r69gj" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.776925 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"5ee4e9c9-306f-4499-a53e-77cd55f1d28e\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.776977 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ee4e9c9-306f-4499-a53e-77cd55f1d28e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5ee4e9c9-306f-4499-a53e-77cd55f1d28e\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.777001 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ee4e9c9-306f-4499-a53e-77cd55f1d28e-logs\") pod \"glance-default-internal-api-0\" (UID: \"5ee4e9c9-306f-4499-a53e-77cd55f1d28e\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.777065 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgmvt\" (UniqueName: \"kubernetes.io/projected/5ee4e9c9-306f-4499-a53e-77cd55f1d28e-kube-api-access-qgmvt\") pod \"glance-default-internal-api-0\" (UID: \"5ee4e9c9-306f-4499-a53e-77cd55f1d28e\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.777091 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/840ae99a-2a91-4c3e-9fab-b05787d3b97b-config\") pod \"dnsmasq-dns-f58898b5f-r69gj\" (UID: \"840ae99a-2a91-4c3e-9fab-b05787d3b97b\") " pod="openstack/dnsmasq-dns-f58898b5f-r69gj" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.777123 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/840ae99a-2a91-4c3e-9fab-b05787d3b97b-ovsdbserver-sb\") pod \"dnsmasq-dns-f58898b5f-r69gj\" (UID: \"840ae99a-2a91-4c3e-9fab-b05787d3b97b\") " pod="openstack/dnsmasq-dns-f58898b5f-r69gj" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.777301 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/840ae99a-2a91-4c3e-9fab-b05787d3b97b-dns-svc\") pod \"dnsmasq-dns-f58898b5f-r69gj\" (UID: \"840ae99a-2a91-4c3e-9fab-b05787d3b97b\") " pod="openstack/dnsmasq-dns-f58898b5f-r69gj" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.777332 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mss5k\" (UniqueName: \"kubernetes.io/projected/8c46304b-1b60-493b-918f-678451f13057-kube-api-access-mss5k\") pod \"barbican-db-sync-r4h4k\" (UID: \"8c46304b-1b60-493b-918f-678451f13057\") " pod="openstack/barbican-db-sync-r4h4k" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.777378 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ee4e9c9-306f-4499-a53e-77cd55f1d28e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5ee4e9c9-306f-4499-a53e-77cd55f1d28e\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.777407 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c46304b-1b60-493b-918f-678451f13057-combined-ca-bundle\") pod \"barbican-db-sync-r4h4k\" (UID: \"8c46304b-1b60-493b-918f-678451f13057\") " pod="openstack/barbican-db-sync-r4h4k" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.777486 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/840ae99a-2a91-4c3e-9fab-b05787d3b97b-dns-swift-storage-0\") pod \"dnsmasq-dns-f58898b5f-r69gj\" (UID: \"840ae99a-2a91-4c3e-9fab-b05787d3b97b\") " pod="openstack/dnsmasq-dns-f58898b5f-r69gj" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.777500 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nnrt\" (UniqueName: \"kubernetes.io/projected/840ae99a-2a91-4c3e-9fab-b05787d3b97b-kube-api-access-5nnrt\") pod \"dnsmasq-dns-f58898b5f-r69gj\" (UID: \"840ae99a-2a91-4c3e-9fab-b05787d3b97b\") " pod="openstack/dnsmasq-dns-f58898b5f-r69gj" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.779564 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/840ae99a-2a91-4c3e-9fab-b05787d3b97b-config\") pod \"dnsmasq-dns-f58898b5f-r69gj\" (UID: \"840ae99a-2a91-4c3e-9fab-b05787d3b97b\") " pod="openstack/dnsmasq-dns-f58898b5f-r69gj" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.780200 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/840ae99a-2a91-4c3e-9fab-b05787d3b97b-ovsdbserver-nb\") pod \"dnsmasq-dns-f58898b5f-r69gj\" (UID: \"840ae99a-2a91-4c3e-9fab-b05787d3b97b\") " pod="openstack/dnsmasq-dns-f58898b5f-r69gj" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.780829 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/840ae99a-2a91-4c3e-9fab-b05787d3b97b-dns-swift-storage-0\") pod \"dnsmasq-dns-f58898b5f-r69gj\" (UID: \"840ae99a-2a91-4c3e-9fab-b05787d3b97b\") " pod="openstack/dnsmasq-dns-f58898b5f-r69gj" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.780940 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/840ae99a-2a91-4c3e-9fab-b05787d3b97b-dns-svc\") pod \"dnsmasq-dns-f58898b5f-r69gj\" (UID: \"840ae99a-2a91-4c3e-9fab-b05787d3b97b\") " pod="openstack/dnsmasq-dns-f58898b5f-r69gj" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.780984 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/840ae99a-2a91-4c3e-9fab-b05787d3b97b-ovsdbserver-sb\") pod \"dnsmasq-dns-f58898b5f-r69gj\" (UID: \"840ae99a-2a91-4c3e-9fab-b05787d3b97b\") " pod="openstack/dnsmasq-dns-f58898b5f-r69gj" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.787519 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.788101 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8c46304b-1b60-493b-918f-678451f13057-db-sync-config-data\") pod \"barbican-db-sync-r4h4k\" (UID: \"8c46304b-1b60-493b-918f-678451f13057\") " pod="openstack/barbican-db-sync-r4h4k" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.789366 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c46304b-1b60-493b-918f-678451f13057-combined-ca-bundle\") pod \"barbican-db-sync-r4h4k\" (UID: \"8c46304b-1b60-493b-918f-678451f13057\") " pod="openstack/barbican-db-sync-r4h4k" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.796822 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nnrt\" (UniqueName: \"kubernetes.io/projected/840ae99a-2a91-4c3e-9fab-b05787d3b97b-kube-api-access-5nnrt\") pod \"dnsmasq-dns-f58898b5f-r69gj\" (UID: \"840ae99a-2a91-4c3e-9fab-b05787d3b97b\") " pod="openstack/dnsmasq-dns-f58898b5f-r69gj" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.798953 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mss5k\" (UniqueName: \"kubernetes.io/projected/8c46304b-1b60-493b-918f-678451f13057-kube-api-access-mss5k\") pod \"barbican-db-sync-r4h4k\" (UID: \"8c46304b-1b60-493b-918f-678451f13057\") " pod="openstack/barbican-db-sync-r4h4k" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.878965 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ee4e9c9-306f-4499-a53e-77cd55f1d28e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5ee4e9c9-306f-4499-a53e-77cd55f1d28e\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.879025 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ee4e9c9-306f-4499-a53e-77cd55f1d28e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5ee4e9c9-306f-4499-a53e-77cd55f1d28e\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.879054 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5ee4e9c9-306f-4499-a53e-77cd55f1d28e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5ee4e9c9-306f-4499-a53e-77cd55f1d28e\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.879079 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"5ee4e9c9-306f-4499-a53e-77cd55f1d28e\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.879112 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ee4e9c9-306f-4499-a53e-77cd55f1d28e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5ee4e9c9-306f-4499-a53e-77cd55f1d28e\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.879131 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ee4e9c9-306f-4499-a53e-77cd55f1d28e-logs\") pod \"glance-default-internal-api-0\" (UID: \"5ee4e9c9-306f-4499-a53e-77cd55f1d28e\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.879196 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgmvt\" (UniqueName: \"kubernetes.io/projected/5ee4e9c9-306f-4499-a53e-77cd55f1d28e-kube-api-access-qgmvt\") pod \"glance-default-internal-api-0\" (UID: \"5ee4e9c9-306f-4499-a53e-77cd55f1d28e\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.879251 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ee4e9c9-306f-4499-a53e-77cd55f1d28e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5ee4e9c9-306f-4499-a53e-77cd55f1d28e\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.880231 4556 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"5ee4e9c9-306f-4499-a53e-77cd55f1d28e\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.884141 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ee4e9c9-306f-4499-a53e-77cd55f1d28e-logs\") pod \"glance-default-internal-api-0\" (UID: \"5ee4e9c9-306f-4499-a53e-77cd55f1d28e\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.884825 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5ee4e9c9-306f-4499-a53e-77cd55f1d28e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5ee4e9c9-306f-4499-a53e-77cd55f1d28e\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.887319 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ee4e9c9-306f-4499-a53e-77cd55f1d28e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5ee4e9c9-306f-4499-a53e-77cd55f1d28e\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.887810 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ee4e9c9-306f-4499-a53e-77cd55f1d28e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5ee4e9c9-306f-4499-a53e-77cd55f1d28e\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.889685 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ee4e9c9-306f-4499-a53e-77cd55f1d28e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5ee4e9c9-306f-4499-a53e-77cd55f1d28e\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.892309 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ee4e9c9-306f-4499-a53e-77cd55f1d28e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5ee4e9c9-306f-4499-a53e-77cd55f1d28e\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.903774 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgmvt\" (UniqueName: \"kubernetes.io/projected/5ee4e9c9-306f-4499-a53e-77cd55f1d28e-kube-api-access-qgmvt\") pod \"glance-default-internal-api-0\" (UID: \"5ee4e9c9-306f-4499-a53e-77cd55f1d28e\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.917820 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"5ee4e9c9-306f-4499-a53e-77cd55f1d28e\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.959618 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f58898b5f-r69gj" Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.969147 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-77cf5996cc-dj8s4"] Feb 18 09:17:44 crc kubenswrapper[4556]: I0218 09:17:44.986662 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-r4h4k" Feb 18 09:17:45 crc kubenswrapper[4556]: I0218 09:17:45.002870 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 18 09:17:45 crc kubenswrapper[4556]: I0218 09:17:45.080257 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-hgqj7"] Feb 18 09:17:45 crc kubenswrapper[4556]: I0218 09:17:45.219748 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-4k7vg"] Feb 18 09:17:45 crc kubenswrapper[4556]: W0218 09:17:45.226050 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6bf90aa4_605f_4b7d_abde_2a1f3236a6af.slice/crio-cfd07d686eaf20681a09ee7a134c96b1944015bcecbe62d365f5d03747a5fa25 WatchSource:0}: Error finding container cfd07d686eaf20681a09ee7a134c96b1944015bcecbe62d365f5d03747a5fa25: Status 404 returned error can't find the container with id cfd07d686eaf20681a09ee7a134c96b1944015bcecbe62d365f5d03747a5fa25 Feb 18 09:17:45 crc kubenswrapper[4556]: I0218 09:17:45.353671 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-74c5b6b6c5-q8djs"] Feb 18 09:17:45 crc kubenswrapper[4556]: I0218 09:17:45.361274 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 18 09:17:45 crc kubenswrapper[4556]: I0218 09:17:45.376271 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-xhbg8"] Feb 18 09:17:45 crc kubenswrapper[4556]: I0218 09:17:45.406340 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-dtq7t"] Feb 18 09:17:45 crc kubenswrapper[4556]: I0218 09:17:45.517893 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-74c5b6b6c5-q8djs" event={"ID":"50228011-f908-465e-a754-d4b6b626dda9","Type":"ContainerStarted","Data":"d7824966b854b4e05a496fcf7987339d08df598f934fcdf6099cd480c21cd503"} Feb 18 09:17:45 crc kubenswrapper[4556]: I0218 09:17:45.519386 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-4k7vg" event={"ID":"6bf90aa4-605f-4b7d-abde-2a1f3236a6af","Type":"ContainerStarted","Data":"63868b1e8917bd5b78ec577b9090971cf503941ce12fce83e038cadfaa8c7368"} Feb 18 09:17:45 crc kubenswrapper[4556]: I0218 09:17:45.519444 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-4k7vg" event={"ID":"6bf90aa4-605f-4b7d-abde-2a1f3236a6af","Type":"ContainerStarted","Data":"cfd07d686eaf20681a09ee7a134c96b1944015bcecbe62d365f5d03747a5fa25"} Feb 18 09:17:45 crc kubenswrapper[4556]: I0218 09:17:45.521464 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-dtq7t" event={"ID":"cb11270a-c754-44b2-b0b1-937cff515ddd","Type":"ContainerStarted","Data":"95f6e23c65f73eadf9ca8c1d78cda62edfcc71e1c4a5eeee591612e6bd856108"} Feb 18 09:17:45 crc kubenswrapper[4556]: I0218 09:17:45.522414 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e0739f72-891b-46e1-bce5-e17c54a407ab","Type":"ContainerStarted","Data":"f429b65a29cc2a73cc0988982cefba8e5a956982006946c9fcfc145106a29642"} Feb 18 09:17:45 crc kubenswrapper[4556]: I0218 09:17:45.523180 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-xhbg8" event={"ID":"2b111dd2-11a6-47cc-ba9e-2886186c676b","Type":"ContainerStarted","Data":"1f75efa9811d0d1c1df4aa765d0b73d8685adb22da219fb07a779b3fd46403bb"} Feb 18 09:17:45 crc kubenswrapper[4556]: I0218 09:17:45.524370 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-77cf5996cc-dj8s4" event={"ID":"66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98","Type":"ContainerStarted","Data":"0f49d8a03148f93f5cf7a0e584b50845c6d3288e42f4a093af888c34aea04a84"} Feb 18 09:17:45 crc kubenswrapper[4556]: I0218 09:17:45.526086 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-hgqj7" event={"ID":"12c97a89-6f28-4eab-add6-448279050305","Type":"ContainerStarted","Data":"72fff07c09446faeccec54641807858690433ad99d4ec281b539e023cf897518"} Feb 18 09:17:45 crc kubenswrapper[4556]: I0218 09:17:45.526106 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-hgqj7" event={"ID":"12c97a89-6f28-4eab-add6-448279050305","Type":"ContainerStarted","Data":"e7c7b1ca7e98d2b1641d31a99977afdd684299e479f768a0177d73bd25054c2e"} Feb 18 09:17:45 crc kubenswrapper[4556]: I0218 09:17:45.528531 4556 generic.go:334] "Generic (PLEG): container finished" podID="77955a10-5158-461d-a0fa-702e45c6eb87" containerID="f34782bd709e8ff17548a12e354d2d3f0cf7ae32095f7659c1cfeb4e92be4e5b" exitCode=0 Feb 18 09:17:45 crc kubenswrapper[4556]: I0218 09:17:45.528614 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-866975879-zxrfh" event={"ID":"77955a10-5158-461d-a0fa-702e45c6eb87","Type":"ContainerDied","Data":"f34782bd709e8ff17548a12e354d2d3f0cf7ae32095f7659c1cfeb4e92be4e5b"} Feb 18 09:17:45 crc kubenswrapper[4556]: I0218 09:17:45.528671 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-866975879-zxrfh" event={"ID":"77955a10-5158-461d-a0fa-702e45c6eb87","Type":"ContainerStarted","Data":"344a5c17e5b6f7e3a6aa2ce6ad3e8d7f7c258ea2c44da5ccd047011b0524809d"} Feb 18 09:17:45 crc kubenswrapper[4556]: I0218 09:17:45.535045 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-4k7vg" podStartSLOduration=1.535020833 podStartE2EDuration="1.535020833s" podCreationTimestamp="2026-02-18 09:17:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:17:45.534946403 +0000 UTC m=+822.551907403" watchObservedRunningTime="2026-02-18 09:17:45.535020833 +0000 UTC m=+822.551981813" Feb 18 09:17:45 crc kubenswrapper[4556]: I0218 09:17:45.572630 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-hgqj7" podStartSLOduration=2.572611361 podStartE2EDuration="2.572611361s" podCreationTimestamp="2026-02-18 09:17:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:17:45.572259116 +0000 UTC m=+822.589220096" watchObservedRunningTime="2026-02-18 09:17:45.572611361 +0000 UTC m=+822.589572341" Feb 18 09:17:45 crc kubenswrapper[4556]: I0218 09:17:45.637006 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f58898b5f-r69gj"] Feb 18 09:17:45 crc kubenswrapper[4556]: I0218 09:17:45.645274 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 18 09:17:45 crc kubenswrapper[4556]: I0218 09:17:45.667797 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-r4h4k"] Feb 18 09:17:45 crc kubenswrapper[4556]: I0218 09:17:45.753457 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 18 09:17:45 crc kubenswrapper[4556]: I0218 09:17:45.811257 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-866975879-zxrfh" Feb 18 09:17:45 crc kubenswrapper[4556]: I0218 09:17:45.921011 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ksgt5\" (UniqueName: \"kubernetes.io/projected/77955a10-5158-461d-a0fa-702e45c6eb87-kube-api-access-ksgt5\") pod \"77955a10-5158-461d-a0fa-702e45c6eb87\" (UID: \"77955a10-5158-461d-a0fa-702e45c6eb87\") " Feb 18 09:17:45 crc kubenswrapper[4556]: I0218 09:17:45.921235 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77955a10-5158-461d-a0fa-702e45c6eb87-config\") pod \"77955a10-5158-461d-a0fa-702e45c6eb87\" (UID: \"77955a10-5158-461d-a0fa-702e45c6eb87\") " Feb 18 09:17:45 crc kubenswrapper[4556]: I0218 09:17:45.921302 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/77955a10-5158-461d-a0fa-702e45c6eb87-dns-svc\") pod \"77955a10-5158-461d-a0fa-702e45c6eb87\" (UID: \"77955a10-5158-461d-a0fa-702e45c6eb87\") " Feb 18 09:17:45 crc kubenswrapper[4556]: I0218 09:17:45.921321 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/77955a10-5158-461d-a0fa-702e45c6eb87-ovsdbserver-nb\") pod \"77955a10-5158-461d-a0fa-702e45c6eb87\" (UID: \"77955a10-5158-461d-a0fa-702e45c6eb87\") " Feb 18 09:17:45 crc kubenswrapper[4556]: I0218 09:17:45.921395 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/77955a10-5158-461d-a0fa-702e45c6eb87-dns-swift-storage-0\") pod \"77955a10-5158-461d-a0fa-702e45c6eb87\" (UID: \"77955a10-5158-461d-a0fa-702e45c6eb87\") " Feb 18 09:17:45 crc kubenswrapper[4556]: I0218 09:17:45.921524 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/77955a10-5158-461d-a0fa-702e45c6eb87-ovsdbserver-sb\") pod \"77955a10-5158-461d-a0fa-702e45c6eb87\" (UID: \"77955a10-5158-461d-a0fa-702e45c6eb87\") " Feb 18 09:17:45 crc kubenswrapper[4556]: I0218 09:17:45.927885 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77955a10-5158-461d-a0fa-702e45c6eb87-kube-api-access-ksgt5" (OuterVolumeSpecName: "kube-api-access-ksgt5") pod "77955a10-5158-461d-a0fa-702e45c6eb87" (UID: "77955a10-5158-461d-a0fa-702e45c6eb87"). InnerVolumeSpecName "kube-api-access-ksgt5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:17:45 crc kubenswrapper[4556]: I0218 09:17:45.945682 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/77955a10-5158-461d-a0fa-702e45c6eb87-config" (OuterVolumeSpecName: "config") pod "77955a10-5158-461d-a0fa-702e45c6eb87" (UID: "77955a10-5158-461d-a0fa-702e45c6eb87"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:17:45 crc kubenswrapper[4556]: I0218 09:17:45.945719 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/77955a10-5158-461d-a0fa-702e45c6eb87-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "77955a10-5158-461d-a0fa-702e45c6eb87" (UID: "77955a10-5158-461d-a0fa-702e45c6eb87"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:17:45 crc kubenswrapper[4556]: I0218 09:17:45.952645 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/77955a10-5158-461d-a0fa-702e45c6eb87-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "77955a10-5158-461d-a0fa-702e45c6eb87" (UID: "77955a10-5158-461d-a0fa-702e45c6eb87"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:17:45 crc kubenswrapper[4556]: I0218 09:17:45.960076 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/77955a10-5158-461d-a0fa-702e45c6eb87-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "77955a10-5158-461d-a0fa-702e45c6eb87" (UID: "77955a10-5158-461d-a0fa-702e45c6eb87"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:17:45 crc kubenswrapper[4556]: I0218 09:17:45.967252 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/77955a10-5158-461d-a0fa-702e45c6eb87-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "77955a10-5158-461d-a0fa-702e45c6eb87" (UID: "77955a10-5158-461d-a0fa-702e45c6eb87"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:17:46 crc kubenswrapper[4556]: I0218 09:17:46.024119 4556 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/77955a10-5158-461d-a0fa-702e45c6eb87-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:46 crc kubenswrapper[4556]: I0218 09:17:46.024144 4556 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/77955a10-5158-461d-a0fa-702e45c6eb87-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:46 crc kubenswrapper[4556]: I0218 09:17:46.024167 4556 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/77955a10-5158-461d-a0fa-702e45c6eb87-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:46 crc kubenswrapper[4556]: I0218 09:17:46.024179 4556 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/77955a10-5158-461d-a0fa-702e45c6eb87-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:46 crc kubenswrapper[4556]: I0218 09:17:46.024188 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ksgt5\" (UniqueName: \"kubernetes.io/projected/77955a10-5158-461d-a0fa-702e45c6eb87-kube-api-access-ksgt5\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:46 crc kubenswrapper[4556]: I0218 09:17:46.024196 4556 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77955a10-5158-461d-a0fa-702e45c6eb87-config\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:46 crc kubenswrapper[4556]: I0218 09:17:46.437676 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 18 09:17:46 crc kubenswrapper[4556]: I0218 09:17:46.457127 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-74c5b6b6c5-q8djs"] Feb 18 09:17:46 crc kubenswrapper[4556]: I0218 09:17:46.478362 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-79dd98bf77-fdbv7"] Feb 18 09:17:46 crc kubenswrapper[4556]: E0218 09:17:46.478972 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77955a10-5158-461d-a0fa-702e45c6eb87" containerName="init" Feb 18 09:17:46 crc kubenswrapper[4556]: I0218 09:17:46.478992 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="77955a10-5158-461d-a0fa-702e45c6eb87" containerName="init" Feb 18 09:17:46 crc kubenswrapper[4556]: I0218 09:17:46.479271 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="77955a10-5158-461d-a0fa-702e45c6eb87" containerName="init" Feb 18 09:17:46 crc kubenswrapper[4556]: I0218 09:17:46.480533 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-79dd98bf77-fdbv7" Feb 18 09:17:46 crc kubenswrapper[4556]: I0218 09:17:46.498450 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 18 09:17:46 crc kubenswrapper[4556]: I0218 09:17:46.514087 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-79dd98bf77-fdbv7"] Feb 18 09:17:46 crc kubenswrapper[4556]: I0218 09:17:46.544198 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 18 09:17:46 crc kubenswrapper[4556]: I0218 09:17:46.544634 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5fec43c-7b17-45d5-91c4-609b6d2ca1b9-logs\") pod \"horizon-79dd98bf77-fdbv7\" (UID: \"c5fec43c-7b17-45d5-91c4-609b6d2ca1b9\") " pod="openstack/horizon-79dd98bf77-fdbv7" Feb 18 09:17:46 crc kubenswrapper[4556]: I0218 09:17:46.544686 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c5fec43c-7b17-45d5-91c4-609b6d2ca1b9-config-data\") pod \"horizon-79dd98bf77-fdbv7\" (UID: \"c5fec43c-7b17-45d5-91c4-609b6d2ca1b9\") " pod="openstack/horizon-79dd98bf77-fdbv7" Feb 18 09:17:46 crc kubenswrapper[4556]: I0218 09:17:46.544805 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8tpm2\" (UniqueName: \"kubernetes.io/projected/c5fec43c-7b17-45d5-91c4-609b6d2ca1b9-kube-api-access-8tpm2\") pod \"horizon-79dd98bf77-fdbv7\" (UID: \"c5fec43c-7b17-45d5-91c4-609b6d2ca1b9\") " pod="openstack/horizon-79dd98bf77-fdbv7" Feb 18 09:17:46 crc kubenswrapper[4556]: I0218 09:17:46.544864 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c5fec43c-7b17-45d5-91c4-609b6d2ca1b9-horizon-secret-key\") pod \"horizon-79dd98bf77-fdbv7\" (UID: \"c5fec43c-7b17-45d5-91c4-609b6d2ca1b9\") " pod="openstack/horizon-79dd98bf77-fdbv7" Feb 18 09:17:46 crc kubenswrapper[4556]: I0218 09:17:46.544912 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c5fec43c-7b17-45d5-91c4-609b6d2ca1b9-scripts\") pod \"horizon-79dd98bf77-fdbv7\" (UID: \"c5fec43c-7b17-45d5-91c4-609b6d2ca1b9\") " pod="openstack/horizon-79dd98bf77-fdbv7" Feb 18 09:17:46 crc kubenswrapper[4556]: I0218 09:17:46.556218 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d6cb6fa8-2338-4694-a50b-9f7ce0073e07","Type":"ContainerStarted","Data":"8e79eb8d29159033c9cff4d3f55c8268d4137c8a564954665b97128c6071a87e"} Feb 18 09:17:46 crc kubenswrapper[4556]: I0218 09:17:46.556273 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d6cb6fa8-2338-4694-a50b-9f7ce0073e07","Type":"ContainerStarted","Data":"c913bac115ee944d83bd31c8abc1196f7e475429b19306879be59ded55365dba"} Feb 18 09:17:46 crc kubenswrapper[4556]: I0218 09:17:46.558321 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-r4h4k" event={"ID":"8c46304b-1b60-493b-918f-678451f13057","Type":"ContainerStarted","Data":"2ef9a757b29b8804aea1b53755203b640f5b983a618f22d01e71c2a8661ad43a"} Feb 18 09:17:46 crc kubenswrapper[4556]: I0218 09:17:46.576648 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-866975879-zxrfh" Feb 18 09:17:46 crc kubenswrapper[4556]: I0218 09:17:46.576670 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-866975879-zxrfh" event={"ID":"77955a10-5158-461d-a0fa-702e45c6eb87","Type":"ContainerDied","Data":"344a5c17e5b6f7e3a6aa2ce6ad3e8d7f7c258ea2c44da5ccd047011b0524809d"} Feb 18 09:17:46 crc kubenswrapper[4556]: I0218 09:17:46.576731 4556 scope.go:117] "RemoveContainer" containerID="f34782bd709e8ff17548a12e354d2d3f0cf7ae32095f7659c1cfeb4e92be4e5b" Feb 18 09:17:46 crc kubenswrapper[4556]: I0218 09:17:46.581771 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5ee4e9c9-306f-4499-a53e-77cd55f1d28e","Type":"ContainerStarted","Data":"0d0f7d564279c563df5df5bf338470f0248ded1d833ce97f03daa4fe1e42455b"} Feb 18 09:17:46 crc kubenswrapper[4556]: I0218 09:17:46.581803 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5ee4e9c9-306f-4499-a53e-77cd55f1d28e","Type":"ContainerStarted","Data":"a4abc960b39b5f7647430759d873ce20bf67f05c9674065c82b9eb582cfd3b7b"} Feb 18 09:17:46 crc kubenswrapper[4556]: I0218 09:17:46.597068 4556 generic.go:334] "Generic (PLEG): container finished" podID="840ae99a-2a91-4c3e-9fab-b05787d3b97b" containerID="031d0ab58c1cf6e0f2a156e287886e3fb123b3baf49e0b05f7123414e3a8048f" exitCode=0 Feb 18 09:17:46 crc kubenswrapper[4556]: I0218 09:17:46.597264 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f58898b5f-r69gj" event={"ID":"840ae99a-2a91-4c3e-9fab-b05787d3b97b","Type":"ContainerDied","Data":"031d0ab58c1cf6e0f2a156e287886e3fb123b3baf49e0b05f7123414e3a8048f"} Feb 18 09:17:46 crc kubenswrapper[4556]: I0218 09:17:46.597329 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f58898b5f-r69gj" event={"ID":"840ae99a-2a91-4c3e-9fab-b05787d3b97b","Type":"ContainerStarted","Data":"1e33760599d03c798ff9357cdc78861f65926769817d4d7440761aaf7945dc91"} Feb 18 09:17:46 crc kubenswrapper[4556]: I0218 09:17:46.651103 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8tpm2\" (UniqueName: \"kubernetes.io/projected/c5fec43c-7b17-45d5-91c4-609b6d2ca1b9-kube-api-access-8tpm2\") pod \"horizon-79dd98bf77-fdbv7\" (UID: \"c5fec43c-7b17-45d5-91c4-609b6d2ca1b9\") " pod="openstack/horizon-79dd98bf77-fdbv7" Feb 18 09:17:46 crc kubenswrapper[4556]: I0218 09:17:46.651211 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c5fec43c-7b17-45d5-91c4-609b6d2ca1b9-horizon-secret-key\") pod \"horizon-79dd98bf77-fdbv7\" (UID: \"c5fec43c-7b17-45d5-91c4-609b6d2ca1b9\") " pod="openstack/horizon-79dd98bf77-fdbv7" Feb 18 09:17:46 crc kubenswrapper[4556]: I0218 09:17:46.651264 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c5fec43c-7b17-45d5-91c4-609b6d2ca1b9-scripts\") pod \"horizon-79dd98bf77-fdbv7\" (UID: \"c5fec43c-7b17-45d5-91c4-609b6d2ca1b9\") " pod="openstack/horizon-79dd98bf77-fdbv7" Feb 18 09:17:46 crc kubenswrapper[4556]: I0218 09:17:46.651686 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5fec43c-7b17-45d5-91c4-609b6d2ca1b9-logs\") pod \"horizon-79dd98bf77-fdbv7\" (UID: \"c5fec43c-7b17-45d5-91c4-609b6d2ca1b9\") " pod="openstack/horizon-79dd98bf77-fdbv7" Feb 18 09:17:46 crc kubenswrapper[4556]: I0218 09:17:46.651706 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c5fec43c-7b17-45d5-91c4-609b6d2ca1b9-config-data\") pod \"horizon-79dd98bf77-fdbv7\" (UID: \"c5fec43c-7b17-45d5-91c4-609b6d2ca1b9\") " pod="openstack/horizon-79dd98bf77-fdbv7" Feb 18 09:17:46 crc kubenswrapper[4556]: I0218 09:17:46.657704 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5fec43c-7b17-45d5-91c4-609b6d2ca1b9-logs\") pod \"horizon-79dd98bf77-fdbv7\" (UID: \"c5fec43c-7b17-45d5-91c4-609b6d2ca1b9\") " pod="openstack/horizon-79dd98bf77-fdbv7" Feb 18 09:17:46 crc kubenswrapper[4556]: I0218 09:17:46.660251 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c5fec43c-7b17-45d5-91c4-609b6d2ca1b9-config-data\") pod \"horizon-79dd98bf77-fdbv7\" (UID: \"c5fec43c-7b17-45d5-91c4-609b6d2ca1b9\") " pod="openstack/horizon-79dd98bf77-fdbv7" Feb 18 09:17:46 crc kubenswrapper[4556]: I0218 09:17:46.661528 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c5fec43c-7b17-45d5-91c4-609b6d2ca1b9-scripts\") pod \"horizon-79dd98bf77-fdbv7\" (UID: \"c5fec43c-7b17-45d5-91c4-609b6d2ca1b9\") " pod="openstack/horizon-79dd98bf77-fdbv7" Feb 18 09:17:46 crc kubenswrapper[4556]: I0218 09:17:46.672817 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-866975879-zxrfh"] Feb 18 09:17:46 crc kubenswrapper[4556]: I0218 09:17:46.680466 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c5fec43c-7b17-45d5-91c4-609b6d2ca1b9-horizon-secret-key\") pod \"horizon-79dd98bf77-fdbv7\" (UID: \"c5fec43c-7b17-45d5-91c4-609b6d2ca1b9\") " pod="openstack/horizon-79dd98bf77-fdbv7" Feb 18 09:17:46 crc kubenswrapper[4556]: I0218 09:17:46.688599 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-866975879-zxrfh"] Feb 18 09:17:46 crc kubenswrapper[4556]: I0218 09:17:46.701410 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8tpm2\" (UniqueName: \"kubernetes.io/projected/c5fec43c-7b17-45d5-91c4-609b6d2ca1b9-kube-api-access-8tpm2\") pod \"horizon-79dd98bf77-fdbv7\" (UID: \"c5fec43c-7b17-45d5-91c4-609b6d2ca1b9\") " pod="openstack/horizon-79dd98bf77-fdbv7" Feb 18 09:17:46 crc kubenswrapper[4556]: I0218 09:17:46.805258 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-79dd98bf77-fdbv7" Feb 18 09:17:47 crc kubenswrapper[4556]: I0218 09:17:47.294900 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77955a10-5158-461d-a0fa-702e45c6eb87" path="/var/lib/kubelet/pods/77955a10-5158-461d-a0fa-702e45c6eb87/volumes" Feb 18 09:17:47 crc kubenswrapper[4556]: I0218 09:17:47.381383 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-79dd98bf77-fdbv7"] Feb 18 09:17:47 crc kubenswrapper[4556]: W0218 09:17:47.416341 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc5fec43c_7b17_45d5_91c4_609b6d2ca1b9.slice/crio-6db6e7d4ac741e11b00a64d7861d1d97bfb5f5b1197764f1bac8b0fb6ff4df54 WatchSource:0}: Error finding container 6db6e7d4ac741e11b00a64d7861d1d97bfb5f5b1197764f1bac8b0fb6ff4df54: Status 404 returned error can't find the container with id 6db6e7d4ac741e11b00a64d7861d1d97bfb5f5b1197764f1bac8b0fb6ff4df54 Feb 18 09:17:47 crc kubenswrapper[4556]: I0218 09:17:47.615018 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d6cb6fa8-2338-4694-a50b-9f7ce0073e07","Type":"ContainerStarted","Data":"eecd9d3c29edc063eecf04c7d53530d0b1756f207cd6aae930b468b55cf920d9"} Feb 18 09:17:47 crc kubenswrapper[4556]: I0218 09:17:47.615095 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="d6cb6fa8-2338-4694-a50b-9f7ce0073e07" containerName="glance-log" containerID="cri-o://8e79eb8d29159033c9cff4d3f55c8268d4137c8a564954665b97128c6071a87e" gracePeriod=30 Feb 18 09:17:47 crc kubenswrapper[4556]: I0218 09:17:47.615434 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="d6cb6fa8-2338-4694-a50b-9f7ce0073e07" containerName="glance-httpd" containerID="cri-o://eecd9d3c29edc063eecf04c7d53530d0b1756f207cd6aae930b468b55cf920d9" gracePeriod=30 Feb 18 09:17:47 crc kubenswrapper[4556]: I0218 09:17:47.618898 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79dd98bf77-fdbv7" event={"ID":"c5fec43c-7b17-45d5-91c4-609b6d2ca1b9","Type":"ContainerStarted","Data":"6db6e7d4ac741e11b00a64d7861d1d97bfb5f5b1197764f1bac8b0fb6ff4df54"} Feb 18 09:17:47 crc kubenswrapper[4556]: I0218 09:17:47.625329 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5ee4e9c9-306f-4499-a53e-77cd55f1d28e","Type":"ContainerStarted","Data":"f35ac54859852987842a810c2e9f5843928d4bb1318b8941835f01a04c028c00"} Feb 18 09:17:47 crc kubenswrapper[4556]: I0218 09:17:47.625506 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="5ee4e9c9-306f-4499-a53e-77cd55f1d28e" containerName="glance-log" containerID="cri-o://0d0f7d564279c563df5df5bf338470f0248ded1d833ce97f03daa4fe1e42455b" gracePeriod=30 Feb 18 09:17:47 crc kubenswrapper[4556]: I0218 09:17:47.625550 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="5ee4e9c9-306f-4499-a53e-77cd55f1d28e" containerName="glance-httpd" containerID="cri-o://f35ac54859852987842a810c2e9f5843928d4bb1318b8941835f01a04c028c00" gracePeriod=30 Feb 18 09:17:47 crc kubenswrapper[4556]: I0218 09:17:47.639974 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f58898b5f-r69gj" event={"ID":"840ae99a-2a91-4c3e-9fab-b05787d3b97b","Type":"ContainerStarted","Data":"c607ad2bcebbaf1277b72f198c43bdc45e51329ad3f67a2c9ef45c2c139be0fc"} Feb 18 09:17:47 crc kubenswrapper[4556]: I0218 09:17:47.640416 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.6403742169999997 podStartE2EDuration="3.640374217s" podCreationTimestamp="2026-02-18 09:17:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:17:47.635814922 +0000 UTC m=+824.652775902" watchObservedRunningTime="2026-02-18 09:17:47.640374217 +0000 UTC m=+824.657335197" Feb 18 09:17:47 crc kubenswrapper[4556]: I0218 09:17:47.643188 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-f58898b5f-r69gj" Feb 18 09:17:47 crc kubenswrapper[4556]: I0218 09:17:47.666633 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.6666098529999998 podStartE2EDuration="3.666609853s" podCreationTimestamp="2026-02-18 09:17:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:17:47.658055769 +0000 UTC m=+824.675016748" watchObservedRunningTime="2026-02-18 09:17:47.666609853 +0000 UTC m=+824.683570833" Feb 18 09:17:47 crc kubenswrapper[4556]: I0218 09:17:47.685366 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-f58898b5f-r69gj" podStartSLOduration=3.6853466470000003 podStartE2EDuration="3.685346647s" podCreationTimestamp="2026-02-18 09:17:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:17:47.676148035 +0000 UTC m=+824.693109015" watchObservedRunningTime="2026-02-18 09:17:47.685346647 +0000 UTC m=+824.702307627" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.265498 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.409453 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6cb6fa8-2338-4694-a50b-9f7ce0073e07-logs\") pod \"d6cb6fa8-2338-4694-a50b-9f7ce0073e07\" (UID: \"d6cb6fa8-2338-4694-a50b-9f7ce0073e07\") " Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.409525 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d6cb6fa8-2338-4694-a50b-9f7ce0073e07-httpd-run\") pod \"d6cb6fa8-2338-4694-a50b-9f7ce0073e07\" (UID: \"d6cb6fa8-2338-4694-a50b-9f7ce0073e07\") " Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.409600 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6cb6fa8-2338-4694-a50b-9f7ce0073e07-combined-ca-bundle\") pod \"d6cb6fa8-2338-4694-a50b-9f7ce0073e07\" (UID: \"d6cb6fa8-2338-4694-a50b-9f7ce0073e07\") " Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.409631 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6cb6fa8-2338-4694-a50b-9f7ce0073e07-scripts\") pod \"d6cb6fa8-2338-4694-a50b-9f7ce0073e07\" (UID: \"d6cb6fa8-2338-4694-a50b-9f7ce0073e07\") " Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.409741 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6cb6fa8-2338-4694-a50b-9f7ce0073e07-config-data\") pod \"d6cb6fa8-2338-4694-a50b-9f7ce0073e07\" (UID: \"d6cb6fa8-2338-4694-a50b-9f7ce0073e07\") " Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.409797 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"d6cb6fa8-2338-4694-a50b-9f7ce0073e07\" (UID: \"d6cb6fa8-2338-4694-a50b-9f7ce0073e07\") " Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.409827 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6cb6fa8-2338-4694-a50b-9f7ce0073e07-public-tls-certs\") pod \"d6cb6fa8-2338-4694-a50b-9f7ce0073e07\" (UID: \"d6cb6fa8-2338-4694-a50b-9f7ce0073e07\") " Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.409862 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qw7lk\" (UniqueName: \"kubernetes.io/projected/d6cb6fa8-2338-4694-a50b-9f7ce0073e07-kube-api-access-qw7lk\") pod \"d6cb6fa8-2338-4694-a50b-9f7ce0073e07\" (UID: \"d6cb6fa8-2338-4694-a50b-9f7ce0073e07\") " Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.410051 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6cb6fa8-2338-4694-a50b-9f7ce0073e07-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "d6cb6fa8-2338-4694-a50b-9f7ce0073e07" (UID: "d6cb6fa8-2338-4694-a50b-9f7ce0073e07"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.410248 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6cb6fa8-2338-4694-a50b-9f7ce0073e07-logs" (OuterVolumeSpecName: "logs") pod "d6cb6fa8-2338-4694-a50b-9f7ce0073e07" (UID: "d6cb6fa8-2338-4694-a50b-9f7ce0073e07"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.410843 4556 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6cb6fa8-2338-4694-a50b-9f7ce0073e07-logs\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.410866 4556 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d6cb6fa8-2338-4694-a50b-9f7ce0073e07-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.415971 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6cb6fa8-2338-4694-a50b-9f7ce0073e07-kube-api-access-qw7lk" (OuterVolumeSpecName: "kube-api-access-qw7lk") pod "d6cb6fa8-2338-4694-a50b-9f7ce0073e07" (UID: "d6cb6fa8-2338-4694-a50b-9f7ce0073e07"). InnerVolumeSpecName "kube-api-access-qw7lk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.416969 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "d6cb6fa8-2338-4694-a50b-9f7ce0073e07" (UID: "d6cb6fa8-2338-4694-a50b-9f7ce0073e07"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.420840 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6cb6fa8-2338-4694-a50b-9f7ce0073e07-scripts" (OuterVolumeSpecName: "scripts") pod "d6cb6fa8-2338-4694-a50b-9f7ce0073e07" (UID: "d6cb6fa8-2338-4694-a50b-9f7ce0073e07"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.442858 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6cb6fa8-2338-4694-a50b-9f7ce0073e07-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d6cb6fa8-2338-4694-a50b-9f7ce0073e07" (UID: "d6cb6fa8-2338-4694-a50b-9f7ce0073e07"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.455538 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6cb6fa8-2338-4694-a50b-9f7ce0073e07-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "d6cb6fa8-2338-4694-a50b-9f7ce0073e07" (UID: "d6cb6fa8-2338-4694-a50b-9f7ce0073e07"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.461256 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6cb6fa8-2338-4694-a50b-9f7ce0073e07-config-data" (OuterVolumeSpecName: "config-data") pod "d6cb6fa8-2338-4694-a50b-9f7ce0073e07" (UID: "d6cb6fa8-2338-4694-a50b-9f7ce0073e07"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.508242 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.512737 4556 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.512778 4556 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6cb6fa8-2338-4694-a50b-9f7ce0073e07-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.512789 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qw7lk\" (UniqueName: \"kubernetes.io/projected/d6cb6fa8-2338-4694-a50b-9f7ce0073e07-kube-api-access-qw7lk\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.512799 4556 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6cb6fa8-2338-4694-a50b-9f7ce0073e07-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.512808 4556 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6cb6fa8-2338-4694-a50b-9f7ce0073e07-scripts\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.512820 4556 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6cb6fa8-2338-4694-a50b-9f7ce0073e07-config-data\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.530658 4556 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.614029 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ee4e9c9-306f-4499-a53e-77cd55f1d28e-scripts\") pod \"5ee4e9c9-306f-4499-a53e-77cd55f1d28e\" (UID: \"5ee4e9c9-306f-4499-a53e-77cd55f1d28e\") " Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.614093 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"5ee4e9c9-306f-4499-a53e-77cd55f1d28e\" (UID: \"5ee4e9c9-306f-4499-a53e-77cd55f1d28e\") " Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.614167 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ee4e9c9-306f-4499-a53e-77cd55f1d28e-combined-ca-bundle\") pod \"5ee4e9c9-306f-4499-a53e-77cd55f1d28e\" (UID: \"5ee4e9c9-306f-4499-a53e-77cd55f1d28e\") " Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.614182 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ee4e9c9-306f-4499-a53e-77cd55f1d28e-internal-tls-certs\") pod \"5ee4e9c9-306f-4499-a53e-77cd55f1d28e\" (UID: \"5ee4e9c9-306f-4499-a53e-77cd55f1d28e\") " Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.614228 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ee4e9c9-306f-4499-a53e-77cd55f1d28e-config-data\") pod \"5ee4e9c9-306f-4499-a53e-77cd55f1d28e\" (UID: \"5ee4e9c9-306f-4499-a53e-77cd55f1d28e\") " Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.614260 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qgmvt\" (UniqueName: \"kubernetes.io/projected/5ee4e9c9-306f-4499-a53e-77cd55f1d28e-kube-api-access-qgmvt\") pod \"5ee4e9c9-306f-4499-a53e-77cd55f1d28e\" (UID: \"5ee4e9c9-306f-4499-a53e-77cd55f1d28e\") " Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.614291 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ee4e9c9-306f-4499-a53e-77cd55f1d28e-logs\") pod \"5ee4e9c9-306f-4499-a53e-77cd55f1d28e\" (UID: \"5ee4e9c9-306f-4499-a53e-77cd55f1d28e\") " Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.614380 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5ee4e9c9-306f-4499-a53e-77cd55f1d28e-httpd-run\") pod \"5ee4e9c9-306f-4499-a53e-77cd55f1d28e\" (UID: \"5ee4e9c9-306f-4499-a53e-77cd55f1d28e\") " Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.614801 4556 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.619702 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ee4e9c9-306f-4499-a53e-77cd55f1d28e-scripts" (OuterVolumeSpecName: "scripts") pod "5ee4e9c9-306f-4499-a53e-77cd55f1d28e" (UID: "5ee4e9c9-306f-4499-a53e-77cd55f1d28e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.622276 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ee4e9c9-306f-4499-a53e-77cd55f1d28e-logs" (OuterVolumeSpecName: "logs") pod "5ee4e9c9-306f-4499-a53e-77cd55f1d28e" (UID: "5ee4e9c9-306f-4499-a53e-77cd55f1d28e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.624837 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ee4e9c9-306f-4499-a53e-77cd55f1d28e-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "5ee4e9c9-306f-4499-a53e-77cd55f1d28e" (UID: "5ee4e9c9-306f-4499-a53e-77cd55f1d28e"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.625512 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "5ee4e9c9-306f-4499-a53e-77cd55f1d28e" (UID: "5ee4e9c9-306f-4499-a53e-77cd55f1d28e"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.647491 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ee4e9c9-306f-4499-a53e-77cd55f1d28e-kube-api-access-qgmvt" (OuterVolumeSpecName: "kube-api-access-qgmvt") pod "5ee4e9c9-306f-4499-a53e-77cd55f1d28e" (UID: "5ee4e9c9-306f-4499-a53e-77cd55f1d28e"). InnerVolumeSpecName "kube-api-access-qgmvt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.661307 4556 generic.go:334] "Generic (PLEG): container finished" podID="d6cb6fa8-2338-4694-a50b-9f7ce0073e07" containerID="eecd9d3c29edc063eecf04c7d53530d0b1756f207cd6aae930b468b55cf920d9" exitCode=0 Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.661340 4556 generic.go:334] "Generic (PLEG): container finished" podID="d6cb6fa8-2338-4694-a50b-9f7ce0073e07" containerID="8e79eb8d29159033c9cff4d3f55c8268d4137c8a564954665b97128c6071a87e" exitCode=143 Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.661351 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.661401 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d6cb6fa8-2338-4694-a50b-9f7ce0073e07","Type":"ContainerDied","Data":"eecd9d3c29edc063eecf04c7d53530d0b1756f207cd6aae930b468b55cf920d9"} Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.661437 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d6cb6fa8-2338-4694-a50b-9f7ce0073e07","Type":"ContainerDied","Data":"8e79eb8d29159033c9cff4d3f55c8268d4137c8a564954665b97128c6071a87e"} Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.661449 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d6cb6fa8-2338-4694-a50b-9f7ce0073e07","Type":"ContainerDied","Data":"c913bac115ee944d83bd31c8abc1196f7e475429b19306879be59ded55365dba"} Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.661466 4556 scope.go:117] "RemoveContainer" containerID="eecd9d3c29edc063eecf04c7d53530d0b1756f207cd6aae930b468b55cf920d9" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.669735 4556 generic.go:334] "Generic (PLEG): container finished" podID="12c97a89-6f28-4eab-add6-448279050305" containerID="72fff07c09446faeccec54641807858690433ad99d4ec281b539e023cf897518" exitCode=0 Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.669822 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-hgqj7" event={"ID":"12c97a89-6f28-4eab-add6-448279050305","Type":"ContainerDied","Data":"72fff07c09446faeccec54641807858690433ad99d4ec281b539e023cf897518"} Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.676231 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ee4e9c9-306f-4499-a53e-77cd55f1d28e-config-data" (OuterVolumeSpecName: "config-data") pod "5ee4e9c9-306f-4499-a53e-77cd55f1d28e" (UID: "5ee4e9c9-306f-4499-a53e-77cd55f1d28e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.681585 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ee4e9c9-306f-4499-a53e-77cd55f1d28e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5ee4e9c9-306f-4499-a53e-77cd55f1d28e" (UID: "5ee4e9c9-306f-4499-a53e-77cd55f1d28e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.684007 4556 generic.go:334] "Generic (PLEG): container finished" podID="5ee4e9c9-306f-4499-a53e-77cd55f1d28e" containerID="f35ac54859852987842a810c2e9f5843928d4bb1318b8941835f01a04c028c00" exitCode=0 Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.684035 4556 generic.go:334] "Generic (PLEG): container finished" podID="5ee4e9c9-306f-4499-a53e-77cd55f1d28e" containerID="0d0f7d564279c563df5df5bf338470f0248ded1d833ce97f03daa4fe1e42455b" exitCode=143 Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.684104 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.684146 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5ee4e9c9-306f-4499-a53e-77cd55f1d28e","Type":"ContainerDied","Data":"f35ac54859852987842a810c2e9f5843928d4bb1318b8941835f01a04c028c00"} Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.684200 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5ee4e9c9-306f-4499-a53e-77cd55f1d28e","Type":"ContainerDied","Data":"0d0f7d564279c563df5df5bf338470f0248ded1d833ce97f03daa4fe1e42455b"} Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.684213 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5ee4e9c9-306f-4499-a53e-77cd55f1d28e","Type":"ContainerDied","Data":"a4abc960b39b5f7647430759d873ce20bf67f05c9674065c82b9eb582cfd3b7b"} Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.708660 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ee4e9c9-306f-4499-a53e-77cd55f1d28e-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "5ee4e9c9-306f-4499-a53e-77cd55f1d28e" (UID: "5ee4e9c9-306f-4499-a53e-77cd55f1d28e"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.721486 4556 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ee4e9c9-306f-4499-a53e-77cd55f1d28e-scripts\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.721535 4556 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.721552 4556 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ee4e9c9-306f-4499-a53e-77cd55f1d28e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.721563 4556 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ee4e9c9-306f-4499-a53e-77cd55f1d28e-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.721571 4556 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ee4e9c9-306f-4499-a53e-77cd55f1d28e-config-data\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.721579 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qgmvt\" (UniqueName: \"kubernetes.io/projected/5ee4e9c9-306f-4499-a53e-77cd55f1d28e-kube-api-access-qgmvt\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.721588 4556 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ee4e9c9-306f-4499-a53e-77cd55f1d28e-logs\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.721596 4556 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5ee4e9c9-306f-4499-a53e-77cd55f1d28e-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.724815 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.737220 4556 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.742353 4556 scope.go:117] "RemoveContainer" containerID="8e79eb8d29159033c9cff4d3f55c8268d4137c8a564954665b97128c6071a87e" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.744629 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.752104 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 18 09:17:48 crc kubenswrapper[4556]: E0218 09:17:48.752453 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6cb6fa8-2338-4694-a50b-9f7ce0073e07" containerName="glance-httpd" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.752470 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6cb6fa8-2338-4694-a50b-9f7ce0073e07" containerName="glance-httpd" Feb 18 09:17:48 crc kubenswrapper[4556]: E0218 09:17:48.752480 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ee4e9c9-306f-4499-a53e-77cd55f1d28e" containerName="glance-httpd" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.752485 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ee4e9c9-306f-4499-a53e-77cd55f1d28e" containerName="glance-httpd" Feb 18 09:17:48 crc kubenswrapper[4556]: E0218 09:17:48.752500 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ee4e9c9-306f-4499-a53e-77cd55f1d28e" containerName="glance-log" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.752505 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ee4e9c9-306f-4499-a53e-77cd55f1d28e" containerName="glance-log" Feb 18 09:17:48 crc kubenswrapper[4556]: E0218 09:17:48.752515 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6cb6fa8-2338-4694-a50b-9f7ce0073e07" containerName="glance-log" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.752520 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6cb6fa8-2338-4694-a50b-9f7ce0073e07" containerName="glance-log" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.752670 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6cb6fa8-2338-4694-a50b-9f7ce0073e07" containerName="glance-httpd" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.752678 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ee4e9c9-306f-4499-a53e-77cd55f1d28e" containerName="glance-log" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.752691 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ee4e9c9-306f-4499-a53e-77cd55f1d28e" containerName="glance-httpd" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.752704 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6cb6fa8-2338-4694-a50b-9f7ce0073e07" containerName="glance-log" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.753519 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.756394 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.757125 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.759628 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.783928 4556 scope.go:117] "RemoveContainer" containerID="eecd9d3c29edc063eecf04c7d53530d0b1756f207cd6aae930b468b55cf920d9" Feb 18 09:17:48 crc kubenswrapper[4556]: E0218 09:17:48.785940 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eecd9d3c29edc063eecf04c7d53530d0b1756f207cd6aae930b468b55cf920d9\": container with ID starting with eecd9d3c29edc063eecf04c7d53530d0b1756f207cd6aae930b468b55cf920d9 not found: ID does not exist" containerID="eecd9d3c29edc063eecf04c7d53530d0b1756f207cd6aae930b468b55cf920d9" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.785974 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eecd9d3c29edc063eecf04c7d53530d0b1756f207cd6aae930b468b55cf920d9"} err="failed to get container status \"eecd9d3c29edc063eecf04c7d53530d0b1756f207cd6aae930b468b55cf920d9\": rpc error: code = NotFound desc = could not find container \"eecd9d3c29edc063eecf04c7d53530d0b1756f207cd6aae930b468b55cf920d9\": container with ID starting with eecd9d3c29edc063eecf04c7d53530d0b1756f207cd6aae930b468b55cf920d9 not found: ID does not exist" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.785998 4556 scope.go:117] "RemoveContainer" containerID="8e79eb8d29159033c9cff4d3f55c8268d4137c8a564954665b97128c6071a87e" Feb 18 09:17:48 crc kubenswrapper[4556]: E0218 09:17:48.786356 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e79eb8d29159033c9cff4d3f55c8268d4137c8a564954665b97128c6071a87e\": container with ID starting with 8e79eb8d29159033c9cff4d3f55c8268d4137c8a564954665b97128c6071a87e not found: ID does not exist" containerID="8e79eb8d29159033c9cff4d3f55c8268d4137c8a564954665b97128c6071a87e" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.786378 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e79eb8d29159033c9cff4d3f55c8268d4137c8a564954665b97128c6071a87e"} err="failed to get container status \"8e79eb8d29159033c9cff4d3f55c8268d4137c8a564954665b97128c6071a87e\": rpc error: code = NotFound desc = could not find container \"8e79eb8d29159033c9cff4d3f55c8268d4137c8a564954665b97128c6071a87e\": container with ID starting with 8e79eb8d29159033c9cff4d3f55c8268d4137c8a564954665b97128c6071a87e not found: ID does not exist" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.786399 4556 scope.go:117] "RemoveContainer" containerID="eecd9d3c29edc063eecf04c7d53530d0b1756f207cd6aae930b468b55cf920d9" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.786855 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eecd9d3c29edc063eecf04c7d53530d0b1756f207cd6aae930b468b55cf920d9"} err="failed to get container status \"eecd9d3c29edc063eecf04c7d53530d0b1756f207cd6aae930b468b55cf920d9\": rpc error: code = NotFound desc = could not find container \"eecd9d3c29edc063eecf04c7d53530d0b1756f207cd6aae930b468b55cf920d9\": container with ID starting with eecd9d3c29edc063eecf04c7d53530d0b1756f207cd6aae930b468b55cf920d9 not found: ID does not exist" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.786873 4556 scope.go:117] "RemoveContainer" containerID="8e79eb8d29159033c9cff4d3f55c8268d4137c8a564954665b97128c6071a87e" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.787166 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e79eb8d29159033c9cff4d3f55c8268d4137c8a564954665b97128c6071a87e"} err="failed to get container status \"8e79eb8d29159033c9cff4d3f55c8268d4137c8a564954665b97128c6071a87e\": rpc error: code = NotFound desc = could not find container \"8e79eb8d29159033c9cff4d3f55c8268d4137c8a564954665b97128c6071a87e\": container with ID starting with 8e79eb8d29159033c9cff4d3f55c8268d4137c8a564954665b97128c6071a87e not found: ID does not exist" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.787184 4556 scope.go:117] "RemoveContainer" containerID="f35ac54859852987842a810c2e9f5843928d4bb1318b8941835f01a04c028c00" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.820002 4556 scope.go:117] "RemoveContainer" containerID="0d0f7d564279c563df5df5bf338470f0248ded1d833ce97f03daa4fe1e42455b" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.825299 4556 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.866332 4556 scope.go:117] "RemoveContainer" containerID="f35ac54859852987842a810c2e9f5843928d4bb1318b8941835f01a04c028c00" Feb 18 09:17:48 crc kubenswrapper[4556]: E0218 09:17:48.866791 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f35ac54859852987842a810c2e9f5843928d4bb1318b8941835f01a04c028c00\": container with ID starting with f35ac54859852987842a810c2e9f5843928d4bb1318b8941835f01a04c028c00 not found: ID does not exist" containerID="f35ac54859852987842a810c2e9f5843928d4bb1318b8941835f01a04c028c00" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.866832 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f35ac54859852987842a810c2e9f5843928d4bb1318b8941835f01a04c028c00"} err="failed to get container status \"f35ac54859852987842a810c2e9f5843928d4bb1318b8941835f01a04c028c00\": rpc error: code = NotFound desc = could not find container \"f35ac54859852987842a810c2e9f5843928d4bb1318b8941835f01a04c028c00\": container with ID starting with f35ac54859852987842a810c2e9f5843928d4bb1318b8941835f01a04c028c00 not found: ID does not exist" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.867874 4556 scope.go:117] "RemoveContainer" containerID="0d0f7d564279c563df5df5bf338470f0248ded1d833ce97f03daa4fe1e42455b" Feb 18 09:17:48 crc kubenswrapper[4556]: E0218 09:17:48.871077 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d0f7d564279c563df5df5bf338470f0248ded1d833ce97f03daa4fe1e42455b\": container with ID starting with 0d0f7d564279c563df5df5bf338470f0248ded1d833ce97f03daa4fe1e42455b not found: ID does not exist" containerID="0d0f7d564279c563df5df5bf338470f0248ded1d833ce97f03daa4fe1e42455b" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.871134 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d0f7d564279c563df5df5bf338470f0248ded1d833ce97f03daa4fe1e42455b"} err="failed to get container status \"0d0f7d564279c563df5df5bf338470f0248ded1d833ce97f03daa4fe1e42455b\": rpc error: code = NotFound desc = could not find container \"0d0f7d564279c563df5df5bf338470f0248ded1d833ce97f03daa4fe1e42455b\": container with ID starting with 0d0f7d564279c563df5df5bf338470f0248ded1d833ce97f03daa4fe1e42455b not found: ID does not exist" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.871192 4556 scope.go:117] "RemoveContainer" containerID="f35ac54859852987842a810c2e9f5843928d4bb1318b8941835f01a04c028c00" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.871733 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f35ac54859852987842a810c2e9f5843928d4bb1318b8941835f01a04c028c00"} err="failed to get container status \"f35ac54859852987842a810c2e9f5843928d4bb1318b8941835f01a04c028c00\": rpc error: code = NotFound desc = could not find container \"f35ac54859852987842a810c2e9f5843928d4bb1318b8941835f01a04c028c00\": container with ID starting with f35ac54859852987842a810c2e9f5843928d4bb1318b8941835f01a04c028c00 not found: ID does not exist" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.871800 4556 scope.go:117] "RemoveContainer" containerID="0d0f7d564279c563df5df5bf338470f0248ded1d833ce97f03daa4fe1e42455b" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.872406 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d0f7d564279c563df5df5bf338470f0248ded1d833ce97f03daa4fe1e42455b"} err="failed to get container status \"0d0f7d564279c563df5df5bf338470f0248ded1d833ce97f03daa4fe1e42455b\": rpc error: code = NotFound desc = could not find container \"0d0f7d564279c563df5df5bf338470f0248ded1d833ce97f03daa4fe1e42455b\": container with ID starting with 0d0f7d564279c563df5df5bf338470f0248ded1d833ce97f03daa4fe1e42455b not found: ID does not exist" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.929929 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7dd567df-a88f-4cfd-b465-f949d75ca4a2-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"7dd567df-a88f-4cfd-b465-f949d75ca4a2\") " pod="openstack/glance-default-external-api-0" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.930170 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7dd567df-a88f-4cfd-b465-f949d75ca4a2-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"7dd567df-a88f-4cfd-b465-f949d75ca4a2\") " pod="openstack/glance-default-external-api-0" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.930202 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7dd567df-a88f-4cfd-b465-f949d75ca4a2-scripts\") pod \"glance-default-external-api-0\" (UID: \"7dd567df-a88f-4cfd-b465-f949d75ca4a2\") " pod="openstack/glance-default-external-api-0" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.930812 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhg6z\" (UniqueName: \"kubernetes.io/projected/7dd567df-a88f-4cfd-b465-f949d75ca4a2-kube-api-access-dhg6z\") pod \"glance-default-external-api-0\" (UID: \"7dd567df-a88f-4cfd-b465-f949d75ca4a2\") " pod="openstack/glance-default-external-api-0" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.930897 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7dd567df-a88f-4cfd-b465-f949d75ca4a2-logs\") pod \"glance-default-external-api-0\" (UID: \"7dd567df-a88f-4cfd-b465-f949d75ca4a2\") " pod="openstack/glance-default-external-api-0" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.930951 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"7dd567df-a88f-4cfd-b465-f949d75ca4a2\") " pod="openstack/glance-default-external-api-0" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.931003 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7dd567df-a88f-4cfd-b465-f949d75ca4a2-config-data\") pod \"glance-default-external-api-0\" (UID: \"7dd567df-a88f-4cfd-b465-f949d75ca4a2\") " pod="openstack/glance-default-external-api-0" Feb 18 09:17:48 crc kubenswrapper[4556]: I0218 09:17:48.931216 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dd567df-a88f-4cfd-b465-f949d75ca4a2-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"7dd567df-a88f-4cfd-b465-f949d75ca4a2\") " pod="openstack/glance-default-external-api-0" Feb 18 09:17:49 crc kubenswrapper[4556]: I0218 09:17:49.033771 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7dd567df-a88f-4cfd-b465-f949d75ca4a2-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"7dd567df-a88f-4cfd-b465-f949d75ca4a2\") " pod="openstack/glance-default-external-api-0" Feb 18 09:17:49 crc kubenswrapper[4556]: I0218 09:17:49.033820 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7dd567df-a88f-4cfd-b465-f949d75ca4a2-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"7dd567df-a88f-4cfd-b465-f949d75ca4a2\") " pod="openstack/glance-default-external-api-0" Feb 18 09:17:49 crc kubenswrapper[4556]: I0218 09:17:49.033839 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7dd567df-a88f-4cfd-b465-f949d75ca4a2-scripts\") pod \"glance-default-external-api-0\" (UID: \"7dd567df-a88f-4cfd-b465-f949d75ca4a2\") " pod="openstack/glance-default-external-api-0" Feb 18 09:17:49 crc kubenswrapper[4556]: I0218 09:17:49.033910 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhg6z\" (UniqueName: \"kubernetes.io/projected/7dd567df-a88f-4cfd-b465-f949d75ca4a2-kube-api-access-dhg6z\") pod \"glance-default-external-api-0\" (UID: \"7dd567df-a88f-4cfd-b465-f949d75ca4a2\") " pod="openstack/glance-default-external-api-0" Feb 18 09:17:49 crc kubenswrapper[4556]: I0218 09:17:49.033967 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7dd567df-a88f-4cfd-b465-f949d75ca4a2-logs\") pod \"glance-default-external-api-0\" (UID: \"7dd567df-a88f-4cfd-b465-f949d75ca4a2\") " pod="openstack/glance-default-external-api-0" Feb 18 09:17:49 crc kubenswrapper[4556]: I0218 09:17:49.033997 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"7dd567df-a88f-4cfd-b465-f949d75ca4a2\") " pod="openstack/glance-default-external-api-0" Feb 18 09:17:49 crc kubenswrapper[4556]: I0218 09:17:49.034028 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7dd567df-a88f-4cfd-b465-f949d75ca4a2-config-data\") pod \"glance-default-external-api-0\" (UID: \"7dd567df-a88f-4cfd-b465-f949d75ca4a2\") " pod="openstack/glance-default-external-api-0" Feb 18 09:17:49 crc kubenswrapper[4556]: I0218 09:17:49.034131 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dd567df-a88f-4cfd-b465-f949d75ca4a2-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"7dd567df-a88f-4cfd-b465-f949d75ca4a2\") " pod="openstack/glance-default-external-api-0" Feb 18 09:17:49 crc kubenswrapper[4556]: I0218 09:17:49.034711 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 18 09:17:49 crc kubenswrapper[4556]: I0218 09:17:49.035416 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7dd567df-a88f-4cfd-b465-f949d75ca4a2-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"7dd567df-a88f-4cfd-b465-f949d75ca4a2\") " pod="openstack/glance-default-external-api-0" Feb 18 09:17:49 crc kubenswrapper[4556]: I0218 09:17:49.048471 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dd567df-a88f-4cfd-b465-f949d75ca4a2-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"7dd567df-a88f-4cfd-b465-f949d75ca4a2\") " pod="openstack/glance-default-external-api-0" Feb 18 09:17:49 crc kubenswrapper[4556]: I0218 09:17:49.048589 4556 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"7dd567df-a88f-4cfd-b465-f949d75ca4a2\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-external-api-0" Feb 18 09:17:49 crc kubenswrapper[4556]: I0218 09:17:49.049169 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7dd567df-a88f-4cfd-b465-f949d75ca4a2-logs\") pod \"glance-default-external-api-0\" (UID: \"7dd567df-a88f-4cfd-b465-f949d75ca4a2\") " pod="openstack/glance-default-external-api-0" Feb 18 09:17:49 crc kubenswrapper[4556]: I0218 09:17:49.059787 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7dd567df-a88f-4cfd-b465-f949d75ca4a2-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"7dd567df-a88f-4cfd-b465-f949d75ca4a2\") " pod="openstack/glance-default-external-api-0" Feb 18 09:17:49 crc kubenswrapper[4556]: I0218 09:17:49.070978 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 18 09:17:49 crc kubenswrapper[4556]: I0218 09:17:49.072805 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhg6z\" (UniqueName: \"kubernetes.io/projected/7dd567df-a88f-4cfd-b465-f949d75ca4a2-kube-api-access-dhg6z\") pod \"glance-default-external-api-0\" (UID: \"7dd567df-a88f-4cfd-b465-f949d75ca4a2\") " pod="openstack/glance-default-external-api-0" Feb 18 09:17:49 crc kubenswrapper[4556]: I0218 09:17:49.082531 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7dd567df-a88f-4cfd-b465-f949d75ca4a2-config-data\") pod \"glance-default-external-api-0\" (UID: \"7dd567df-a88f-4cfd-b465-f949d75ca4a2\") " pod="openstack/glance-default-external-api-0" Feb 18 09:17:49 crc kubenswrapper[4556]: I0218 09:17:49.096215 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7dd567df-a88f-4cfd-b465-f949d75ca4a2-scripts\") pod \"glance-default-external-api-0\" (UID: \"7dd567df-a88f-4cfd-b465-f949d75ca4a2\") " pod="openstack/glance-default-external-api-0" Feb 18 09:17:49 crc kubenswrapper[4556]: I0218 09:17:49.117214 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 18 09:17:49 crc kubenswrapper[4556]: I0218 09:17:49.120023 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 18 09:17:49 crc kubenswrapper[4556]: I0218 09:17:49.120331 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"7dd567df-a88f-4cfd-b465-f949d75ca4a2\") " pod="openstack/glance-default-external-api-0" Feb 18 09:17:49 crc kubenswrapper[4556]: I0218 09:17:49.127065 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 18 09:17:49 crc kubenswrapper[4556]: I0218 09:17:49.127234 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Feb 18 09:17:49 crc kubenswrapper[4556]: I0218 09:17:49.175143 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 18 09:17:49 crc kubenswrapper[4556]: I0218 09:17:49.238555 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"40566cc9-f4ba-4201-836a-811e07642dd8\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:17:49 crc kubenswrapper[4556]: I0218 09:17:49.238662 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/40566cc9-f4ba-4201-836a-811e07642dd8-logs\") pod \"glance-default-internal-api-0\" (UID: \"40566cc9-f4ba-4201-836a-811e07642dd8\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:17:49 crc kubenswrapper[4556]: I0218 09:17:49.238697 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/40566cc9-f4ba-4201-836a-811e07642dd8-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"40566cc9-f4ba-4201-836a-811e07642dd8\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:17:49 crc kubenswrapper[4556]: I0218 09:17:49.238730 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/40566cc9-f4ba-4201-836a-811e07642dd8-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"40566cc9-f4ba-4201-836a-811e07642dd8\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:17:49 crc kubenswrapper[4556]: I0218 09:17:49.238821 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40566cc9-f4ba-4201-836a-811e07642dd8-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"40566cc9-f4ba-4201-836a-811e07642dd8\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:17:49 crc kubenswrapper[4556]: I0218 09:17:49.238846 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40566cc9-f4ba-4201-836a-811e07642dd8-config-data\") pod \"glance-default-internal-api-0\" (UID: \"40566cc9-f4ba-4201-836a-811e07642dd8\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:17:49 crc kubenswrapper[4556]: I0218 09:17:49.238900 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/40566cc9-f4ba-4201-836a-811e07642dd8-scripts\") pod \"glance-default-internal-api-0\" (UID: \"40566cc9-f4ba-4201-836a-811e07642dd8\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:17:49 crc kubenswrapper[4556]: I0218 09:17:49.238919 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pdxkd\" (UniqueName: \"kubernetes.io/projected/40566cc9-f4ba-4201-836a-811e07642dd8-kube-api-access-pdxkd\") pod \"glance-default-internal-api-0\" (UID: \"40566cc9-f4ba-4201-836a-811e07642dd8\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:17:49 crc kubenswrapper[4556]: I0218 09:17:49.291763 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ee4e9c9-306f-4499-a53e-77cd55f1d28e" path="/var/lib/kubelet/pods/5ee4e9c9-306f-4499-a53e-77cd55f1d28e/volumes" Feb 18 09:17:49 crc kubenswrapper[4556]: I0218 09:17:49.292801 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6cb6fa8-2338-4694-a50b-9f7ce0073e07" path="/var/lib/kubelet/pods/d6cb6fa8-2338-4694-a50b-9f7ce0073e07/volumes" Feb 18 09:17:49 crc kubenswrapper[4556]: I0218 09:17:49.340905 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/40566cc9-f4ba-4201-836a-811e07642dd8-scripts\") pod \"glance-default-internal-api-0\" (UID: \"40566cc9-f4ba-4201-836a-811e07642dd8\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:17:49 crc kubenswrapper[4556]: I0218 09:17:49.340943 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pdxkd\" (UniqueName: \"kubernetes.io/projected/40566cc9-f4ba-4201-836a-811e07642dd8-kube-api-access-pdxkd\") pod \"glance-default-internal-api-0\" (UID: \"40566cc9-f4ba-4201-836a-811e07642dd8\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:17:49 crc kubenswrapper[4556]: I0218 09:17:49.341044 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"40566cc9-f4ba-4201-836a-811e07642dd8\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:17:49 crc kubenswrapper[4556]: I0218 09:17:49.341143 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/40566cc9-f4ba-4201-836a-811e07642dd8-logs\") pod \"glance-default-internal-api-0\" (UID: \"40566cc9-f4ba-4201-836a-811e07642dd8\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:17:49 crc kubenswrapper[4556]: I0218 09:17:49.341188 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/40566cc9-f4ba-4201-836a-811e07642dd8-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"40566cc9-f4ba-4201-836a-811e07642dd8\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:17:49 crc kubenswrapper[4556]: I0218 09:17:49.341217 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/40566cc9-f4ba-4201-836a-811e07642dd8-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"40566cc9-f4ba-4201-836a-811e07642dd8\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:17:49 crc kubenswrapper[4556]: I0218 09:17:49.341315 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40566cc9-f4ba-4201-836a-811e07642dd8-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"40566cc9-f4ba-4201-836a-811e07642dd8\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:17:49 crc kubenswrapper[4556]: I0218 09:17:49.341334 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40566cc9-f4ba-4201-836a-811e07642dd8-config-data\") pod \"glance-default-internal-api-0\" (UID: \"40566cc9-f4ba-4201-836a-811e07642dd8\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:17:49 crc kubenswrapper[4556]: I0218 09:17:49.342000 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/40566cc9-f4ba-4201-836a-811e07642dd8-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"40566cc9-f4ba-4201-836a-811e07642dd8\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:17:49 crc kubenswrapper[4556]: I0218 09:17:49.342656 4556 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"40566cc9-f4ba-4201-836a-811e07642dd8\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Feb 18 09:17:49 crc kubenswrapper[4556]: I0218 09:17:49.344644 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/40566cc9-f4ba-4201-836a-811e07642dd8-logs\") pod \"glance-default-internal-api-0\" (UID: \"40566cc9-f4ba-4201-836a-811e07642dd8\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:17:49 crc kubenswrapper[4556]: I0218 09:17:49.345714 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/40566cc9-f4ba-4201-836a-811e07642dd8-scripts\") pod \"glance-default-internal-api-0\" (UID: \"40566cc9-f4ba-4201-836a-811e07642dd8\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:17:49 crc kubenswrapper[4556]: I0218 09:17:49.346211 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/40566cc9-f4ba-4201-836a-811e07642dd8-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"40566cc9-f4ba-4201-836a-811e07642dd8\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:17:49 crc kubenswrapper[4556]: I0218 09:17:49.346220 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40566cc9-f4ba-4201-836a-811e07642dd8-config-data\") pod \"glance-default-internal-api-0\" (UID: \"40566cc9-f4ba-4201-836a-811e07642dd8\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:17:49 crc kubenswrapper[4556]: I0218 09:17:49.349263 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40566cc9-f4ba-4201-836a-811e07642dd8-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"40566cc9-f4ba-4201-836a-811e07642dd8\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:17:49 crc kubenswrapper[4556]: I0218 09:17:49.356245 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdxkd\" (UniqueName: \"kubernetes.io/projected/40566cc9-f4ba-4201-836a-811e07642dd8-kube-api-access-pdxkd\") pod \"glance-default-internal-api-0\" (UID: \"40566cc9-f4ba-4201-836a-811e07642dd8\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:17:49 crc kubenswrapper[4556]: I0218 09:17:49.364715 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"40566cc9-f4ba-4201-836a-811e07642dd8\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:17:49 crc kubenswrapper[4556]: I0218 09:17:49.373722 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 18 09:17:49 crc kubenswrapper[4556]: I0218 09:17:49.526698 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 18 09:17:53 crc kubenswrapper[4556]: I0218 09:17:53.693239 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-77cf5996cc-dj8s4"] Feb 18 09:17:53 crc kubenswrapper[4556]: I0218 09:17:53.718063 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 18 09:17:53 crc kubenswrapper[4556]: I0218 09:17:53.739164 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-85bb8c66f4-fhblh"] Feb 18 09:17:53 crc kubenswrapper[4556]: I0218 09:17:53.740546 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-85bb8c66f4-fhblh" Feb 18 09:17:53 crc kubenswrapper[4556]: I0218 09:17:53.743367 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Feb 18 09:17:53 crc kubenswrapper[4556]: I0218 09:17:53.750486 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-85bb8c66f4-fhblh"] Feb 18 09:17:53 crc kubenswrapper[4556]: I0218 09:17:53.769428 4556 generic.go:334] "Generic (PLEG): container finished" podID="6bf90aa4-605f-4b7d-abde-2a1f3236a6af" containerID="63868b1e8917bd5b78ec577b9090971cf503941ce12fce83e038cadfaa8c7368" exitCode=0 Feb 18 09:17:53 crc kubenswrapper[4556]: I0218 09:17:53.769476 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-4k7vg" event={"ID":"6bf90aa4-605f-4b7d-abde-2a1f3236a6af","Type":"ContainerDied","Data":"63868b1e8917bd5b78ec577b9090971cf503941ce12fce83e038cadfaa8c7368"} Feb 18 09:17:53 crc kubenswrapper[4556]: I0218 09:17:53.793593 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-79dd98bf77-fdbv7"] Feb 18 09:17:53 crc kubenswrapper[4556]: I0218 09:17:53.803204 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 18 09:17:53 crc kubenswrapper[4556]: I0218 09:17:53.848696 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3354b930-6b00-4aa7-a4bd-97ec410cc863-combined-ca-bundle\") pod \"horizon-85bb8c66f4-fhblh\" (UID: \"3354b930-6b00-4aa7-a4bd-97ec410cc863\") " pod="openstack/horizon-85bb8c66f4-fhblh" Feb 18 09:17:53 crc kubenswrapper[4556]: I0218 09:17:53.849075 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffnjj\" (UniqueName: \"kubernetes.io/projected/3354b930-6b00-4aa7-a4bd-97ec410cc863-kube-api-access-ffnjj\") pod \"horizon-85bb8c66f4-fhblh\" (UID: \"3354b930-6b00-4aa7-a4bd-97ec410cc863\") " pod="openstack/horizon-85bb8c66f4-fhblh" Feb 18 09:17:53 crc kubenswrapper[4556]: I0218 09:17:53.849294 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3354b930-6b00-4aa7-a4bd-97ec410cc863-config-data\") pod \"horizon-85bb8c66f4-fhblh\" (UID: \"3354b930-6b00-4aa7-a4bd-97ec410cc863\") " pod="openstack/horizon-85bb8c66f4-fhblh" Feb 18 09:17:53 crc kubenswrapper[4556]: I0218 09:17:53.849555 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/3354b930-6b00-4aa7-a4bd-97ec410cc863-horizon-tls-certs\") pod \"horizon-85bb8c66f4-fhblh\" (UID: \"3354b930-6b00-4aa7-a4bd-97ec410cc863\") " pod="openstack/horizon-85bb8c66f4-fhblh" Feb 18 09:17:53 crc kubenswrapper[4556]: I0218 09:17:53.849845 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3354b930-6b00-4aa7-a4bd-97ec410cc863-scripts\") pod \"horizon-85bb8c66f4-fhblh\" (UID: \"3354b930-6b00-4aa7-a4bd-97ec410cc863\") " pod="openstack/horizon-85bb8c66f4-fhblh" Feb 18 09:17:53 crc kubenswrapper[4556]: I0218 09:17:53.849981 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3354b930-6b00-4aa7-a4bd-97ec410cc863-logs\") pod \"horizon-85bb8c66f4-fhblh\" (UID: \"3354b930-6b00-4aa7-a4bd-97ec410cc863\") " pod="openstack/horizon-85bb8c66f4-fhblh" Feb 18 09:17:53 crc kubenswrapper[4556]: I0218 09:17:53.850211 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3354b930-6b00-4aa7-a4bd-97ec410cc863-horizon-secret-key\") pod \"horizon-85bb8c66f4-fhblh\" (UID: \"3354b930-6b00-4aa7-a4bd-97ec410cc863\") " pod="openstack/horizon-85bb8c66f4-fhblh" Feb 18 09:17:53 crc kubenswrapper[4556]: I0218 09:17:53.853046 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-595566bb8b-6b5h6"] Feb 18 09:17:53 crc kubenswrapper[4556]: I0218 09:17:53.861701 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-595566bb8b-6b5h6" Feb 18 09:17:53 crc kubenswrapper[4556]: I0218 09:17:53.866370 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-595566bb8b-6b5h6"] Feb 18 09:17:53 crc kubenswrapper[4556]: I0218 09:17:53.952591 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffnjj\" (UniqueName: \"kubernetes.io/projected/3354b930-6b00-4aa7-a4bd-97ec410cc863-kube-api-access-ffnjj\") pod \"horizon-85bb8c66f4-fhblh\" (UID: \"3354b930-6b00-4aa7-a4bd-97ec410cc863\") " pod="openstack/horizon-85bb8c66f4-fhblh" Feb 18 09:17:53 crc kubenswrapper[4556]: I0218 09:17:53.952692 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3354b930-6b00-4aa7-a4bd-97ec410cc863-config-data\") pod \"horizon-85bb8c66f4-fhblh\" (UID: \"3354b930-6b00-4aa7-a4bd-97ec410cc863\") " pod="openstack/horizon-85bb8c66f4-fhblh" Feb 18 09:17:53 crc kubenswrapper[4556]: I0218 09:17:53.952752 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/30fffa0b-72a5-4bbe-a1fb-fa8b26d1decf-config-data\") pod \"horizon-595566bb8b-6b5h6\" (UID: \"30fffa0b-72a5-4bbe-a1fb-fa8b26d1decf\") " pod="openstack/horizon-595566bb8b-6b5h6" Feb 18 09:17:53 crc kubenswrapper[4556]: I0218 09:17:53.952773 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/30fffa0b-72a5-4bbe-a1fb-fa8b26d1decf-logs\") pod \"horizon-595566bb8b-6b5h6\" (UID: \"30fffa0b-72a5-4bbe-a1fb-fa8b26d1decf\") " pod="openstack/horizon-595566bb8b-6b5h6" Feb 18 09:17:53 crc kubenswrapper[4556]: I0218 09:17:53.952815 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rh9cn\" (UniqueName: \"kubernetes.io/projected/30fffa0b-72a5-4bbe-a1fb-fa8b26d1decf-kube-api-access-rh9cn\") pod \"horizon-595566bb8b-6b5h6\" (UID: \"30fffa0b-72a5-4bbe-a1fb-fa8b26d1decf\") " pod="openstack/horizon-595566bb8b-6b5h6" Feb 18 09:17:53 crc kubenswrapper[4556]: I0218 09:17:53.952902 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/3354b930-6b00-4aa7-a4bd-97ec410cc863-horizon-tls-certs\") pod \"horizon-85bb8c66f4-fhblh\" (UID: \"3354b930-6b00-4aa7-a4bd-97ec410cc863\") " pod="openstack/horizon-85bb8c66f4-fhblh" Feb 18 09:17:53 crc kubenswrapper[4556]: I0218 09:17:53.952987 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3354b930-6b00-4aa7-a4bd-97ec410cc863-scripts\") pod \"horizon-85bb8c66f4-fhblh\" (UID: \"3354b930-6b00-4aa7-a4bd-97ec410cc863\") " pod="openstack/horizon-85bb8c66f4-fhblh" Feb 18 09:17:53 crc kubenswrapper[4556]: I0218 09:17:53.953017 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/30fffa0b-72a5-4bbe-a1fb-fa8b26d1decf-horizon-secret-key\") pod \"horizon-595566bb8b-6b5h6\" (UID: \"30fffa0b-72a5-4bbe-a1fb-fa8b26d1decf\") " pod="openstack/horizon-595566bb8b-6b5h6" Feb 18 09:17:53 crc kubenswrapper[4556]: I0218 09:17:53.953047 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3354b930-6b00-4aa7-a4bd-97ec410cc863-logs\") pod \"horizon-85bb8c66f4-fhblh\" (UID: \"3354b930-6b00-4aa7-a4bd-97ec410cc863\") " pod="openstack/horizon-85bb8c66f4-fhblh" Feb 18 09:17:53 crc kubenswrapper[4556]: I0218 09:17:53.953070 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/30fffa0b-72a5-4bbe-a1fb-fa8b26d1decf-scripts\") pod \"horizon-595566bb8b-6b5h6\" (UID: \"30fffa0b-72a5-4bbe-a1fb-fa8b26d1decf\") " pod="openstack/horizon-595566bb8b-6b5h6" Feb 18 09:17:53 crc kubenswrapper[4556]: I0218 09:17:53.953123 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30fffa0b-72a5-4bbe-a1fb-fa8b26d1decf-combined-ca-bundle\") pod \"horizon-595566bb8b-6b5h6\" (UID: \"30fffa0b-72a5-4bbe-a1fb-fa8b26d1decf\") " pod="openstack/horizon-595566bb8b-6b5h6" Feb 18 09:17:53 crc kubenswrapper[4556]: I0218 09:17:53.953189 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3354b930-6b00-4aa7-a4bd-97ec410cc863-horizon-secret-key\") pod \"horizon-85bb8c66f4-fhblh\" (UID: \"3354b930-6b00-4aa7-a4bd-97ec410cc863\") " pod="openstack/horizon-85bb8c66f4-fhblh" Feb 18 09:17:53 crc kubenswrapper[4556]: I0218 09:17:53.953218 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/30fffa0b-72a5-4bbe-a1fb-fa8b26d1decf-horizon-tls-certs\") pod \"horizon-595566bb8b-6b5h6\" (UID: \"30fffa0b-72a5-4bbe-a1fb-fa8b26d1decf\") " pod="openstack/horizon-595566bb8b-6b5h6" Feb 18 09:17:53 crc kubenswrapper[4556]: I0218 09:17:53.953258 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3354b930-6b00-4aa7-a4bd-97ec410cc863-combined-ca-bundle\") pod \"horizon-85bb8c66f4-fhblh\" (UID: \"3354b930-6b00-4aa7-a4bd-97ec410cc863\") " pod="openstack/horizon-85bb8c66f4-fhblh" Feb 18 09:17:53 crc kubenswrapper[4556]: I0218 09:17:53.954004 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3354b930-6b00-4aa7-a4bd-97ec410cc863-scripts\") pod \"horizon-85bb8c66f4-fhblh\" (UID: \"3354b930-6b00-4aa7-a4bd-97ec410cc863\") " pod="openstack/horizon-85bb8c66f4-fhblh" Feb 18 09:17:53 crc kubenswrapper[4556]: I0218 09:17:53.954204 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3354b930-6b00-4aa7-a4bd-97ec410cc863-logs\") pod \"horizon-85bb8c66f4-fhblh\" (UID: \"3354b930-6b00-4aa7-a4bd-97ec410cc863\") " pod="openstack/horizon-85bb8c66f4-fhblh" Feb 18 09:17:53 crc kubenswrapper[4556]: I0218 09:17:53.954333 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3354b930-6b00-4aa7-a4bd-97ec410cc863-config-data\") pod \"horizon-85bb8c66f4-fhblh\" (UID: \"3354b930-6b00-4aa7-a4bd-97ec410cc863\") " pod="openstack/horizon-85bb8c66f4-fhblh" Feb 18 09:17:53 crc kubenswrapper[4556]: I0218 09:17:53.959138 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3354b930-6b00-4aa7-a4bd-97ec410cc863-horizon-secret-key\") pod \"horizon-85bb8c66f4-fhblh\" (UID: \"3354b930-6b00-4aa7-a4bd-97ec410cc863\") " pod="openstack/horizon-85bb8c66f4-fhblh" Feb 18 09:17:53 crc kubenswrapper[4556]: I0218 09:17:53.959476 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3354b930-6b00-4aa7-a4bd-97ec410cc863-combined-ca-bundle\") pod \"horizon-85bb8c66f4-fhblh\" (UID: \"3354b930-6b00-4aa7-a4bd-97ec410cc863\") " pod="openstack/horizon-85bb8c66f4-fhblh" Feb 18 09:17:53 crc kubenswrapper[4556]: I0218 09:17:53.966852 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffnjj\" (UniqueName: \"kubernetes.io/projected/3354b930-6b00-4aa7-a4bd-97ec410cc863-kube-api-access-ffnjj\") pod \"horizon-85bb8c66f4-fhblh\" (UID: \"3354b930-6b00-4aa7-a4bd-97ec410cc863\") " pod="openstack/horizon-85bb8c66f4-fhblh" Feb 18 09:17:53 crc kubenswrapper[4556]: I0218 09:17:53.976166 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/3354b930-6b00-4aa7-a4bd-97ec410cc863-horizon-tls-certs\") pod \"horizon-85bb8c66f4-fhblh\" (UID: \"3354b930-6b00-4aa7-a4bd-97ec410cc863\") " pod="openstack/horizon-85bb8c66f4-fhblh" Feb 18 09:17:54 crc kubenswrapper[4556]: I0218 09:17:54.054905 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/30fffa0b-72a5-4bbe-a1fb-fa8b26d1decf-horizon-tls-certs\") pod \"horizon-595566bb8b-6b5h6\" (UID: \"30fffa0b-72a5-4bbe-a1fb-fa8b26d1decf\") " pod="openstack/horizon-595566bb8b-6b5h6" Feb 18 09:17:54 crc kubenswrapper[4556]: I0218 09:17:54.055098 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/30fffa0b-72a5-4bbe-a1fb-fa8b26d1decf-config-data\") pod \"horizon-595566bb8b-6b5h6\" (UID: \"30fffa0b-72a5-4bbe-a1fb-fa8b26d1decf\") " pod="openstack/horizon-595566bb8b-6b5h6" Feb 18 09:17:54 crc kubenswrapper[4556]: I0218 09:17:54.055129 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/30fffa0b-72a5-4bbe-a1fb-fa8b26d1decf-logs\") pod \"horizon-595566bb8b-6b5h6\" (UID: \"30fffa0b-72a5-4bbe-a1fb-fa8b26d1decf\") " pod="openstack/horizon-595566bb8b-6b5h6" Feb 18 09:17:54 crc kubenswrapper[4556]: I0218 09:17:54.055180 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rh9cn\" (UniqueName: \"kubernetes.io/projected/30fffa0b-72a5-4bbe-a1fb-fa8b26d1decf-kube-api-access-rh9cn\") pod \"horizon-595566bb8b-6b5h6\" (UID: \"30fffa0b-72a5-4bbe-a1fb-fa8b26d1decf\") " pod="openstack/horizon-595566bb8b-6b5h6" Feb 18 09:17:54 crc kubenswrapper[4556]: I0218 09:17:54.055299 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/30fffa0b-72a5-4bbe-a1fb-fa8b26d1decf-horizon-secret-key\") pod \"horizon-595566bb8b-6b5h6\" (UID: \"30fffa0b-72a5-4bbe-a1fb-fa8b26d1decf\") " pod="openstack/horizon-595566bb8b-6b5h6" Feb 18 09:17:54 crc kubenswrapper[4556]: I0218 09:17:54.055337 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/30fffa0b-72a5-4bbe-a1fb-fa8b26d1decf-scripts\") pod \"horizon-595566bb8b-6b5h6\" (UID: \"30fffa0b-72a5-4bbe-a1fb-fa8b26d1decf\") " pod="openstack/horizon-595566bb8b-6b5h6" Feb 18 09:17:54 crc kubenswrapper[4556]: I0218 09:17:54.055396 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30fffa0b-72a5-4bbe-a1fb-fa8b26d1decf-combined-ca-bundle\") pod \"horizon-595566bb8b-6b5h6\" (UID: \"30fffa0b-72a5-4bbe-a1fb-fa8b26d1decf\") " pod="openstack/horizon-595566bb8b-6b5h6" Feb 18 09:17:54 crc kubenswrapper[4556]: I0218 09:17:54.057646 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/30fffa0b-72a5-4bbe-a1fb-fa8b26d1decf-logs\") pod \"horizon-595566bb8b-6b5h6\" (UID: \"30fffa0b-72a5-4bbe-a1fb-fa8b26d1decf\") " pod="openstack/horizon-595566bb8b-6b5h6" Feb 18 09:17:54 crc kubenswrapper[4556]: I0218 09:17:54.058027 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/30fffa0b-72a5-4bbe-a1fb-fa8b26d1decf-scripts\") pod \"horizon-595566bb8b-6b5h6\" (UID: \"30fffa0b-72a5-4bbe-a1fb-fa8b26d1decf\") " pod="openstack/horizon-595566bb8b-6b5h6" Feb 18 09:17:54 crc kubenswrapper[4556]: I0218 09:17:54.058565 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/30fffa0b-72a5-4bbe-a1fb-fa8b26d1decf-config-data\") pod \"horizon-595566bb8b-6b5h6\" (UID: \"30fffa0b-72a5-4bbe-a1fb-fa8b26d1decf\") " pod="openstack/horizon-595566bb8b-6b5h6" Feb 18 09:17:54 crc kubenswrapper[4556]: I0218 09:17:54.059985 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30fffa0b-72a5-4bbe-a1fb-fa8b26d1decf-combined-ca-bundle\") pod \"horizon-595566bb8b-6b5h6\" (UID: \"30fffa0b-72a5-4bbe-a1fb-fa8b26d1decf\") " pod="openstack/horizon-595566bb8b-6b5h6" Feb 18 09:17:54 crc kubenswrapper[4556]: I0218 09:17:54.060060 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/30fffa0b-72a5-4bbe-a1fb-fa8b26d1decf-horizon-secret-key\") pod \"horizon-595566bb8b-6b5h6\" (UID: \"30fffa0b-72a5-4bbe-a1fb-fa8b26d1decf\") " pod="openstack/horizon-595566bb8b-6b5h6" Feb 18 09:17:54 crc kubenswrapper[4556]: I0218 09:17:54.060761 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/30fffa0b-72a5-4bbe-a1fb-fa8b26d1decf-horizon-tls-certs\") pod \"horizon-595566bb8b-6b5h6\" (UID: \"30fffa0b-72a5-4bbe-a1fb-fa8b26d1decf\") " pod="openstack/horizon-595566bb8b-6b5h6" Feb 18 09:17:54 crc kubenswrapper[4556]: I0218 09:17:54.069323 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-85bb8c66f4-fhblh" Feb 18 09:17:54 crc kubenswrapper[4556]: I0218 09:17:54.069761 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rh9cn\" (UniqueName: \"kubernetes.io/projected/30fffa0b-72a5-4bbe-a1fb-fa8b26d1decf-kube-api-access-rh9cn\") pod \"horizon-595566bb8b-6b5h6\" (UID: \"30fffa0b-72a5-4bbe-a1fb-fa8b26d1decf\") " pod="openstack/horizon-595566bb8b-6b5h6" Feb 18 09:17:54 crc kubenswrapper[4556]: I0218 09:17:54.181094 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-595566bb8b-6b5h6" Feb 18 09:17:54 crc kubenswrapper[4556]: I0218 09:17:54.962729 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-f58898b5f-r69gj" Feb 18 09:17:55 crc kubenswrapper[4556]: I0218 09:17:55.036038 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54cc5c87cc-ck95n"] Feb 18 09:17:55 crc kubenswrapper[4556]: I0218 09:17:55.036295 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-54cc5c87cc-ck95n" podUID="269247cf-79a0-49a1-8fc9-da82411da8e5" containerName="dnsmasq-dns" containerID="cri-o://a62979af6d8d1b2e9b80ec0dfe2c502d3c4c4c1c4f03145c6c6e9d54c955a58b" gracePeriod=10 Feb 18 09:17:55 crc kubenswrapper[4556]: I0218 09:17:55.789885 4556 generic.go:334] "Generic (PLEG): container finished" podID="269247cf-79a0-49a1-8fc9-da82411da8e5" containerID="a62979af6d8d1b2e9b80ec0dfe2c502d3c4c4c1c4f03145c6c6e9d54c955a58b" exitCode=0 Feb 18 09:17:55 crc kubenswrapper[4556]: I0218 09:17:55.789988 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54cc5c87cc-ck95n" event={"ID":"269247cf-79a0-49a1-8fc9-da82411da8e5","Type":"ContainerDied","Data":"a62979af6d8d1b2e9b80ec0dfe2c502d3c4c4c1c4f03145c6c6e9d54c955a58b"} Feb 18 09:17:58 crc kubenswrapper[4556]: I0218 09:17:58.461144 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-hgqj7" Feb 18 09:17:58 crc kubenswrapper[4556]: I0218 09:17:58.580090 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6rwkn\" (UniqueName: \"kubernetes.io/projected/12c97a89-6f28-4eab-add6-448279050305-kube-api-access-6rwkn\") pod \"12c97a89-6f28-4eab-add6-448279050305\" (UID: \"12c97a89-6f28-4eab-add6-448279050305\") " Feb 18 09:17:58 crc kubenswrapper[4556]: I0218 09:17:58.580149 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12c97a89-6f28-4eab-add6-448279050305-combined-ca-bundle\") pod \"12c97a89-6f28-4eab-add6-448279050305\" (UID: \"12c97a89-6f28-4eab-add6-448279050305\") " Feb 18 09:17:58 crc kubenswrapper[4556]: I0218 09:17:58.580196 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12c97a89-6f28-4eab-add6-448279050305-scripts\") pod \"12c97a89-6f28-4eab-add6-448279050305\" (UID: \"12c97a89-6f28-4eab-add6-448279050305\") " Feb 18 09:17:58 crc kubenswrapper[4556]: I0218 09:17:58.580228 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/12c97a89-6f28-4eab-add6-448279050305-credential-keys\") pod \"12c97a89-6f28-4eab-add6-448279050305\" (UID: \"12c97a89-6f28-4eab-add6-448279050305\") " Feb 18 09:17:58 crc kubenswrapper[4556]: I0218 09:17:58.580271 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12c97a89-6f28-4eab-add6-448279050305-config-data\") pod \"12c97a89-6f28-4eab-add6-448279050305\" (UID: \"12c97a89-6f28-4eab-add6-448279050305\") " Feb 18 09:17:58 crc kubenswrapper[4556]: I0218 09:17:58.580295 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/12c97a89-6f28-4eab-add6-448279050305-fernet-keys\") pod \"12c97a89-6f28-4eab-add6-448279050305\" (UID: \"12c97a89-6f28-4eab-add6-448279050305\") " Feb 18 09:17:58 crc kubenswrapper[4556]: I0218 09:17:58.587016 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12c97a89-6f28-4eab-add6-448279050305-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "12c97a89-6f28-4eab-add6-448279050305" (UID: "12c97a89-6f28-4eab-add6-448279050305"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:17:58 crc kubenswrapper[4556]: I0218 09:17:58.587407 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12c97a89-6f28-4eab-add6-448279050305-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "12c97a89-6f28-4eab-add6-448279050305" (UID: "12c97a89-6f28-4eab-add6-448279050305"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:17:58 crc kubenswrapper[4556]: I0218 09:17:58.602639 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12c97a89-6f28-4eab-add6-448279050305-config-data" (OuterVolumeSpecName: "config-data") pod "12c97a89-6f28-4eab-add6-448279050305" (UID: "12c97a89-6f28-4eab-add6-448279050305"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:17:58 crc kubenswrapper[4556]: I0218 09:17:58.604048 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12c97a89-6f28-4eab-add6-448279050305-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "12c97a89-6f28-4eab-add6-448279050305" (UID: "12c97a89-6f28-4eab-add6-448279050305"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:17:58 crc kubenswrapper[4556]: I0218 09:17:58.606314 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12c97a89-6f28-4eab-add6-448279050305-kube-api-access-6rwkn" (OuterVolumeSpecName: "kube-api-access-6rwkn") pod "12c97a89-6f28-4eab-add6-448279050305" (UID: "12c97a89-6f28-4eab-add6-448279050305"). InnerVolumeSpecName "kube-api-access-6rwkn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:17:58 crc kubenswrapper[4556]: I0218 09:17:58.609357 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12c97a89-6f28-4eab-add6-448279050305-scripts" (OuterVolumeSpecName: "scripts") pod "12c97a89-6f28-4eab-add6-448279050305" (UID: "12c97a89-6f28-4eab-add6-448279050305"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:17:58 crc kubenswrapper[4556]: I0218 09:17:58.682817 4556 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/12c97a89-6f28-4eab-add6-448279050305-credential-keys\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:58 crc kubenswrapper[4556]: I0218 09:17:58.682847 4556 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12c97a89-6f28-4eab-add6-448279050305-config-data\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:58 crc kubenswrapper[4556]: I0218 09:17:58.682857 4556 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/12c97a89-6f28-4eab-add6-448279050305-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:58 crc kubenswrapper[4556]: I0218 09:17:58.682865 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6rwkn\" (UniqueName: \"kubernetes.io/projected/12c97a89-6f28-4eab-add6-448279050305-kube-api-access-6rwkn\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:58 crc kubenswrapper[4556]: I0218 09:17:58.682895 4556 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12c97a89-6f28-4eab-add6-448279050305-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:58 crc kubenswrapper[4556]: I0218 09:17:58.682903 4556 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12c97a89-6f28-4eab-add6-448279050305-scripts\") on node \"crc\" DevicePath \"\"" Feb 18 09:17:58 crc kubenswrapper[4556]: I0218 09:17:58.820294 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-hgqj7" event={"ID":"12c97a89-6f28-4eab-add6-448279050305","Type":"ContainerDied","Data":"e7c7b1ca7e98d2b1641d31a99977afdd684299e479f768a0177d73bd25054c2e"} Feb 18 09:17:58 crc kubenswrapper[4556]: I0218 09:17:58.820334 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e7c7b1ca7e98d2b1641d31a99977afdd684299e479f768a0177d73bd25054c2e" Feb 18 09:17:58 crc kubenswrapper[4556]: I0218 09:17:58.820399 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-hgqj7" Feb 18 09:17:59 crc kubenswrapper[4556]: I0218 09:17:59.058012 4556 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-54cc5c87cc-ck95n" podUID="269247cf-79a0-49a1-8fc9-da82411da8e5" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.123:5353: connect: connection refused" Feb 18 09:17:59 crc kubenswrapper[4556]: I0218 09:17:59.532741 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-hgqj7"] Feb 18 09:17:59 crc kubenswrapper[4556]: I0218 09:17:59.538700 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-hgqj7"] Feb 18 09:17:59 crc kubenswrapper[4556]: I0218 09:17:59.627731 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-brdr4"] Feb 18 09:17:59 crc kubenswrapper[4556]: E0218 09:17:59.628281 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12c97a89-6f28-4eab-add6-448279050305" containerName="keystone-bootstrap" Feb 18 09:17:59 crc kubenswrapper[4556]: I0218 09:17:59.628301 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="12c97a89-6f28-4eab-add6-448279050305" containerName="keystone-bootstrap" Feb 18 09:17:59 crc kubenswrapper[4556]: I0218 09:17:59.628542 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="12c97a89-6f28-4eab-add6-448279050305" containerName="keystone-bootstrap" Feb 18 09:17:59 crc kubenswrapper[4556]: I0218 09:17:59.629145 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-brdr4" Feb 18 09:17:59 crc kubenswrapper[4556]: I0218 09:17:59.631084 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 18 09:17:59 crc kubenswrapper[4556]: I0218 09:17:59.631379 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 18 09:17:59 crc kubenswrapper[4556]: I0218 09:17:59.631433 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-dpsn5" Feb 18 09:17:59 crc kubenswrapper[4556]: I0218 09:17:59.631396 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Feb 18 09:17:59 crc kubenswrapper[4556]: I0218 09:17:59.631531 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 18 09:17:59 crc kubenswrapper[4556]: I0218 09:17:59.638113 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-brdr4"] Feb 18 09:17:59 crc kubenswrapper[4556]: I0218 09:17:59.700602 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6347657-1897-458a-9a02-36d8655525f2-config-data\") pod \"keystone-bootstrap-brdr4\" (UID: \"a6347657-1897-458a-9a02-36d8655525f2\") " pod="openstack/keystone-bootstrap-brdr4" Feb 18 09:17:59 crc kubenswrapper[4556]: I0218 09:17:59.700657 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a6347657-1897-458a-9a02-36d8655525f2-credential-keys\") pod \"keystone-bootstrap-brdr4\" (UID: \"a6347657-1897-458a-9a02-36d8655525f2\") " pod="openstack/keystone-bootstrap-brdr4" Feb 18 09:17:59 crc kubenswrapper[4556]: I0218 09:17:59.700764 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6347657-1897-458a-9a02-36d8655525f2-combined-ca-bundle\") pod \"keystone-bootstrap-brdr4\" (UID: \"a6347657-1897-458a-9a02-36d8655525f2\") " pod="openstack/keystone-bootstrap-brdr4" Feb 18 09:17:59 crc kubenswrapper[4556]: I0218 09:17:59.700886 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27blw\" (UniqueName: \"kubernetes.io/projected/a6347657-1897-458a-9a02-36d8655525f2-kube-api-access-27blw\") pod \"keystone-bootstrap-brdr4\" (UID: \"a6347657-1897-458a-9a02-36d8655525f2\") " pod="openstack/keystone-bootstrap-brdr4" Feb 18 09:17:59 crc kubenswrapper[4556]: I0218 09:17:59.700960 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6347657-1897-458a-9a02-36d8655525f2-scripts\") pod \"keystone-bootstrap-brdr4\" (UID: \"a6347657-1897-458a-9a02-36d8655525f2\") " pod="openstack/keystone-bootstrap-brdr4" Feb 18 09:17:59 crc kubenswrapper[4556]: I0218 09:17:59.700986 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a6347657-1897-458a-9a02-36d8655525f2-fernet-keys\") pod \"keystone-bootstrap-brdr4\" (UID: \"a6347657-1897-458a-9a02-36d8655525f2\") " pod="openstack/keystone-bootstrap-brdr4" Feb 18 09:17:59 crc kubenswrapper[4556]: E0218 09:17:59.759887 4556 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api@sha256:657020ed78b5d92505b0b4187dfcf078515484304fd39ce38702d4fb06f4ca36" Feb 18 09:17:59 crc kubenswrapper[4556]: E0218 09:17:59.760014 4556 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api@sha256:657020ed78b5d92505b0b4187dfcf078515484304fd39ce38702d4fb06f4ca36,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-x5dj6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-xhbg8_openstack(2b111dd2-11a6-47cc-ba9e-2886186c676b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 18 09:17:59 crc kubenswrapper[4556]: E0218 09:17:59.761445 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-xhbg8" podUID="2b111dd2-11a6-47cc-ba9e-2886186c676b" Feb 18 09:17:59 crc kubenswrapper[4556]: I0218 09:17:59.802686 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6347657-1897-458a-9a02-36d8655525f2-config-data\") pod \"keystone-bootstrap-brdr4\" (UID: \"a6347657-1897-458a-9a02-36d8655525f2\") " pod="openstack/keystone-bootstrap-brdr4" Feb 18 09:17:59 crc kubenswrapper[4556]: I0218 09:17:59.802758 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a6347657-1897-458a-9a02-36d8655525f2-credential-keys\") pod \"keystone-bootstrap-brdr4\" (UID: \"a6347657-1897-458a-9a02-36d8655525f2\") " pod="openstack/keystone-bootstrap-brdr4" Feb 18 09:17:59 crc kubenswrapper[4556]: I0218 09:17:59.802827 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6347657-1897-458a-9a02-36d8655525f2-combined-ca-bundle\") pod \"keystone-bootstrap-brdr4\" (UID: \"a6347657-1897-458a-9a02-36d8655525f2\") " pod="openstack/keystone-bootstrap-brdr4" Feb 18 09:17:59 crc kubenswrapper[4556]: I0218 09:17:59.802876 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27blw\" (UniqueName: \"kubernetes.io/projected/a6347657-1897-458a-9a02-36d8655525f2-kube-api-access-27blw\") pod \"keystone-bootstrap-brdr4\" (UID: \"a6347657-1897-458a-9a02-36d8655525f2\") " pod="openstack/keystone-bootstrap-brdr4" Feb 18 09:17:59 crc kubenswrapper[4556]: I0218 09:17:59.802908 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6347657-1897-458a-9a02-36d8655525f2-scripts\") pod \"keystone-bootstrap-brdr4\" (UID: \"a6347657-1897-458a-9a02-36d8655525f2\") " pod="openstack/keystone-bootstrap-brdr4" Feb 18 09:17:59 crc kubenswrapper[4556]: I0218 09:17:59.802928 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a6347657-1897-458a-9a02-36d8655525f2-fernet-keys\") pod \"keystone-bootstrap-brdr4\" (UID: \"a6347657-1897-458a-9a02-36d8655525f2\") " pod="openstack/keystone-bootstrap-brdr4" Feb 18 09:17:59 crc kubenswrapper[4556]: I0218 09:17:59.807982 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6347657-1897-458a-9a02-36d8655525f2-scripts\") pod \"keystone-bootstrap-brdr4\" (UID: \"a6347657-1897-458a-9a02-36d8655525f2\") " pod="openstack/keystone-bootstrap-brdr4" Feb 18 09:17:59 crc kubenswrapper[4556]: I0218 09:17:59.808263 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a6347657-1897-458a-9a02-36d8655525f2-fernet-keys\") pod \"keystone-bootstrap-brdr4\" (UID: \"a6347657-1897-458a-9a02-36d8655525f2\") " pod="openstack/keystone-bootstrap-brdr4" Feb 18 09:17:59 crc kubenswrapper[4556]: I0218 09:17:59.808442 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6347657-1897-458a-9a02-36d8655525f2-config-data\") pod \"keystone-bootstrap-brdr4\" (UID: \"a6347657-1897-458a-9a02-36d8655525f2\") " pod="openstack/keystone-bootstrap-brdr4" Feb 18 09:17:59 crc kubenswrapper[4556]: I0218 09:17:59.808944 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6347657-1897-458a-9a02-36d8655525f2-combined-ca-bundle\") pod \"keystone-bootstrap-brdr4\" (UID: \"a6347657-1897-458a-9a02-36d8655525f2\") " pod="openstack/keystone-bootstrap-brdr4" Feb 18 09:17:59 crc kubenswrapper[4556]: I0218 09:17:59.815763 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a6347657-1897-458a-9a02-36d8655525f2-credential-keys\") pod \"keystone-bootstrap-brdr4\" (UID: \"a6347657-1897-458a-9a02-36d8655525f2\") " pod="openstack/keystone-bootstrap-brdr4" Feb 18 09:17:59 crc kubenswrapper[4556]: I0218 09:17:59.819655 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27blw\" (UniqueName: \"kubernetes.io/projected/a6347657-1897-458a-9a02-36d8655525f2-kube-api-access-27blw\") pod \"keystone-bootstrap-brdr4\" (UID: \"a6347657-1897-458a-9a02-36d8655525f2\") " pod="openstack/keystone-bootstrap-brdr4" Feb 18 09:17:59 crc kubenswrapper[4556]: I0218 09:17:59.834136 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-4k7vg" event={"ID":"6bf90aa4-605f-4b7d-abde-2a1f3236a6af","Type":"ContainerDied","Data":"cfd07d686eaf20681a09ee7a134c96b1944015bcecbe62d365f5d03747a5fa25"} Feb 18 09:17:59 crc kubenswrapper[4556]: I0218 09:17:59.834199 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cfd07d686eaf20681a09ee7a134c96b1944015bcecbe62d365f5d03747a5fa25" Feb 18 09:17:59 crc kubenswrapper[4556]: E0218 09:17:59.836502 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api@sha256:657020ed78b5d92505b0b4187dfcf078515484304fd39ce38702d4fb06f4ca36\\\"\"" pod="openstack/placement-db-sync-xhbg8" podUID="2b111dd2-11a6-47cc-ba9e-2886186c676b" Feb 18 09:17:59 crc kubenswrapper[4556]: I0218 09:17:59.878535 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-4k7vg" Feb 18 09:17:59 crc kubenswrapper[4556]: I0218 09:17:59.952270 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-brdr4" Feb 18 09:18:00 crc kubenswrapper[4556]: I0218 09:18:00.006728 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hmcgv\" (UniqueName: \"kubernetes.io/projected/6bf90aa4-605f-4b7d-abde-2a1f3236a6af-kube-api-access-hmcgv\") pod \"6bf90aa4-605f-4b7d-abde-2a1f3236a6af\" (UID: \"6bf90aa4-605f-4b7d-abde-2a1f3236a6af\") " Feb 18 09:18:00 crc kubenswrapper[4556]: I0218 09:18:00.006959 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6bf90aa4-605f-4b7d-abde-2a1f3236a6af-config\") pod \"6bf90aa4-605f-4b7d-abde-2a1f3236a6af\" (UID: \"6bf90aa4-605f-4b7d-abde-2a1f3236a6af\") " Feb 18 09:18:00 crc kubenswrapper[4556]: I0218 09:18:00.007017 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bf90aa4-605f-4b7d-abde-2a1f3236a6af-combined-ca-bundle\") pod \"6bf90aa4-605f-4b7d-abde-2a1f3236a6af\" (UID: \"6bf90aa4-605f-4b7d-abde-2a1f3236a6af\") " Feb 18 09:18:00 crc kubenswrapper[4556]: I0218 09:18:00.024525 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6bf90aa4-605f-4b7d-abde-2a1f3236a6af-kube-api-access-hmcgv" (OuterVolumeSpecName: "kube-api-access-hmcgv") pod "6bf90aa4-605f-4b7d-abde-2a1f3236a6af" (UID: "6bf90aa4-605f-4b7d-abde-2a1f3236a6af"). InnerVolumeSpecName "kube-api-access-hmcgv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:18:00 crc kubenswrapper[4556]: I0218 09:18:00.026825 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bf90aa4-605f-4b7d-abde-2a1f3236a6af-config" (OuterVolumeSpecName: "config") pod "6bf90aa4-605f-4b7d-abde-2a1f3236a6af" (UID: "6bf90aa4-605f-4b7d-abde-2a1f3236a6af"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:18:00 crc kubenswrapper[4556]: I0218 09:18:00.032830 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bf90aa4-605f-4b7d-abde-2a1f3236a6af-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6bf90aa4-605f-4b7d-abde-2a1f3236a6af" (UID: "6bf90aa4-605f-4b7d-abde-2a1f3236a6af"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:18:00 crc kubenswrapper[4556]: I0218 09:18:00.109955 4556 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bf90aa4-605f-4b7d-abde-2a1f3236a6af-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:00 crc kubenswrapper[4556]: I0218 09:18:00.109984 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hmcgv\" (UniqueName: \"kubernetes.io/projected/6bf90aa4-605f-4b7d-abde-2a1f3236a6af-kube-api-access-hmcgv\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:00 crc kubenswrapper[4556]: I0218 09:18:00.109997 4556 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/6bf90aa4-605f-4b7d-abde-2a1f3236a6af-config\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:00 crc kubenswrapper[4556]: I0218 09:18:00.841124 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-4k7vg" Feb 18 09:18:01 crc kubenswrapper[4556]: I0218 09:18:01.003053 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86dd6887f5-kqgml"] Feb 18 09:18:01 crc kubenswrapper[4556]: E0218 09:18:01.003468 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bf90aa4-605f-4b7d-abde-2a1f3236a6af" containerName="neutron-db-sync" Feb 18 09:18:01 crc kubenswrapper[4556]: I0218 09:18:01.003488 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bf90aa4-605f-4b7d-abde-2a1f3236a6af" containerName="neutron-db-sync" Feb 18 09:18:01 crc kubenswrapper[4556]: I0218 09:18:01.003674 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="6bf90aa4-605f-4b7d-abde-2a1f3236a6af" containerName="neutron-db-sync" Feb 18 09:18:01 crc kubenswrapper[4556]: I0218 09:18:01.004510 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86dd6887f5-kqgml" Feb 18 09:18:01 crc kubenswrapper[4556]: I0218 09:18:01.035530 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86dd6887f5-kqgml"] Feb 18 09:18:01 crc kubenswrapper[4556]: I0218 09:18:01.135057 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/02b5c258-0253-4d75-852b-30c2b26b5598-dns-swift-storage-0\") pod \"dnsmasq-dns-86dd6887f5-kqgml\" (UID: \"02b5c258-0253-4d75-852b-30c2b26b5598\") " pod="openstack/dnsmasq-dns-86dd6887f5-kqgml" Feb 18 09:18:01 crc kubenswrapper[4556]: I0218 09:18:01.135168 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/02b5c258-0253-4d75-852b-30c2b26b5598-dns-svc\") pod \"dnsmasq-dns-86dd6887f5-kqgml\" (UID: \"02b5c258-0253-4d75-852b-30c2b26b5598\") " pod="openstack/dnsmasq-dns-86dd6887f5-kqgml" Feb 18 09:18:01 crc kubenswrapper[4556]: I0218 09:18:01.135256 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02b5c258-0253-4d75-852b-30c2b26b5598-config\") pod \"dnsmasq-dns-86dd6887f5-kqgml\" (UID: \"02b5c258-0253-4d75-852b-30c2b26b5598\") " pod="openstack/dnsmasq-dns-86dd6887f5-kqgml" Feb 18 09:18:01 crc kubenswrapper[4556]: I0218 09:18:01.135338 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/02b5c258-0253-4d75-852b-30c2b26b5598-ovsdbserver-sb\") pod \"dnsmasq-dns-86dd6887f5-kqgml\" (UID: \"02b5c258-0253-4d75-852b-30c2b26b5598\") " pod="openstack/dnsmasq-dns-86dd6887f5-kqgml" Feb 18 09:18:01 crc kubenswrapper[4556]: I0218 09:18:01.135673 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4hph\" (UniqueName: \"kubernetes.io/projected/02b5c258-0253-4d75-852b-30c2b26b5598-kube-api-access-c4hph\") pod \"dnsmasq-dns-86dd6887f5-kqgml\" (UID: \"02b5c258-0253-4d75-852b-30c2b26b5598\") " pod="openstack/dnsmasq-dns-86dd6887f5-kqgml" Feb 18 09:18:01 crc kubenswrapper[4556]: I0218 09:18:01.135733 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/02b5c258-0253-4d75-852b-30c2b26b5598-ovsdbserver-nb\") pod \"dnsmasq-dns-86dd6887f5-kqgml\" (UID: \"02b5c258-0253-4d75-852b-30c2b26b5598\") " pod="openstack/dnsmasq-dns-86dd6887f5-kqgml" Feb 18 09:18:01 crc kubenswrapper[4556]: I0218 09:18:01.237801 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4hph\" (UniqueName: \"kubernetes.io/projected/02b5c258-0253-4d75-852b-30c2b26b5598-kube-api-access-c4hph\") pod \"dnsmasq-dns-86dd6887f5-kqgml\" (UID: \"02b5c258-0253-4d75-852b-30c2b26b5598\") " pod="openstack/dnsmasq-dns-86dd6887f5-kqgml" Feb 18 09:18:01 crc kubenswrapper[4556]: I0218 09:18:01.237869 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/02b5c258-0253-4d75-852b-30c2b26b5598-ovsdbserver-nb\") pod \"dnsmasq-dns-86dd6887f5-kqgml\" (UID: \"02b5c258-0253-4d75-852b-30c2b26b5598\") " pod="openstack/dnsmasq-dns-86dd6887f5-kqgml" Feb 18 09:18:01 crc kubenswrapper[4556]: I0218 09:18:01.237903 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/02b5c258-0253-4d75-852b-30c2b26b5598-dns-swift-storage-0\") pod \"dnsmasq-dns-86dd6887f5-kqgml\" (UID: \"02b5c258-0253-4d75-852b-30c2b26b5598\") " pod="openstack/dnsmasq-dns-86dd6887f5-kqgml" Feb 18 09:18:01 crc kubenswrapper[4556]: I0218 09:18:01.237921 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/02b5c258-0253-4d75-852b-30c2b26b5598-dns-svc\") pod \"dnsmasq-dns-86dd6887f5-kqgml\" (UID: \"02b5c258-0253-4d75-852b-30c2b26b5598\") " pod="openstack/dnsmasq-dns-86dd6887f5-kqgml" Feb 18 09:18:01 crc kubenswrapper[4556]: I0218 09:18:01.237978 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02b5c258-0253-4d75-852b-30c2b26b5598-config\") pod \"dnsmasq-dns-86dd6887f5-kqgml\" (UID: \"02b5c258-0253-4d75-852b-30c2b26b5598\") " pod="openstack/dnsmasq-dns-86dd6887f5-kqgml" Feb 18 09:18:01 crc kubenswrapper[4556]: I0218 09:18:01.238019 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/02b5c258-0253-4d75-852b-30c2b26b5598-ovsdbserver-sb\") pod \"dnsmasq-dns-86dd6887f5-kqgml\" (UID: \"02b5c258-0253-4d75-852b-30c2b26b5598\") " pod="openstack/dnsmasq-dns-86dd6887f5-kqgml" Feb 18 09:18:01 crc kubenswrapper[4556]: I0218 09:18:01.239201 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/02b5c258-0253-4d75-852b-30c2b26b5598-dns-swift-storage-0\") pod \"dnsmasq-dns-86dd6887f5-kqgml\" (UID: \"02b5c258-0253-4d75-852b-30c2b26b5598\") " pod="openstack/dnsmasq-dns-86dd6887f5-kqgml" Feb 18 09:18:01 crc kubenswrapper[4556]: I0218 09:18:01.239509 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/02b5c258-0253-4d75-852b-30c2b26b5598-dns-svc\") pod \"dnsmasq-dns-86dd6887f5-kqgml\" (UID: \"02b5c258-0253-4d75-852b-30c2b26b5598\") " pod="openstack/dnsmasq-dns-86dd6887f5-kqgml" Feb 18 09:18:01 crc kubenswrapper[4556]: I0218 09:18:01.240260 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02b5c258-0253-4d75-852b-30c2b26b5598-config\") pod \"dnsmasq-dns-86dd6887f5-kqgml\" (UID: \"02b5c258-0253-4d75-852b-30c2b26b5598\") " pod="openstack/dnsmasq-dns-86dd6887f5-kqgml" Feb 18 09:18:01 crc kubenswrapper[4556]: I0218 09:18:01.240415 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/02b5c258-0253-4d75-852b-30c2b26b5598-ovsdbserver-nb\") pod \"dnsmasq-dns-86dd6887f5-kqgml\" (UID: \"02b5c258-0253-4d75-852b-30c2b26b5598\") " pod="openstack/dnsmasq-dns-86dd6887f5-kqgml" Feb 18 09:18:01 crc kubenswrapper[4556]: I0218 09:18:01.240870 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/02b5c258-0253-4d75-852b-30c2b26b5598-ovsdbserver-sb\") pod \"dnsmasq-dns-86dd6887f5-kqgml\" (UID: \"02b5c258-0253-4d75-852b-30c2b26b5598\") " pod="openstack/dnsmasq-dns-86dd6887f5-kqgml" Feb 18 09:18:01 crc kubenswrapper[4556]: I0218 09:18:01.245966 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-d8cc75c64-fwzfq"] Feb 18 09:18:01 crc kubenswrapper[4556]: I0218 09:18:01.248196 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d8cc75c64-fwzfq" Feb 18 09:18:01 crc kubenswrapper[4556]: I0218 09:18:01.252462 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Feb 18 09:18:01 crc kubenswrapper[4556]: I0218 09:18:01.252847 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Feb 18 09:18:01 crc kubenswrapper[4556]: I0218 09:18:01.254433 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-bcc55" Feb 18 09:18:01 crc kubenswrapper[4556]: I0218 09:18:01.256140 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Feb 18 09:18:01 crc kubenswrapper[4556]: I0218 09:18:01.256714 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-d8cc75c64-fwzfq"] Feb 18 09:18:01 crc kubenswrapper[4556]: I0218 09:18:01.263261 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4hph\" (UniqueName: \"kubernetes.io/projected/02b5c258-0253-4d75-852b-30c2b26b5598-kube-api-access-c4hph\") pod \"dnsmasq-dns-86dd6887f5-kqgml\" (UID: \"02b5c258-0253-4d75-852b-30c2b26b5598\") " pod="openstack/dnsmasq-dns-86dd6887f5-kqgml" Feb 18 09:18:01 crc kubenswrapper[4556]: I0218 09:18:01.296391 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12c97a89-6f28-4eab-add6-448279050305" path="/var/lib/kubelet/pods/12c97a89-6f28-4eab-add6-448279050305/volumes" Feb 18 09:18:01 crc kubenswrapper[4556]: I0218 09:18:01.336043 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86dd6887f5-kqgml" Feb 18 09:18:01 crc kubenswrapper[4556]: I0218 09:18:01.341266 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2bb44e10-7532-4173-a140-140b161601a4-config\") pod \"neutron-d8cc75c64-fwzfq\" (UID: \"2bb44e10-7532-4173-a140-140b161601a4\") " pod="openstack/neutron-d8cc75c64-fwzfq" Feb 18 09:18:01 crc kubenswrapper[4556]: I0218 09:18:01.341358 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/2bb44e10-7532-4173-a140-140b161601a4-httpd-config\") pod \"neutron-d8cc75c64-fwzfq\" (UID: \"2bb44e10-7532-4173-a140-140b161601a4\") " pod="openstack/neutron-d8cc75c64-fwzfq" Feb 18 09:18:01 crc kubenswrapper[4556]: I0218 09:18:01.341420 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6c8xm\" (UniqueName: \"kubernetes.io/projected/2bb44e10-7532-4173-a140-140b161601a4-kube-api-access-6c8xm\") pod \"neutron-d8cc75c64-fwzfq\" (UID: \"2bb44e10-7532-4173-a140-140b161601a4\") " pod="openstack/neutron-d8cc75c64-fwzfq" Feb 18 09:18:01 crc kubenswrapper[4556]: I0218 09:18:01.341653 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2bb44e10-7532-4173-a140-140b161601a4-ovndb-tls-certs\") pod \"neutron-d8cc75c64-fwzfq\" (UID: \"2bb44e10-7532-4173-a140-140b161601a4\") " pod="openstack/neutron-d8cc75c64-fwzfq" Feb 18 09:18:01 crc kubenswrapper[4556]: I0218 09:18:01.341912 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bb44e10-7532-4173-a140-140b161601a4-combined-ca-bundle\") pod \"neutron-d8cc75c64-fwzfq\" (UID: \"2bb44e10-7532-4173-a140-140b161601a4\") " pod="openstack/neutron-d8cc75c64-fwzfq" Feb 18 09:18:01 crc kubenswrapper[4556]: I0218 09:18:01.443767 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bb44e10-7532-4173-a140-140b161601a4-combined-ca-bundle\") pod \"neutron-d8cc75c64-fwzfq\" (UID: \"2bb44e10-7532-4173-a140-140b161601a4\") " pod="openstack/neutron-d8cc75c64-fwzfq" Feb 18 09:18:01 crc kubenswrapper[4556]: I0218 09:18:01.443833 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2bb44e10-7532-4173-a140-140b161601a4-config\") pod \"neutron-d8cc75c64-fwzfq\" (UID: \"2bb44e10-7532-4173-a140-140b161601a4\") " pod="openstack/neutron-d8cc75c64-fwzfq" Feb 18 09:18:01 crc kubenswrapper[4556]: I0218 09:18:01.443858 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/2bb44e10-7532-4173-a140-140b161601a4-httpd-config\") pod \"neutron-d8cc75c64-fwzfq\" (UID: \"2bb44e10-7532-4173-a140-140b161601a4\") " pod="openstack/neutron-d8cc75c64-fwzfq" Feb 18 09:18:01 crc kubenswrapper[4556]: I0218 09:18:01.443886 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6c8xm\" (UniqueName: \"kubernetes.io/projected/2bb44e10-7532-4173-a140-140b161601a4-kube-api-access-6c8xm\") pod \"neutron-d8cc75c64-fwzfq\" (UID: \"2bb44e10-7532-4173-a140-140b161601a4\") " pod="openstack/neutron-d8cc75c64-fwzfq" Feb 18 09:18:01 crc kubenswrapper[4556]: I0218 09:18:01.443938 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2bb44e10-7532-4173-a140-140b161601a4-ovndb-tls-certs\") pod \"neutron-d8cc75c64-fwzfq\" (UID: \"2bb44e10-7532-4173-a140-140b161601a4\") " pod="openstack/neutron-d8cc75c64-fwzfq" Feb 18 09:18:01 crc kubenswrapper[4556]: I0218 09:18:01.447266 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/2bb44e10-7532-4173-a140-140b161601a4-httpd-config\") pod \"neutron-d8cc75c64-fwzfq\" (UID: \"2bb44e10-7532-4173-a140-140b161601a4\") " pod="openstack/neutron-d8cc75c64-fwzfq" Feb 18 09:18:01 crc kubenswrapper[4556]: I0218 09:18:01.448103 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2bb44e10-7532-4173-a140-140b161601a4-ovndb-tls-certs\") pod \"neutron-d8cc75c64-fwzfq\" (UID: \"2bb44e10-7532-4173-a140-140b161601a4\") " pod="openstack/neutron-d8cc75c64-fwzfq" Feb 18 09:18:01 crc kubenswrapper[4556]: I0218 09:18:01.449479 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/2bb44e10-7532-4173-a140-140b161601a4-config\") pod \"neutron-d8cc75c64-fwzfq\" (UID: \"2bb44e10-7532-4173-a140-140b161601a4\") " pod="openstack/neutron-d8cc75c64-fwzfq" Feb 18 09:18:01 crc kubenswrapper[4556]: I0218 09:18:01.449999 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bb44e10-7532-4173-a140-140b161601a4-combined-ca-bundle\") pod \"neutron-d8cc75c64-fwzfq\" (UID: \"2bb44e10-7532-4173-a140-140b161601a4\") " pod="openstack/neutron-d8cc75c64-fwzfq" Feb 18 09:18:01 crc kubenswrapper[4556]: I0218 09:18:01.457428 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6c8xm\" (UniqueName: \"kubernetes.io/projected/2bb44e10-7532-4173-a140-140b161601a4-kube-api-access-6c8xm\") pod \"neutron-d8cc75c64-fwzfq\" (UID: \"2bb44e10-7532-4173-a140-140b161601a4\") " pod="openstack/neutron-d8cc75c64-fwzfq" Feb 18 09:18:01 crc kubenswrapper[4556]: I0218 09:18:01.593428 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d8cc75c64-fwzfq" Feb 18 09:18:03 crc kubenswrapper[4556]: I0218 09:18:03.237119 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6fd7956ddf-b647f"] Feb 18 09:18:03 crc kubenswrapper[4556]: I0218 09:18:03.239509 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6fd7956ddf-b647f" Feb 18 09:18:03 crc kubenswrapper[4556]: I0218 09:18:03.244549 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Feb 18 09:18:03 crc kubenswrapper[4556]: I0218 09:18:03.244776 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Feb 18 09:18:03 crc kubenswrapper[4556]: I0218 09:18:03.247769 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6fd7956ddf-b647f"] Feb 18 09:18:03 crc kubenswrapper[4556]: I0218 09:18:03.292625 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b92f250f-e888-4ff5-ae26-9e31c7be3379-internal-tls-certs\") pod \"neutron-6fd7956ddf-b647f\" (UID: \"b92f250f-e888-4ff5-ae26-9e31c7be3379\") " pod="openstack/neutron-6fd7956ddf-b647f" Feb 18 09:18:03 crc kubenswrapper[4556]: I0218 09:18:03.292734 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b92f250f-e888-4ff5-ae26-9e31c7be3379-config\") pod \"neutron-6fd7956ddf-b647f\" (UID: \"b92f250f-e888-4ff5-ae26-9e31c7be3379\") " pod="openstack/neutron-6fd7956ddf-b647f" Feb 18 09:18:03 crc kubenswrapper[4556]: I0218 09:18:03.292801 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b92f250f-e888-4ff5-ae26-9e31c7be3379-public-tls-certs\") pod \"neutron-6fd7956ddf-b647f\" (UID: \"b92f250f-e888-4ff5-ae26-9e31c7be3379\") " pod="openstack/neutron-6fd7956ddf-b647f" Feb 18 09:18:03 crc kubenswrapper[4556]: I0218 09:18:03.292904 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b92f250f-e888-4ff5-ae26-9e31c7be3379-httpd-config\") pod \"neutron-6fd7956ddf-b647f\" (UID: \"b92f250f-e888-4ff5-ae26-9e31c7be3379\") " pod="openstack/neutron-6fd7956ddf-b647f" Feb 18 09:18:03 crc kubenswrapper[4556]: I0218 09:18:03.292940 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b92f250f-e888-4ff5-ae26-9e31c7be3379-combined-ca-bundle\") pod \"neutron-6fd7956ddf-b647f\" (UID: \"b92f250f-e888-4ff5-ae26-9e31c7be3379\") " pod="openstack/neutron-6fd7956ddf-b647f" Feb 18 09:18:03 crc kubenswrapper[4556]: I0218 09:18:03.292999 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czjtk\" (UniqueName: \"kubernetes.io/projected/b92f250f-e888-4ff5-ae26-9e31c7be3379-kube-api-access-czjtk\") pod \"neutron-6fd7956ddf-b647f\" (UID: \"b92f250f-e888-4ff5-ae26-9e31c7be3379\") " pod="openstack/neutron-6fd7956ddf-b647f" Feb 18 09:18:03 crc kubenswrapper[4556]: I0218 09:18:03.293066 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b92f250f-e888-4ff5-ae26-9e31c7be3379-ovndb-tls-certs\") pod \"neutron-6fd7956ddf-b647f\" (UID: \"b92f250f-e888-4ff5-ae26-9e31c7be3379\") " pod="openstack/neutron-6fd7956ddf-b647f" Feb 18 09:18:03 crc kubenswrapper[4556]: I0218 09:18:03.395996 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b92f250f-e888-4ff5-ae26-9e31c7be3379-httpd-config\") pod \"neutron-6fd7956ddf-b647f\" (UID: \"b92f250f-e888-4ff5-ae26-9e31c7be3379\") " pod="openstack/neutron-6fd7956ddf-b647f" Feb 18 09:18:03 crc kubenswrapper[4556]: I0218 09:18:03.396094 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b92f250f-e888-4ff5-ae26-9e31c7be3379-combined-ca-bundle\") pod \"neutron-6fd7956ddf-b647f\" (UID: \"b92f250f-e888-4ff5-ae26-9e31c7be3379\") " pod="openstack/neutron-6fd7956ddf-b647f" Feb 18 09:18:03 crc kubenswrapper[4556]: I0218 09:18:03.396245 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czjtk\" (UniqueName: \"kubernetes.io/projected/b92f250f-e888-4ff5-ae26-9e31c7be3379-kube-api-access-czjtk\") pod \"neutron-6fd7956ddf-b647f\" (UID: \"b92f250f-e888-4ff5-ae26-9e31c7be3379\") " pod="openstack/neutron-6fd7956ddf-b647f" Feb 18 09:18:03 crc kubenswrapper[4556]: I0218 09:18:03.396394 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b92f250f-e888-4ff5-ae26-9e31c7be3379-ovndb-tls-certs\") pod \"neutron-6fd7956ddf-b647f\" (UID: \"b92f250f-e888-4ff5-ae26-9e31c7be3379\") " pod="openstack/neutron-6fd7956ddf-b647f" Feb 18 09:18:03 crc kubenswrapper[4556]: I0218 09:18:03.396527 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b92f250f-e888-4ff5-ae26-9e31c7be3379-internal-tls-certs\") pod \"neutron-6fd7956ddf-b647f\" (UID: \"b92f250f-e888-4ff5-ae26-9e31c7be3379\") " pod="openstack/neutron-6fd7956ddf-b647f" Feb 18 09:18:03 crc kubenswrapper[4556]: I0218 09:18:03.396599 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b92f250f-e888-4ff5-ae26-9e31c7be3379-config\") pod \"neutron-6fd7956ddf-b647f\" (UID: \"b92f250f-e888-4ff5-ae26-9e31c7be3379\") " pod="openstack/neutron-6fd7956ddf-b647f" Feb 18 09:18:03 crc kubenswrapper[4556]: I0218 09:18:03.397036 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b92f250f-e888-4ff5-ae26-9e31c7be3379-public-tls-certs\") pod \"neutron-6fd7956ddf-b647f\" (UID: \"b92f250f-e888-4ff5-ae26-9e31c7be3379\") " pod="openstack/neutron-6fd7956ddf-b647f" Feb 18 09:18:03 crc kubenswrapper[4556]: I0218 09:18:03.398951 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Feb 18 09:18:03 crc kubenswrapper[4556]: I0218 09:18:03.400420 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Feb 18 09:18:03 crc kubenswrapper[4556]: I0218 09:18:03.403522 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b92f250f-e888-4ff5-ae26-9e31c7be3379-httpd-config\") pod \"neutron-6fd7956ddf-b647f\" (UID: \"b92f250f-e888-4ff5-ae26-9e31c7be3379\") " pod="openstack/neutron-6fd7956ddf-b647f" Feb 18 09:18:03 crc kubenswrapper[4556]: I0218 09:18:03.404535 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b92f250f-e888-4ff5-ae26-9e31c7be3379-ovndb-tls-certs\") pod \"neutron-6fd7956ddf-b647f\" (UID: \"b92f250f-e888-4ff5-ae26-9e31c7be3379\") " pod="openstack/neutron-6fd7956ddf-b647f" Feb 18 09:18:03 crc kubenswrapper[4556]: I0218 09:18:03.405070 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/b92f250f-e888-4ff5-ae26-9e31c7be3379-config\") pod \"neutron-6fd7956ddf-b647f\" (UID: \"b92f250f-e888-4ff5-ae26-9e31c7be3379\") " pod="openstack/neutron-6fd7956ddf-b647f" Feb 18 09:18:03 crc kubenswrapper[4556]: I0218 09:18:03.411389 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b92f250f-e888-4ff5-ae26-9e31c7be3379-public-tls-certs\") pod \"neutron-6fd7956ddf-b647f\" (UID: \"b92f250f-e888-4ff5-ae26-9e31c7be3379\") " pod="openstack/neutron-6fd7956ddf-b647f" Feb 18 09:18:03 crc kubenswrapper[4556]: I0218 09:18:03.411862 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b92f250f-e888-4ff5-ae26-9e31c7be3379-combined-ca-bundle\") pod \"neutron-6fd7956ddf-b647f\" (UID: \"b92f250f-e888-4ff5-ae26-9e31c7be3379\") " pod="openstack/neutron-6fd7956ddf-b647f" Feb 18 09:18:03 crc kubenswrapper[4556]: I0218 09:18:03.412823 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b92f250f-e888-4ff5-ae26-9e31c7be3379-internal-tls-certs\") pod \"neutron-6fd7956ddf-b647f\" (UID: \"b92f250f-e888-4ff5-ae26-9e31c7be3379\") " pod="openstack/neutron-6fd7956ddf-b647f" Feb 18 09:18:03 crc kubenswrapper[4556]: I0218 09:18:03.414359 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czjtk\" (UniqueName: \"kubernetes.io/projected/b92f250f-e888-4ff5-ae26-9e31c7be3379-kube-api-access-czjtk\") pod \"neutron-6fd7956ddf-b647f\" (UID: \"b92f250f-e888-4ff5-ae26-9e31c7be3379\") " pod="openstack/neutron-6fd7956ddf-b647f" Feb 18 09:18:03 crc kubenswrapper[4556]: I0218 09:18:03.559605 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6fd7956ddf-b647f" Feb 18 09:18:07 crc kubenswrapper[4556]: E0218 09:18:07.005977 4556 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:a5f8855b2ed00a661ac827cc3908e540ed2327354ac5a1d39491f4507237b4ec" Feb 18 09:18:07 crc kubenswrapper[4556]: E0218 09:18:07.006411 4556 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:a5f8855b2ed00a661ac827cc3908e540ed2327354ac5a1d39491f4507237b4ec,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mss5k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-r4h4k_openstack(8c46304b-1b60-493b-918f-678451f13057): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 18 09:18:07 crc kubenswrapper[4556]: E0218 09:18:07.008213 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-r4h4k" podUID="8c46304b-1b60-493b-918f-678451f13057" Feb 18 09:18:07 crc kubenswrapper[4556]: I0218 09:18:07.108216 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54cc5c87cc-ck95n" Feb 18 09:18:07 crc kubenswrapper[4556]: I0218 09:18:07.185920 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/269247cf-79a0-49a1-8fc9-da82411da8e5-dns-swift-storage-0\") pod \"269247cf-79a0-49a1-8fc9-da82411da8e5\" (UID: \"269247cf-79a0-49a1-8fc9-da82411da8e5\") " Feb 18 09:18:07 crc kubenswrapper[4556]: I0218 09:18:07.186043 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/269247cf-79a0-49a1-8fc9-da82411da8e5-config\") pod \"269247cf-79a0-49a1-8fc9-da82411da8e5\" (UID: \"269247cf-79a0-49a1-8fc9-da82411da8e5\") " Feb 18 09:18:07 crc kubenswrapper[4556]: I0218 09:18:07.186213 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/269247cf-79a0-49a1-8fc9-da82411da8e5-dns-svc\") pod \"269247cf-79a0-49a1-8fc9-da82411da8e5\" (UID: \"269247cf-79a0-49a1-8fc9-da82411da8e5\") " Feb 18 09:18:07 crc kubenswrapper[4556]: I0218 09:18:07.186248 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t7cwx\" (UniqueName: \"kubernetes.io/projected/269247cf-79a0-49a1-8fc9-da82411da8e5-kube-api-access-t7cwx\") pod \"269247cf-79a0-49a1-8fc9-da82411da8e5\" (UID: \"269247cf-79a0-49a1-8fc9-da82411da8e5\") " Feb 18 09:18:07 crc kubenswrapper[4556]: I0218 09:18:07.186357 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/269247cf-79a0-49a1-8fc9-da82411da8e5-ovsdbserver-nb\") pod \"269247cf-79a0-49a1-8fc9-da82411da8e5\" (UID: \"269247cf-79a0-49a1-8fc9-da82411da8e5\") " Feb 18 09:18:07 crc kubenswrapper[4556]: I0218 09:18:07.186388 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/269247cf-79a0-49a1-8fc9-da82411da8e5-ovsdbserver-sb\") pod \"269247cf-79a0-49a1-8fc9-da82411da8e5\" (UID: \"269247cf-79a0-49a1-8fc9-da82411da8e5\") " Feb 18 09:18:07 crc kubenswrapper[4556]: I0218 09:18:07.218319 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/269247cf-79a0-49a1-8fc9-da82411da8e5-kube-api-access-t7cwx" (OuterVolumeSpecName: "kube-api-access-t7cwx") pod "269247cf-79a0-49a1-8fc9-da82411da8e5" (UID: "269247cf-79a0-49a1-8fc9-da82411da8e5"). InnerVolumeSpecName "kube-api-access-t7cwx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:18:07 crc kubenswrapper[4556]: I0218 09:18:07.288682 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t7cwx\" (UniqueName: \"kubernetes.io/projected/269247cf-79a0-49a1-8fc9-da82411da8e5-kube-api-access-t7cwx\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:07 crc kubenswrapper[4556]: I0218 09:18:07.321422 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/269247cf-79a0-49a1-8fc9-da82411da8e5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "269247cf-79a0-49a1-8fc9-da82411da8e5" (UID: "269247cf-79a0-49a1-8fc9-da82411da8e5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:18:07 crc kubenswrapper[4556]: I0218 09:18:07.327496 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/269247cf-79a0-49a1-8fc9-da82411da8e5-config" (OuterVolumeSpecName: "config") pod "269247cf-79a0-49a1-8fc9-da82411da8e5" (UID: "269247cf-79a0-49a1-8fc9-da82411da8e5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:18:07 crc kubenswrapper[4556]: I0218 09:18:07.328717 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/269247cf-79a0-49a1-8fc9-da82411da8e5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "269247cf-79a0-49a1-8fc9-da82411da8e5" (UID: "269247cf-79a0-49a1-8fc9-da82411da8e5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:18:07 crc kubenswrapper[4556]: I0218 09:18:07.334279 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/269247cf-79a0-49a1-8fc9-da82411da8e5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "269247cf-79a0-49a1-8fc9-da82411da8e5" (UID: "269247cf-79a0-49a1-8fc9-da82411da8e5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:18:07 crc kubenswrapper[4556]: I0218 09:18:07.334953 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/269247cf-79a0-49a1-8fc9-da82411da8e5-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "269247cf-79a0-49a1-8fc9-da82411da8e5" (UID: "269247cf-79a0-49a1-8fc9-da82411da8e5"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:18:07 crc kubenswrapper[4556]: I0218 09:18:07.391357 4556 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/269247cf-79a0-49a1-8fc9-da82411da8e5-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:07 crc kubenswrapper[4556]: I0218 09:18:07.391392 4556 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/269247cf-79a0-49a1-8fc9-da82411da8e5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:07 crc kubenswrapper[4556]: I0218 09:18:07.391405 4556 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/269247cf-79a0-49a1-8fc9-da82411da8e5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:07 crc kubenswrapper[4556]: I0218 09:18:07.391417 4556 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/269247cf-79a0-49a1-8fc9-da82411da8e5-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:07 crc kubenswrapper[4556]: I0218 09:18:07.391428 4556 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/269247cf-79a0-49a1-8fc9-da82411da8e5-config\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:07 crc kubenswrapper[4556]: I0218 09:18:07.911656 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54cc5c87cc-ck95n" Feb 18 09:18:07 crc kubenswrapper[4556]: I0218 09:18:07.911624 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54cc5c87cc-ck95n" event={"ID":"269247cf-79a0-49a1-8fc9-da82411da8e5","Type":"ContainerDied","Data":"a23111e48eb9ed7e01c5600fc97cdec0e8c7e2fdd91d09e5fbafbb3f2e281d74"} Feb 18 09:18:07 crc kubenswrapper[4556]: I0218 09:18:07.912190 4556 scope.go:117] "RemoveContainer" containerID="a62979af6d8d1b2e9b80ec0dfe2c502d3c4c4c1c4f03145c6c6e9d54c955a58b" Feb 18 09:18:07 crc kubenswrapper[4556]: E0218 09:18:07.913670 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:a5f8855b2ed00a661ac827cc3908e540ed2327354ac5a1d39491f4507237b4ec\\\"\"" pod="openstack/barbican-db-sync-r4h4k" podUID="8c46304b-1b60-493b-918f-678451f13057" Feb 18 09:18:07 crc kubenswrapper[4556]: I0218 09:18:07.959724 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54cc5c87cc-ck95n"] Feb 18 09:18:07 crc kubenswrapper[4556]: I0218 09:18:07.969712 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-54cc5c87cc-ck95n"] Feb 18 09:18:08 crc kubenswrapper[4556]: E0218 09:18:08.205877 4556 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:3fa6e687aa002b92fedbfe2c1ccaa2906b399c58d17bf9ecece2c4cd69a0210b" Feb 18 09:18:08 crc kubenswrapper[4556]: E0218 09:18:08.206061 4556 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:3fa6e687aa002b92fedbfe2c1ccaa2906b399c58d17bf9ecece2c4cd69a0210b,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-46vwv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-dtq7t_openstack(cb11270a-c754-44b2-b0b1-937cff515ddd): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 18 09:18:08 crc kubenswrapper[4556]: E0218 09:18:08.207332 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-dtq7t" podUID="cb11270a-c754-44b2-b0b1-937cff515ddd" Feb 18 09:18:08 crc kubenswrapper[4556]: I0218 09:18:08.287712 4556 scope.go:117] "RemoveContainer" containerID="6b3748e795909615f30929c23ebb9f851741f83ac5a40113175b1337bf80b42f" Feb 18 09:18:08 crc kubenswrapper[4556]: I0218 09:18:08.760119 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 18 09:18:08 crc kubenswrapper[4556]: I0218 09:18:08.897816 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-85bb8c66f4-fhblh"] Feb 18 09:18:08 crc kubenswrapper[4556]: I0218 09:18:08.920560 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-77cf5996cc-dj8s4" event={"ID":"66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98","Type":"ContainerStarted","Data":"ace36a61edb3ea0808af8dd243149192bdb820e0dff28b907b17b501b9e6c790"} Feb 18 09:18:08 crc kubenswrapper[4556]: I0218 09:18:08.923677 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"40566cc9-f4ba-4201-836a-811e07642dd8","Type":"ContainerStarted","Data":"e61af08817ed81d916b4caff2b0fa6dc025a866fee3c758f4ca4dc8a0e350dfe"} Feb 18 09:18:08 crc kubenswrapper[4556]: I0218 09:18:08.927511 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e0739f72-891b-46e1-bce5-e17c54a407ab","Type":"ContainerStarted","Data":"5efdef38fecec015f3f0924c79231a8c7d9c34252e406c92aefa44ef006d639b"} Feb 18 09:18:08 crc kubenswrapper[4556]: I0218 09:18:08.933840 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79dd98bf77-fdbv7" event={"ID":"c5fec43c-7b17-45d5-91c4-609b6d2ca1b9","Type":"ContainerStarted","Data":"2fed1042675c8bc639ff1bf40020b12a7f5f553be9bb216173f3fbe9c18aa7dd"} Feb 18 09:18:08 crc kubenswrapper[4556]: I0218 09:18:08.941035 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-74c5b6b6c5-q8djs" event={"ID":"50228011-f908-465e-a754-d4b6b626dda9","Type":"ContainerStarted","Data":"7147437f196486d78c3232e03d579055aba6c510b6715c9264fd4d2b04e9089e"} Feb 18 09:18:08 crc kubenswrapper[4556]: E0218 09:18:08.942758 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:3fa6e687aa002b92fedbfe2c1ccaa2906b399c58d17bf9ecece2c4cd69a0210b\\\"\"" pod="openstack/cinder-db-sync-dtq7t" podUID="cb11270a-c754-44b2-b0b1-937cff515ddd" Feb 18 09:18:08 crc kubenswrapper[4556]: W0218 09:18:08.951380 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3354b930_6b00_4aa7_a4bd_97ec410cc863.slice/crio-0ab01d8a9a09a2f53fe5451dc6aba62a035bf8862bbcdd87335a79bca55b4e46 WatchSource:0}: Error finding container 0ab01d8a9a09a2f53fe5451dc6aba62a035bf8862bbcdd87335a79bca55b4e46: Status 404 returned error can't find the container with id 0ab01d8a9a09a2f53fe5451dc6aba62a035bf8862bbcdd87335a79bca55b4e46 Feb 18 09:18:09 crc kubenswrapper[4556]: I0218 09:18:09.002334 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 18 09:18:09 crc kubenswrapper[4556]: W0218 09:18:09.016720 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7dd567df_a88f_4cfd_b465_f949d75ca4a2.slice/crio-ab6fde9a3d76ad1f6265c242c50013db2fb59f906eaa9af459cad071e77c2f0b WatchSource:0}: Error finding container ab6fde9a3d76ad1f6265c242c50013db2fb59f906eaa9af459cad071e77c2f0b: Status 404 returned error can't find the container with id ab6fde9a3d76ad1f6265c242c50013db2fb59f906eaa9af459cad071e77c2f0b Feb 18 09:18:09 crc kubenswrapper[4556]: I0218 09:18:09.058255 4556 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-54cc5c87cc-ck95n" podUID="269247cf-79a0-49a1-8fc9-da82411da8e5" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.123:5353: i/o timeout" Feb 18 09:18:09 crc kubenswrapper[4556]: I0218 09:18:09.074272 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86dd6887f5-kqgml"] Feb 18 09:18:09 crc kubenswrapper[4556]: W0218 09:18:09.085689 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod02b5c258_0253_4d75_852b_30c2b26b5598.slice/crio-b642158754e733cbe4e6d6ca65904149495445382adadd8f335ec5f528572058 WatchSource:0}: Error finding container b642158754e733cbe4e6d6ca65904149495445382adadd8f335ec5f528572058: Status 404 returned error can't find the container with id b642158754e733cbe4e6d6ca65904149495445382adadd8f335ec5f528572058 Feb 18 09:18:09 crc kubenswrapper[4556]: I0218 09:18:09.105431 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-595566bb8b-6b5h6"] Feb 18 09:18:09 crc kubenswrapper[4556]: I0218 09:18:09.122744 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-brdr4"] Feb 18 09:18:09 crc kubenswrapper[4556]: W0218 09:18:09.136326 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda6347657_1897_458a_9a02_36d8655525f2.slice/crio-edbc28809a341505414bc97051fe0667a2952b1a20bf410da5bfbc5c0c2eeb16 WatchSource:0}: Error finding container edbc28809a341505414bc97051fe0667a2952b1a20bf410da5bfbc5c0c2eeb16: Status 404 returned error can't find the container with id edbc28809a341505414bc97051fe0667a2952b1a20bf410da5bfbc5c0c2eeb16 Feb 18 09:18:09 crc kubenswrapper[4556]: I0218 09:18:09.142258 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Feb 18 09:18:09 crc kubenswrapper[4556]: I0218 09:18:09.184535 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6fd7956ddf-b647f"] Feb 18 09:18:09 crc kubenswrapper[4556]: I0218 09:18:09.292562 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="269247cf-79a0-49a1-8fc9-da82411da8e5" path="/var/lib/kubelet/pods/269247cf-79a0-49a1-8fc9-da82411da8e5/volumes" Feb 18 09:18:09 crc kubenswrapper[4556]: I0218 09:18:09.952391 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-d8cc75c64-fwzfq"] Feb 18 09:18:09 crc kubenswrapper[4556]: I0218 09:18:09.959995 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6fd7956ddf-b647f" event={"ID":"b92f250f-e888-4ff5-ae26-9e31c7be3379","Type":"ContainerStarted","Data":"cc6725c2f27c13f0413ab0e4d8de88e810bf5e3b4b491a0eedf920bb5855de30"} Feb 18 09:18:09 crc kubenswrapper[4556]: I0218 09:18:09.960028 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6fd7956ddf-b647f" event={"ID":"b92f250f-e888-4ff5-ae26-9e31c7be3379","Type":"ContainerStarted","Data":"20b200abdb599e8ec5463ea4f53a30e71450839bf07dba61fec5a4eb01b348dd"} Feb 18 09:18:09 crc kubenswrapper[4556]: I0218 09:18:09.960039 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6fd7956ddf-b647f" event={"ID":"b92f250f-e888-4ff5-ae26-9e31c7be3379","Type":"ContainerStarted","Data":"4cf48eebb92b1e8d95ce8b8111c230c3a42d8ec80542347be535fc1a082eee1b"} Feb 18 09:18:09 crc kubenswrapper[4556]: I0218 09:18:09.960290 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-6fd7956ddf-b647f" Feb 18 09:18:09 crc kubenswrapper[4556]: I0218 09:18:09.974392 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"40566cc9-f4ba-4201-836a-811e07642dd8","Type":"ContainerStarted","Data":"a9e0696232cbfb04625e8ba6e496c429dc487bc7b0e719cc31f4d45c5c99ac8a"} Feb 18 09:18:09 crc kubenswrapper[4556]: I0218 09:18:09.978847 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7dd567df-a88f-4cfd-b465-f949d75ca4a2","Type":"ContainerStarted","Data":"3be8265b5007dcf42927d8cd64e27c3f9f427cf821edbc8b0caa611d3753f4bf"} Feb 18 09:18:09 crc kubenswrapper[4556]: I0218 09:18:09.978932 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7dd567df-a88f-4cfd-b465-f949d75ca4a2","Type":"ContainerStarted","Data":"ab6fde9a3d76ad1f6265c242c50013db2fb59f906eaa9af459cad071e77c2f0b"} Feb 18 09:18:09 crc kubenswrapper[4556]: I0218 09:18:09.982120 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-6fd7956ddf-b647f" podStartSLOduration=6.982110733 podStartE2EDuration="6.982110733s" podCreationTimestamp="2026-02-18 09:18:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:18:09.97298842 +0000 UTC m=+846.989949400" watchObservedRunningTime="2026-02-18 09:18:09.982110733 +0000 UTC m=+846.999071713" Feb 18 09:18:09 crc kubenswrapper[4556]: I0218 09:18:09.987424 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79dd98bf77-fdbv7" event={"ID":"c5fec43c-7b17-45d5-91c4-609b6d2ca1b9","Type":"ContainerStarted","Data":"a6e43edd5803076f4c76dc543e9ef0bdba7eda69af36b2cf7df94a0200eacf09"} Feb 18 09:18:09 crc kubenswrapper[4556]: I0218 09:18:09.987528 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-79dd98bf77-fdbv7" podUID="c5fec43c-7b17-45d5-91c4-609b6d2ca1b9" containerName="horizon-log" containerID="cri-o://2fed1042675c8bc639ff1bf40020b12a7f5f553be9bb216173f3fbe9c18aa7dd" gracePeriod=30 Feb 18 09:18:09 crc kubenswrapper[4556]: I0218 09:18:09.987937 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-79dd98bf77-fdbv7" podUID="c5fec43c-7b17-45d5-91c4-609b6d2ca1b9" containerName="horizon" containerID="cri-o://a6e43edd5803076f4c76dc543e9ef0bdba7eda69af36b2cf7df94a0200eacf09" gracePeriod=30 Feb 18 09:18:09 crc kubenswrapper[4556]: I0218 09:18:09.993225 4556 generic.go:334] "Generic (PLEG): container finished" podID="02b5c258-0253-4d75-852b-30c2b26b5598" containerID="1888148871a332a7d249fe69193657e3adb1614a874e247e986fe92585d9afa7" exitCode=0 Feb 18 09:18:09 crc kubenswrapper[4556]: I0218 09:18:09.993300 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86dd6887f5-kqgml" event={"ID":"02b5c258-0253-4d75-852b-30c2b26b5598","Type":"ContainerDied","Data":"1888148871a332a7d249fe69193657e3adb1614a874e247e986fe92585d9afa7"} Feb 18 09:18:09 crc kubenswrapper[4556]: I0218 09:18:09.993344 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86dd6887f5-kqgml" event={"ID":"02b5c258-0253-4d75-852b-30c2b26b5598","Type":"ContainerStarted","Data":"b642158754e733cbe4e6d6ca65904149495445382adadd8f335ec5f528572058"} Feb 18 09:18:10 crc kubenswrapper[4556]: I0218 09:18:10.003881 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-595566bb8b-6b5h6" event={"ID":"30fffa0b-72a5-4bbe-a1fb-fa8b26d1decf","Type":"ContainerStarted","Data":"89304d6e4b1965aac81a12052a467e21ae879bb159a75620632cf44afc108d01"} Feb 18 09:18:10 crc kubenswrapper[4556]: I0218 09:18:10.003929 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-595566bb8b-6b5h6" event={"ID":"30fffa0b-72a5-4bbe-a1fb-fa8b26d1decf","Type":"ContainerStarted","Data":"9846e1f66a632ec1a7d8ecce792645bd367d9138eb208cd9db73a58dc672b061"} Feb 18 09:18:10 crc kubenswrapper[4556]: I0218 09:18:10.003944 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-595566bb8b-6b5h6" event={"ID":"30fffa0b-72a5-4bbe-a1fb-fa8b26d1decf","Type":"ContainerStarted","Data":"da82c9db64901a08ff744eeb9accc33074a717bc71d934eec313a56732bc13b5"} Feb 18 09:18:10 crc kubenswrapper[4556]: I0218 09:18:10.007285 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-85bb8c66f4-fhblh" event={"ID":"3354b930-6b00-4aa7-a4bd-97ec410cc863","Type":"ContainerStarted","Data":"4423fda867205940e352d08e4c633d2caee4bc72b78e6280dd5fd52606caa4dd"} Feb 18 09:18:10 crc kubenswrapper[4556]: I0218 09:18:10.007318 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-85bb8c66f4-fhblh" event={"ID":"3354b930-6b00-4aa7-a4bd-97ec410cc863","Type":"ContainerStarted","Data":"c1df47a378b25fad3361d8a29971e5ef0d47d8783acc451f619234849683c0f1"} Feb 18 09:18:10 crc kubenswrapper[4556]: I0218 09:18:10.007328 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-85bb8c66f4-fhblh" event={"ID":"3354b930-6b00-4aa7-a4bd-97ec410cc863","Type":"ContainerStarted","Data":"0ab01d8a9a09a2f53fe5451dc6aba62a035bf8862bbcdd87335a79bca55b4e46"} Feb 18 09:18:10 crc kubenswrapper[4556]: I0218 09:18:10.009471 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-74c5b6b6c5-q8djs" event={"ID":"50228011-f908-465e-a754-d4b6b626dda9","Type":"ContainerStarted","Data":"c354f6bc02d120ffecda388ac1244fdc7228debcaed10748429b9072f8e42df2"} Feb 18 09:18:10 crc kubenswrapper[4556]: I0218 09:18:10.009619 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-74c5b6b6c5-q8djs" podUID="50228011-f908-465e-a754-d4b6b626dda9" containerName="horizon-log" containerID="cri-o://7147437f196486d78c3232e03d579055aba6c510b6715c9264fd4d2b04e9089e" gracePeriod=30 Feb 18 09:18:10 crc kubenswrapper[4556]: I0218 09:18:10.009902 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-74c5b6b6c5-q8djs" podUID="50228011-f908-465e-a754-d4b6b626dda9" containerName="horizon" containerID="cri-o://c354f6bc02d120ffecda388ac1244fdc7228debcaed10748429b9072f8e42df2" gracePeriod=30 Feb 18 09:18:10 crc kubenswrapper[4556]: I0218 09:18:10.011998 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-79dd98bf77-fdbv7" podStartSLOduration=3.16175984 podStartE2EDuration="24.011985077s" podCreationTimestamp="2026-02-18 09:17:46 +0000 UTC" firstStartedPulling="2026-02-18 09:17:47.43777684 +0000 UTC m=+824.454737820" lastFinishedPulling="2026-02-18 09:18:08.288002077 +0000 UTC m=+845.304963057" observedRunningTime="2026-02-18 09:18:10.005560583 +0000 UTC m=+847.022521563" watchObservedRunningTime="2026-02-18 09:18:10.011985077 +0000 UTC m=+847.028946056" Feb 18 09:18:10 crc kubenswrapper[4556]: I0218 09:18:10.015314 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-brdr4" event={"ID":"a6347657-1897-458a-9a02-36d8655525f2","Type":"ContainerStarted","Data":"647910ecbb2afa2329d622e6150e7766438dd9b5e0ec136356c23c9ef4cfc82a"} Feb 18 09:18:10 crc kubenswrapper[4556]: I0218 09:18:10.015341 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-brdr4" event={"ID":"a6347657-1897-458a-9a02-36d8655525f2","Type":"ContainerStarted","Data":"edbc28809a341505414bc97051fe0667a2952b1a20bf410da5bfbc5c0c2eeb16"} Feb 18 09:18:10 crc kubenswrapper[4556]: I0218 09:18:10.020614 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-77cf5996cc-dj8s4" event={"ID":"66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98","Type":"ContainerStarted","Data":"2973401c6ebe4b90a1471e53a5013aae7d75364f2830c2e5cd584b5f74812ad2"} Feb 18 09:18:10 crc kubenswrapper[4556]: I0218 09:18:10.020736 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-77cf5996cc-dj8s4" podUID="66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98" containerName="horizon-log" containerID="cri-o://ace36a61edb3ea0808af8dd243149192bdb820e0dff28b907b17b501b9e6c790" gracePeriod=30 Feb 18 09:18:10 crc kubenswrapper[4556]: I0218 09:18:10.020857 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-77cf5996cc-dj8s4" podUID="66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98" containerName="horizon" containerID="cri-o://2973401c6ebe4b90a1471e53a5013aae7d75364f2830c2e5cd584b5f74812ad2" gracePeriod=30 Feb 18 09:18:10 crc kubenswrapper[4556]: I0218 09:18:10.072839 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-595566bb8b-6b5h6" podStartSLOduration=17.072823388 podStartE2EDuration="17.072823388s" podCreationTimestamp="2026-02-18 09:17:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:18:10.050583301 +0000 UTC m=+847.067544281" watchObservedRunningTime="2026-02-18 09:18:10.072823388 +0000 UTC m=+847.089784368" Feb 18 09:18:10 crc kubenswrapper[4556]: I0218 09:18:10.075184 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-77cf5996cc-dj8s4" podStartSLOduration=2.922563689 podStartE2EDuration="26.07517741s" podCreationTimestamp="2026-02-18 09:17:44 +0000 UTC" firstStartedPulling="2026-02-18 09:17:45.04629883 +0000 UTC m=+822.063259811" lastFinishedPulling="2026-02-18 09:18:08.198912551 +0000 UTC m=+845.215873532" observedRunningTime="2026-02-18 09:18:10.067761667 +0000 UTC m=+847.084722647" watchObservedRunningTime="2026-02-18 09:18:10.07517741 +0000 UTC m=+847.092138380" Feb 18 09:18:10 crc kubenswrapper[4556]: I0218 09:18:10.091042 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-74c5b6b6c5-q8djs" podStartSLOduration=3.164477045 podStartE2EDuration="26.091021036s" podCreationTimestamp="2026-02-18 09:17:44 +0000 UTC" firstStartedPulling="2026-02-18 09:17:45.382490171 +0000 UTC m=+822.399451151" lastFinishedPulling="2026-02-18 09:18:08.309034162 +0000 UTC m=+845.325995142" observedRunningTime="2026-02-18 09:18:10.085567256 +0000 UTC m=+847.102528236" watchObservedRunningTime="2026-02-18 09:18:10.091021036 +0000 UTC m=+847.107982007" Feb 18 09:18:10 crc kubenswrapper[4556]: I0218 09:18:10.104041 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-brdr4" podStartSLOduration=11.104033181 podStartE2EDuration="11.104033181s" podCreationTimestamp="2026-02-18 09:17:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:18:10.099469669 +0000 UTC m=+847.116430650" watchObservedRunningTime="2026-02-18 09:18:10.104033181 +0000 UTC m=+847.120994161" Feb 18 09:18:10 crc kubenswrapper[4556]: I0218 09:18:10.131135 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-85bb8c66f4-fhblh" podStartSLOduration=17.131114092 podStartE2EDuration="17.131114092s" podCreationTimestamp="2026-02-18 09:17:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:18:10.123535711 +0000 UTC m=+847.140496692" watchObservedRunningTime="2026-02-18 09:18:10.131114092 +0000 UTC m=+847.148075073" Feb 18 09:18:11 crc kubenswrapper[4556]: W0218 09:18:11.056687 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2bb44e10_7532_4173_a140_140b161601a4.slice/crio-f097d1897907d95895e9c90beb30d488dd020fb1b4fe76bc9601482a0968b723 WatchSource:0}: Error finding container f097d1897907d95895e9c90beb30d488dd020fb1b4fe76bc9601482a0968b723: Status 404 returned error can't find the container with id f097d1897907d95895e9c90beb30d488dd020fb1b4fe76bc9601482a0968b723 Feb 18 09:18:11 crc kubenswrapper[4556]: I0218 09:18:11.057890 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"40566cc9-f4ba-4201-836a-811e07642dd8","Type":"ContainerStarted","Data":"51fc65cc3d60b91bd3881006a0a7146830369a338802a9c1a6590e6d57e56956"} Feb 18 09:18:11 crc kubenswrapper[4556]: I0218 09:18:11.058326 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="40566cc9-f4ba-4201-836a-811e07642dd8" containerName="glance-log" containerID="cri-o://a9e0696232cbfb04625e8ba6e496c429dc487bc7b0e719cc31f4d45c5c99ac8a" gracePeriod=30 Feb 18 09:18:11 crc kubenswrapper[4556]: I0218 09:18:11.058440 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="40566cc9-f4ba-4201-836a-811e07642dd8" containerName="glance-httpd" containerID="cri-o://51fc65cc3d60b91bd3881006a0a7146830369a338802a9c1a6590e6d57e56956" gracePeriod=30 Feb 18 09:18:11 crc kubenswrapper[4556]: I0218 09:18:11.716056 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 18 09:18:11 crc kubenswrapper[4556]: I0218 09:18:11.822438 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/40566cc9-f4ba-4201-836a-811e07642dd8-internal-tls-certs\") pod \"40566cc9-f4ba-4201-836a-811e07642dd8\" (UID: \"40566cc9-f4ba-4201-836a-811e07642dd8\") " Feb 18 09:18:11 crc kubenswrapper[4556]: I0218 09:18:11.822514 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/40566cc9-f4ba-4201-836a-811e07642dd8-scripts\") pod \"40566cc9-f4ba-4201-836a-811e07642dd8\" (UID: \"40566cc9-f4ba-4201-836a-811e07642dd8\") " Feb 18 09:18:11 crc kubenswrapper[4556]: I0218 09:18:11.823287 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"40566cc9-f4ba-4201-836a-811e07642dd8\" (UID: \"40566cc9-f4ba-4201-836a-811e07642dd8\") " Feb 18 09:18:11 crc kubenswrapper[4556]: I0218 09:18:11.823323 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pdxkd\" (UniqueName: \"kubernetes.io/projected/40566cc9-f4ba-4201-836a-811e07642dd8-kube-api-access-pdxkd\") pod \"40566cc9-f4ba-4201-836a-811e07642dd8\" (UID: \"40566cc9-f4ba-4201-836a-811e07642dd8\") " Feb 18 09:18:11 crc kubenswrapper[4556]: I0218 09:18:11.823441 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/40566cc9-f4ba-4201-836a-811e07642dd8-logs\") pod \"40566cc9-f4ba-4201-836a-811e07642dd8\" (UID: \"40566cc9-f4ba-4201-836a-811e07642dd8\") " Feb 18 09:18:11 crc kubenswrapper[4556]: I0218 09:18:11.823488 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/40566cc9-f4ba-4201-836a-811e07642dd8-httpd-run\") pod \"40566cc9-f4ba-4201-836a-811e07642dd8\" (UID: \"40566cc9-f4ba-4201-836a-811e07642dd8\") " Feb 18 09:18:11 crc kubenswrapper[4556]: I0218 09:18:11.823544 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40566cc9-f4ba-4201-836a-811e07642dd8-combined-ca-bundle\") pod \"40566cc9-f4ba-4201-836a-811e07642dd8\" (UID: \"40566cc9-f4ba-4201-836a-811e07642dd8\") " Feb 18 09:18:11 crc kubenswrapper[4556]: I0218 09:18:11.823619 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40566cc9-f4ba-4201-836a-811e07642dd8-config-data\") pod \"40566cc9-f4ba-4201-836a-811e07642dd8\" (UID: \"40566cc9-f4ba-4201-836a-811e07642dd8\") " Feb 18 09:18:11 crc kubenswrapper[4556]: I0218 09:18:11.824239 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40566cc9-f4ba-4201-836a-811e07642dd8-logs" (OuterVolumeSpecName: "logs") pod "40566cc9-f4ba-4201-836a-811e07642dd8" (UID: "40566cc9-f4ba-4201-836a-811e07642dd8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:18:11 crc kubenswrapper[4556]: I0218 09:18:11.824825 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40566cc9-f4ba-4201-836a-811e07642dd8-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "40566cc9-f4ba-4201-836a-811e07642dd8" (UID: "40566cc9-f4ba-4201-836a-811e07642dd8"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:18:11 crc kubenswrapper[4556]: I0218 09:18:11.834325 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40566cc9-f4ba-4201-836a-811e07642dd8-kube-api-access-pdxkd" (OuterVolumeSpecName: "kube-api-access-pdxkd") pod "40566cc9-f4ba-4201-836a-811e07642dd8" (UID: "40566cc9-f4ba-4201-836a-811e07642dd8"). InnerVolumeSpecName "kube-api-access-pdxkd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:18:11 crc kubenswrapper[4556]: I0218 09:18:11.849501 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40566cc9-f4ba-4201-836a-811e07642dd8-scripts" (OuterVolumeSpecName: "scripts") pod "40566cc9-f4ba-4201-836a-811e07642dd8" (UID: "40566cc9-f4ba-4201-836a-811e07642dd8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:18:11 crc kubenswrapper[4556]: I0218 09:18:11.857699 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "40566cc9-f4ba-4201-836a-811e07642dd8" (UID: "40566cc9-f4ba-4201-836a-811e07642dd8"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 18 09:18:11 crc kubenswrapper[4556]: I0218 09:18:11.881142 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40566cc9-f4ba-4201-836a-811e07642dd8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "40566cc9-f4ba-4201-836a-811e07642dd8" (UID: "40566cc9-f4ba-4201-836a-811e07642dd8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:18:11 crc kubenswrapper[4556]: I0218 09:18:11.922547 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40566cc9-f4ba-4201-836a-811e07642dd8-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "40566cc9-f4ba-4201-836a-811e07642dd8" (UID: "40566cc9-f4ba-4201-836a-811e07642dd8"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:18:11 crc kubenswrapper[4556]: I0218 09:18:11.925671 4556 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Feb 18 09:18:11 crc kubenswrapper[4556]: I0218 09:18:11.925697 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pdxkd\" (UniqueName: \"kubernetes.io/projected/40566cc9-f4ba-4201-836a-811e07642dd8-kube-api-access-pdxkd\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:11 crc kubenswrapper[4556]: I0218 09:18:11.925709 4556 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/40566cc9-f4ba-4201-836a-811e07642dd8-logs\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:11 crc kubenswrapper[4556]: I0218 09:18:11.925719 4556 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/40566cc9-f4ba-4201-836a-811e07642dd8-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:11 crc kubenswrapper[4556]: I0218 09:18:11.925727 4556 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40566cc9-f4ba-4201-836a-811e07642dd8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:11 crc kubenswrapper[4556]: I0218 09:18:11.925734 4556 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/40566cc9-f4ba-4201-836a-811e07642dd8-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:11 crc kubenswrapper[4556]: I0218 09:18:11.925742 4556 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/40566cc9-f4ba-4201-836a-811e07642dd8-scripts\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:11 crc kubenswrapper[4556]: I0218 09:18:11.926403 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40566cc9-f4ba-4201-836a-811e07642dd8-config-data" (OuterVolumeSpecName: "config-data") pod "40566cc9-f4ba-4201-836a-811e07642dd8" (UID: "40566cc9-f4ba-4201-836a-811e07642dd8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:18:11 crc kubenswrapper[4556]: I0218 09:18:11.939854 4556 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.027195 4556 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.027228 4556 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40566cc9-f4ba-4201-836a-811e07642dd8-config-data\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.071335 4556 generic.go:334] "Generic (PLEG): container finished" podID="40566cc9-f4ba-4201-836a-811e07642dd8" containerID="51fc65cc3d60b91bd3881006a0a7146830369a338802a9c1a6590e6d57e56956" exitCode=0 Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.071604 4556 generic.go:334] "Generic (PLEG): container finished" podID="40566cc9-f4ba-4201-836a-811e07642dd8" containerID="a9e0696232cbfb04625e8ba6e496c429dc487bc7b0e719cc31f4d45c5c99ac8a" exitCode=143 Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.071413 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"40566cc9-f4ba-4201-836a-811e07642dd8","Type":"ContainerDied","Data":"51fc65cc3d60b91bd3881006a0a7146830369a338802a9c1a6590e6d57e56956"} Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.071448 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.071688 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"40566cc9-f4ba-4201-836a-811e07642dd8","Type":"ContainerDied","Data":"a9e0696232cbfb04625e8ba6e496c429dc487bc7b0e719cc31f4d45c5c99ac8a"} Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.071725 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"40566cc9-f4ba-4201-836a-811e07642dd8","Type":"ContainerDied","Data":"e61af08817ed81d916b4caff2b0fa6dc025a866fee3c758f4ca4dc8a0e350dfe"} Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.071756 4556 scope.go:117] "RemoveContainer" containerID="51fc65cc3d60b91bd3881006a0a7146830369a338802a9c1a6590e6d57e56956" Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.074442 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d8cc75c64-fwzfq" event={"ID":"2bb44e10-7532-4173-a140-140b161601a4","Type":"ContainerStarted","Data":"4975d3e5e950a678a1bad9398685c9357404c16df42f27b08a96ae18291c4230"} Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.074503 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d8cc75c64-fwzfq" event={"ID":"2bb44e10-7532-4173-a140-140b161601a4","Type":"ContainerStarted","Data":"0d354e35ff2cfc3bc438333dde43e23a4e77b08fcb22a13609433bf275ed1538"} Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.074516 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d8cc75c64-fwzfq" event={"ID":"2bb44e10-7532-4173-a140-140b161601a4","Type":"ContainerStarted","Data":"f097d1897907d95895e9c90beb30d488dd020fb1b4fe76bc9601482a0968b723"} Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.074566 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-d8cc75c64-fwzfq" Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.077195 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e0739f72-891b-46e1-bce5-e17c54a407ab","Type":"ContainerStarted","Data":"f644379761b7c18c34aaa29cf6ae6f5433a7bb623d8ff1f7d0fd70f9539f4a9f"} Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.081917 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7dd567df-a88f-4cfd-b465-f949d75ca4a2","Type":"ContainerStarted","Data":"4512175c8db79a54e8326671deb8adadf5806321eae7ed88c050aba63d66aa9d"} Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.082166 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="7dd567df-a88f-4cfd-b465-f949d75ca4a2" containerName="glance-log" containerID="cri-o://3be8265b5007dcf42927d8cd64e27c3f9f427cf821edbc8b0caa611d3753f4bf" gracePeriod=30 Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.084188 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="7dd567df-a88f-4cfd-b465-f949d75ca4a2" containerName="glance-httpd" containerID="cri-o://4512175c8db79a54e8326671deb8adadf5806321eae7ed88c050aba63d66aa9d" gracePeriod=30 Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.092948 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86dd6887f5-kqgml" event={"ID":"02b5c258-0253-4d75-852b-30c2b26b5598","Type":"ContainerStarted","Data":"0da42c5457683b62830f1b6e7cd6f3910bea6f7d2bc6851bc77ed7aa339d0f0d"} Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.093167 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86dd6887f5-kqgml" Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.098454 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-d8cc75c64-fwzfq" podStartSLOduration=11.098439417 podStartE2EDuration="11.098439417s" podCreationTimestamp="2026-02-18 09:18:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:18:12.096600366 +0000 UTC m=+849.113561347" watchObservedRunningTime="2026-02-18 09:18:12.098439417 +0000 UTC m=+849.115400397" Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.123249 4556 scope.go:117] "RemoveContainer" containerID="a9e0696232cbfb04625e8ba6e496c429dc487bc7b0e719cc31f4d45c5c99ac8a" Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.131165 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.145766 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.158740 4556 scope.go:117] "RemoveContainer" containerID="51fc65cc3d60b91bd3881006a0a7146830369a338802a9c1a6590e6d57e56956" Feb 18 09:18:12 crc kubenswrapper[4556]: E0218 09:18:12.163217 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51fc65cc3d60b91bd3881006a0a7146830369a338802a9c1a6590e6d57e56956\": container with ID starting with 51fc65cc3d60b91bd3881006a0a7146830369a338802a9c1a6590e6d57e56956 not found: ID does not exist" containerID="51fc65cc3d60b91bd3881006a0a7146830369a338802a9c1a6590e6d57e56956" Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.163257 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51fc65cc3d60b91bd3881006a0a7146830369a338802a9c1a6590e6d57e56956"} err="failed to get container status \"51fc65cc3d60b91bd3881006a0a7146830369a338802a9c1a6590e6d57e56956\": rpc error: code = NotFound desc = could not find container \"51fc65cc3d60b91bd3881006a0a7146830369a338802a9c1a6590e6d57e56956\": container with ID starting with 51fc65cc3d60b91bd3881006a0a7146830369a338802a9c1a6590e6d57e56956 not found: ID does not exist" Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.163282 4556 scope.go:117] "RemoveContainer" containerID="a9e0696232cbfb04625e8ba6e496c429dc487bc7b0e719cc31f4d45c5c99ac8a" Feb 18 09:18:12 crc kubenswrapper[4556]: E0218 09:18:12.165206 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9e0696232cbfb04625e8ba6e496c429dc487bc7b0e719cc31f4d45c5c99ac8a\": container with ID starting with a9e0696232cbfb04625e8ba6e496c429dc487bc7b0e719cc31f4d45c5c99ac8a not found: ID does not exist" containerID="a9e0696232cbfb04625e8ba6e496c429dc487bc7b0e719cc31f4d45c5c99ac8a" Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.165254 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9e0696232cbfb04625e8ba6e496c429dc487bc7b0e719cc31f4d45c5c99ac8a"} err="failed to get container status \"a9e0696232cbfb04625e8ba6e496c429dc487bc7b0e719cc31f4d45c5c99ac8a\": rpc error: code = NotFound desc = could not find container \"a9e0696232cbfb04625e8ba6e496c429dc487bc7b0e719cc31f4d45c5c99ac8a\": container with ID starting with a9e0696232cbfb04625e8ba6e496c429dc487bc7b0e719cc31f4d45c5c99ac8a not found: ID does not exist" Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.165290 4556 scope.go:117] "RemoveContainer" containerID="51fc65cc3d60b91bd3881006a0a7146830369a338802a9c1a6590e6d57e56956" Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.174071 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51fc65cc3d60b91bd3881006a0a7146830369a338802a9c1a6590e6d57e56956"} err="failed to get container status \"51fc65cc3d60b91bd3881006a0a7146830369a338802a9c1a6590e6d57e56956\": rpc error: code = NotFound desc = could not find container \"51fc65cc3d60b91bd3881006a0a7146830369a338802a9c1a6590e6d57e56956\": container with ID starting with 51fc65cc3d60b91bd3881006a0a7146830369a338802a9c1a6590e6d57e56956 not found: ID does not exist" Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.174102 4556 scope.go:117] "RemoveContainer" containerID="a9e0696232cbfb04625e8ba6e496c429dc487bc7b0e719cc31f4d45c5c99ac8a" Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.174721 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9e0696232cbfb04625e8ba6e496c429dc487bc7b0e719cc31f4d45c5c99ac8a"} err="failed to get container status \"a9e0696232cbfb04625e8ba6e496c429dc487bc7b0e719cc31f4d45c5c99ac8a\": rpc error: code = NotFound desc = could not find container \"a9e0696232cbfb04625e8ba6e496c429dc487bc7b0e719cc31f4d45c5c99ac8a\": container with ID starting with a9e0696232cbfb04625e8ba6e496c429dc487bc7b0e719cc31f4d45c5c99ac8a not found: ID does not exist" Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.194205 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 18 09:18:12 crc kubenswrapper[4556]: E0218 09:18:12.194665 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="269247cf-79a0-49a1-8fc9-da82411da8e5" containerName="init" Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.194684 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="269247cf-79a0-49a1-8fc9-da82411da8e5" containerName="init" Feb 18 09:18:12 crc kubenswrapper[4556]: E0218 09:18:12.194721 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="269247cf-79a0-49a1-8fc9-da82411da8e5" containerName="dnsmasq-dns" Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.194726 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="269247cf-79a0-49a1-8fc9-da82411da8e5" containerName="dnsmasq-dns" Feb 18 09:18:12 crc kubenswrapper[4556]: E0218 09:18:12.194739 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40566cc9-f4ba-4201-836a-811e07642dd8" containerName="glance-log" Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.194748 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="40566cc9-f4ba-4201-836a-811e07642dd8" containerName="glance-log" Feb 18 09:18:12 crc kubenswrapper[4556]: E0218 09:18:12.194759 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40566cc9-f4ba-4201-836a-811e07642dd8" containerName="glance-httpd" Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.194764 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="40566cc9-f4ba-4201-836a-811e07642dd8" containerName="glance-httpd" Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.194969 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="269247cf-79a0-49a1-8fc9-da82411da8e5" containerName="dnsmasq-dns" Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.194988 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="40566cc9-f4ba-4201-836a-811e07642dd8" containerName="glance-log" Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.195019 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="40566cc9-f4ba-4201-836a-811e07642dd8" containerName="glance-httpd" Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.208728 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=24.208708394 podStartE2EDuration="24.208708394s" podCreationTimestamp="2026-02-18 09:17:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:18:12.159572063 +0000 UTC m=+849.176533034" watchObservedRunningTime="2026-02-18 09:18:12.208708394 +0000 UTC m=+849.225669374" Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.209017 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.211526 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.211785 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.229222 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.234383 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86dd6887f5-kqgml" podStartSLOduration=12.234365358 podStartE2EDuration="12.234365358s" podCreationTimestamp="2026-02-18 09:18:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:18:12.187981602 +0000 UTC m=+849.204942582" watchObservedRunningTime="2026-02-18 09:18:12.234365358 +0000 UTC m=+849.251326338" Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.332653 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/35011d99-3341-4081-9733-a6f42f69a6d2-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"35011d99-3341-4081-9733-a6f42f69a6d2\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.332735 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35011d99-3341-4081-9733-a6f42f69a6d2-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"35011d99-3341-4081-9733-a6f42f69a6d2\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.332760 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35011d99-3341-4081-9733-a6f42f69a6d2-config-data\") pod \"glance-default-internal-api-0\" (UID: \"35011d99-3341-4081-9733-a6f42f69a6d2\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.332793 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35011d99-3341-4081-9733-a6f42f69a6d2-scripts\") pod \"glance-default-internal-api-0\" (UID: \"35011d99-3341-4081-9733-a6f42f69a6d2\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.332833 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"35011d99-3341-4081-9733-a6f42f69a6d2\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.332877 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/35011d99-3341-4081-9733-a6f42f69a6d2-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"35011d99-3341-4081-9733-a6f42f69a6d2\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.332893 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/35011d99-3341-4081-9733-a6f42f69a6d2-logs\") pod \"glance-default-internal-api-0\" (UID: \"35011d99-3341-4081-9733-a6f42f69a6d2\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.332913 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h98ck\" (UniqueName: \"kubernetes.io/projected/35011d99-3341-4081-9733-a6f42f69a6d2-kube-api-access-h98ck\") pod \"glance-default-internal-api-0\" (UID: \"35011d99-3341-4081-9733-a6f42f69a6d2\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.435143 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35011d99-3341-4081-9733-a6f42f69a6d2-config-data\") pod \"glance-default-internal-api-0\" (UID: \"35011d99-3341-4081-9733-a6f42f69a6d2\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.435268 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35011d99-3341-4081-9733-a6f42f69a6d2-scripts\") pod \"glance-default-internal-api-0\" (UID: \"35011d99-3341-4081-9733-a6f42f69a6d2\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.435363 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"35011d99-3341-4081-9733-a6f42f69a6d2\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.435494 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/35011d99-3341-4081-9733-a6f42f69a6d2-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"35011d99-3341-4081-9733-a6f42f69a6d2\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.435540 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/35011d99-3341-4081-9733-a6f42f69a6d2-logs\") pod \"glance-default-internal-api-0\" (UID: \"35011d99-3341-4081-9733-a6f42f69a6d2\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.435723 4556 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"35011d99-3341-4081-9733-a6f42f69a6d2\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.438634 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/35011d99-3341-4081-9733-a6f42f69a6d2-logs\") pod \"glance-default-internal-api-0\" (UID: \"35011d99-3341-4081-9733-a6f42f69a6d2\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.439451 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h98ck\" (UniqueName: \"kubernetes.io/projected/35011d99-3341-4081-9733-a6f42f69a6d2-kube-api-access-h98ck\") pod \"glance-default-internal-api-0\" (UID: \"35011d99-3341-4081-9733-a6f42f69a6d2\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.439598 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/35011d99-3341-4081-9733-a6f42f69a6d2-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"35011d99-3341-4081-9733-a6f42f69a6d2\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.439832 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35011d99-3341-4081-9733-a6f42f69a6d2-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"35011d99-3341-4081-9733-a6f42f69a6d2\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.440115 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/35011d99-3341-4081-9733-a6f42f69a6d2-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"35011d99-3341-4081-9733-a6f42f69a6d2\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.444130 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35011d99-3341-4081-9733-a6f42f69a6d2-config-data\") pod \"glance-default-internal-api-0\" (UID: \"35011d99-3341-4081-9733-a6f42f69a6d2\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.444931 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35011d99-3341-4081-9733-a6f42f69a6d2-scripts\") pod \"glance-default-internal-api-0\" (UID: \"35011d99-3341-4081-9733-a6f42f69a6d2\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.445074 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35011d99-3341-4081-9733-a6f42f69a6d2-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"35011d99-3341-4081-9733-a6f42f69a6d2\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.446631 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/35011d99-3341-4081-9733-a6f42f69a6d2-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"35011d99-3341-4081-9733-a6f42f69a6d2\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.462731 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h98ck\" (UniqueName: \"kubernetes.io/projected/35011d99-3341-4081-9733-a6f42f69a6d2-kube-api-access-h98ck\") pod \"glance-default-internal-api-0\" (UID: \"35011d99-3341-4081-9733-a6f42f69a6d2\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.467705 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"35011d99-3341-4081-9733-a6f42f69a6d2\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:18:12 crc kubenswrapper[4556]: I0218 09:18:12.553827 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 18 09:18:13 crc kubenswrapper[4556]: I0218 09:18:13.104762 4556 generic.go:334] "Generic (PLEG): container finished" podID="a6347657-1897-458a-9a02-36d8655525f2" containerID="647910ecbb2afa2329d622e6150e7766438dd9b5e0ec136356c23c9ef4cfc82a" exitCode=0 Feb 18 09:18:13 crc kubenswrapper[4556]: I0218 09:18:13.104851 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-brdr4" event={"ID":"a6347657-1897-458a-9a02-36d8655525f2","Type":"ContainerDied","Data":"647910ecbb2afa2329d622e6150e7766438dd9b5e0ec136356c23c9ef4cfc82a"} Feb 18 09:18:13 crc kubenswrapper[4556]: I0218 09:18:13.114853 4556 generic.go:334] "Generic (PLEG): container finished" podID="7dd567df-a88f-4cfd-b465-f949d75ca4a2" containerID="4512175c8db79a54e8326671deb8adadf5806321eae7ed88c050aba63d66aa9d" exitCode=0 Feb 18 09:18:13 crc kubenswrapper[4556]: I0218 09:18:13.114882 4556 generic.go:334] "Generic (PLEG): container finished" podID="7dd567df-a88f-4cfd-b465-f949d75ca4a2" containerID="3be8265b5007dcf42927d8cd64e27c3f9f427cf821edbc8b0caa611d3753f4bf" exitCode=143 Feb 18 09:18:13 crc kubenswrapper[4556]: I0218 09:18:13.115066 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7dd567df-a88f-4cfd-b465-f949d75ca4a2","Type":"ContainerDied","Data":"4512175c8db79a54e8326671deb8adadf5806321eae7ed88c050aba63d66aa9d"} Feb 18 09:18:13 crc kubenswrapper[4556]: I0218 09:18:13.115105 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7dd567df-a88f-4cfd-b465-f949d75ca4a2","Type":"ContainerDied","Data":"3be8265b5007dcf42927d8cd64e27c3f9f427cf821edbc8b0caa611d3753f4bf"} Feb 18 09:18:13 crc kubenswrapper[4556]: I0218 09:18:13.199246 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 18 09:18:13 crc kubenswrapper[4556]: I0218 09:18:13.296082 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40566cc9-f4ba-4201-836a-811e07642dd8" path="/var/lib/kubelet/pods/40566cc9-f4ba-4201-836a-811e07642dd8/volumes" Feb 18 09:18:13 crc kubenswrapper[4556]: I0218 09:18:13.300665 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 18 09:18:13 crc kubenswrapper[4556]: I0218 09:18:13.361355 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dd567df-a88f-4cfd-b465-f949d75ca4a2-combined-ca-bundle\") pod \"7dd567df-a88f-4cfd-b465-f949d75ca4a2\" (UID: \"7dd567df-a88f-4cfd-b465-f949d75ca4a2\") " Feb 18 09:18:13 crc kubenswrapper[4556]: I0218 09:18:13.361463 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7dd567df-a88f-4cfd-b465-f949d75ca4a2-logs\") pod \"7dd567df-a88f-4cfd-b465-f949d75ca4a2\" (UID: \"7dd567df-a88f-4cfd-b465-f949d75ca4a2\") " Feb 18 09:18:13 crc kubenswrapper[4556]: I0218 09:18:13.361559 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7dd567df-a88f-4cfd-b465-f949d75ca4a2-public-tls-certs\") pod \"7dd567df-a88f-4cfd-b465-f949d75ca4a2\" (UID: \"7dd567df-a88f-4cfd-b465-f949d75ca4a2\") " Feb 18 09:18:13 crc kubenswrapper[4556]: I0218 09:18:13.361581 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dhg6z\" (UniqueName: \"kubernetes.io/projected/7dd567df-a88f-4cfd-b465-f949d75ca4a2-kube-api-access-dhg6z\") pod \"7dd567df-a88f-4cfd-b465-f949d75ca4a2\" (UID: \"7dd567df-a88f-4cfd-b465-f949d75ca4a2\") " Feb 18 09:18:13 crc kubenswrapper[4556]: I0218 09:18:13.361728 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"7dd567df-a88f-4cfd-b465-f949d75ca4a2\" (UID: \"7dd567df-a88f-4cfd-b465-f949d75ca4a2\") " Feb 18 09:18:13 crc kubenswrapper[4556]: I0218 09:18:13.361787 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7dd567df-a88f-4cfd-b465-f949d75ca4a2-httpd-run\") pod \"7dd567df-a88f-4cfd-b465-f949d75ca4a2\" (UID: \"7dd567df-a88f-4cfd-b465-f949d75ca4a2\") " Feb 18 09:18:13 crc kubenswrapper[4556]: I0218 09:18:13.361818 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7dd567df-a88f-4cfd-b465-f949d75ca4a2-config-data\") pod \"7dd567df-a88f-4cfd-b465-f949d75ca4a2\" (UID: \"7dd567df-a88f-4cfd-b465-f949d75ca4a2\") " Feb 18 09:18:13 crc kubenswrapper[4556]: I0218 09:18:13.361930 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7dd567df-a88f-4cfd-b465-f949d75ca4a2-scripts\") pod \"7dd567df-a88f-4cfd-b465-f949d75ca4a2\" (UID: \"7dd567df-a88f-4cfd-b465-f949d75ca4a2\") " Feb 18 09:18:13 crc kubenswrapper[4556]: I0218 09:18:13.364647 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7dd567df-a88f-4cfd-b465-f949d75ca4a2-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "7dd567df-a88f-4cfd-b465-f949d75ca4a2" (UID: "7dd567df-a88f-4cfd-b465-f949d75ca4a2"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:18:13 crc kubenswrapper[4556]: I0218 09:18:13.370920 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dd567df-a88f-4cfd-b465-f949d75ca4a2-scripts" (OuterVolumeSpecName: "scripts") pod "7dd567df-a88f-4cfd-b465-f949d75ca4a2" (UID: "7dd567df-a88f-4cfd-b465-f949d75ca4a2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:18:13 crc kubenswrapper[4556]: I0218 09:18:13.372748 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7dd567df-a88f-4cfd-b465-f949d75ca4a2-logs" (OuterVolumeSpecName: "logs") pod "7dd567df-a88f-4cfd-b465-f949d75ca4a2" (UID: "7dd567df-a88f-4cfd-b465-f949d75ca4a2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:18:13 crc kubenswrapper[4556]: I0218 09:18:13.382449 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7dd567df-a88f-4cfd-b465-f949d75ca4a2-kube-api-access-dhg6z" (OuterVolumeSpecName: "kube-api-access-dhg6z") pod "7dd567df-a88f-4cfd-b465-f949d75ca4a2" (UID: "7dd567df-a88f-4cfd-b465-f949d75ca4a2"). InnerVolumeSpecName "kube-api-access-dhg6z". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:18:13 crc kubenswrapper[4556]: I0218 09:18:13.389313 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "7dd567df-a88f-4cfd-b465-f949d75ca4a2" (UID: "7dd567df-a88f-4cfd-b465-f949d75ca4a2"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 18 09:18:13 crc kubenswrapper[4556]: I0218 09:18:13.433353 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dd567df-a88f-4cfd-b465-f949d75ca4a2-config-data" (OuterVolumeSpecName: "config-data") pod "7dd567df-a88f-4cfd-b465-f949d75ca4a2" (UID: "7dd567df-a88f-4cfd-b465-f949d75ca4a2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:18:13 crc kubenswrapper[4556]: I0218 09:18:13.435241 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dd567df-a88f-4cfd-b465-f949d75ca4a2-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "7dd567df-a88f-4cfd-b465-f949d75ca4a2" (UID: "7dd567df-a88f-4cfd-b465-f949d75ca4a2"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:18:13 crc kubenswrapper[4556]: I0218 09:18:13.443243 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dd567df-a88f-4cfd-b465-f949d75ca4a2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7dd567df-a88f-4cfd-b465-f949d75ca4a2" (UID: "7dd567df-a88f-4cfd-b465-f949d75ca4a2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:18:13 crc kubenswrapper[4556]: I0218 09:18:13.465344 4556 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Feb 18 09:18:13 crc kubenswrapper[4556]: I0218 09:18:13.465378 4556 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7dd567df-a88f-4cfd-b465-f949d75ca4a2-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:13 crc kubenswrapper[4556]: I0218 09:18:13.465407 4556 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7dd567df-a88f-4cfd-b465-f949d75ca4a2-config-data\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:13 crc kubenswrapper[4556]: I0218 09:18:13.465418 4556 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7dd567df-a88f-4cfd-b465-f949d75ca4a2-scripts\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:13 crc kubenswrapper[4556]: I0218 09:18:13.465430 4556 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dd567df-a88f-4cfd-b465-f949d75ca4a2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:13 crc kubenswrapper[4556]: I0218 09:18:13.465443 4556 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7dd567df-a88f-4cfd-b465-f949d75ca4a2-logs\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:13 crc kubenswrapper[4556]: I0218 09:18:13.465452 4556 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7dd567df-a88f-4cfd-b465-f949d75ca4a2-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:13 crc kubenswrapper[4556]: I0218 09:18:13.465463 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dhg6z\" (UniqueName: \"kubernetes.io/projected/7dd567df-a88f-4cfd-b465-f949d75ca4a2-kube-api-access-dhg6z\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:13 crc kubenswrapper[4556]: I0218 09:18:13.480267 4556 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Feb 18 09:18:13 crc kubenswrapper[4556]: I0218 09:18:13.568481 4556 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.070344 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-85bb8c66f4-fhblh" Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.070418 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-85bb8c66f4-fhblh" Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.140713 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"35011d99-3341-4081-9733-a6f42f69a6d2","Type":"ContainerStarted","Data":"4f6a9471fabb777523494b87d377f21524f2dc8cb583c2a3116675d5d8900f75"} Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.140771 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"35011d99-3341-4081-9733-a6f42f69a6d2","Type":"ContainerStarted","Data":"fbc7dd6512ba3c836f9289719c03dca46f287d8dfca505ce872676d47b0cab7f"} Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.143777 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-xhbg8" event={"ID":"2b111dd2-11a6-47cc-ba9e-2886186c676b","Type":"ContainerStarted","Data":"627b08c9854a12fcffab1aed34bc89bff2f1f8dfd17c4c63a28713fd77d9847b"} Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.151711 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7dd567df-a88f-4cfd-b465-f949d75ca4a2","Type":"ContainerDied","Data":"ab6fde9a3d76ad1f6265c242c50013db2fb59f906eaa9af459cad071e77c2f0b"} Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.151781 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.151798 4556 scope.go:117] "RemoveContainer" containerID="4512175c8db79a54e8326671deb8adadf5806321eae7ed88c050aba63d66aa9d" Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.165189 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-xhbg8" podStartSLOduration=1.735077231 podStartE2EDuration="30.165175959s" podCreationTimestamp="2026-02-18 09:17:44 +0000 UTC" firstStartedPulling="2026-02-18 09:17:45.408413898 +0000 UTC m=+822.425374879" lastFinishedPulling="2026-02-18 09:18:13.838512627 +0000 UTC m=+850.855473607" observedRunningTime="2026-02-18 09:18:14.160482863 +0000 UTC m=+851.177443844" watchObservedRunningTime="2026-02-18 09:18:14.165175959 +0000 UTC m=+851.182136939" Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.181538 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-595566bb8b-6b5h6" Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.182246 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-595566bb8b-6b5h6" Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.196072 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.229742 4556 scope.go:117] "RemoveContainer" containerID="3be8265b5007dcf42927d8cd64e27c3f9f427cf821edbc8b0caa611d3753f4bf" Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.254828 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.264686 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 18 09:18:14 crc kubenswrapper[4556]: E0218 09:18:14.267319 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dd567df-a88f-4cfd-b465-f949d75ca4a2" containerName="glance-httpd" Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.267344 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dd567df-a88f-4cfd-b465-f949d75ca4a2" containerName="glance-httpd" Feb 18 09:18:14 crc kubenswrapper[4556]: E0218 09:18:14.267373 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dd567df-a88f-4cfd-b465-f949d75ca4a2" containerName="glance-log" Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.267380 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dd567df-a88f-4cfd-b465-f949d75ca4a2" containerName="glance-log" Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.267560 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="7dd567df-a88f-4cfd-b465-f949d75ca4a2" containerName="glance-httpd" Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.267577 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="7dd567df-a88f-4cfd-b465-f949d75ca4a2" containerName="glance-log" Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.268449 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.271182 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.276478 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.279946 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.397594 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9\") " pod="openstack/glance-default-external-api-0" Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.397683 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9\") " pod="openstack/glance-default-external-api-0" Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.397712 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9-config-data\") pod \"glance-default-external-api-0\" (UID: \"2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9\") " pod="openstack/glance-default-external-api-0" Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.397818 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9-logs\") pod \"glance-default-external-api-0\" (UID: \"2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9\") " pod="openstack/glance-default-external-api-0" Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.397843 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9\") " pod="openstack/glance-default-external-api-0" Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.397901 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkcrq\" (UniqueName: \"kubernetes.io/projected/2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9-kube-api-access-gkcrq\") pod \"glance-default-external-api-0\" (UID: \"2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9\") " pod="openstack/glance-default-external-api-0" Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.397928 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9\") " pod="openstack/glance-default-external-api-0" Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.398199 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9-scripts\") pod \"glance-default-external-api-0\" (UID: \"2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9\") " pod="openstack/glance-default-external-api-0" Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.420261 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-77cf5996cc-dj8s4" Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.504036 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9\") " pod="openstack/glance-default-external-api-0" Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.504192 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9\") " pod="openstack/glance-default-external-api-0" Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.504222 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9-config-data\") pod \"glance-default-external-api-0\" (UID: \"2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9\") " pod="openstack/glance-default-external-api-0" Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.504349 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9-logs\") pod \"glance-default-external-api-0\" (UID: \"2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9\") " pod="openstack/glance-default-external-api-0" Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.504380 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9\") " pod="openstack/glance-default-external-api-0" Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.504468 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkcrq\" (UniqueName: \"kubernetes.io/projected/2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9-kube-api-access-gkcrq\") pod \"glance-default-external-api-0\" (UID: \"2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9\") " pod="openstack/glance-default-external-api-0" Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.504504 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9\") " pod="openstack/glance-default-external-api-0" Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.504687 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9-scripts\") pod \"glance-default-external-api-0\" (UID: \"2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9\") " pod="openstack/glance-default-external-api-0" Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.505793 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9-logs\") pod \"glance-default-external-api-0\" (UID: \"2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9\") " pod="openstack/glance-default-external-api-0" Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.506079 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9\") " pod="openstack/glance-default-external-api-0" Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.506225 4556 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-external-api-0" Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.512669 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9-config-data\") pod \"glance-default-external-api-0\" (UID: \"2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9\") " pod="openstack/glance-default-external-api-0" Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.513128 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9\") " pod="openstack/glance-default-external-api-0" Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.513545 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9-scripts\") pod \"glance-default-external-api-0\" (UID: \"2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9\") " pod="openstack/glance-default-external-api-0" Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.513921 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9\") " pod="openstack/glance-default-external-api-0" Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.529936 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkcrq\" (UniqueName: \"kubernetes.io/projected/2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9-kube-api-access-gkcrq\") pod \"glance-default-external-api-0\" (UID: \"2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9\") " pod="openstack/glance-default-external-api-0" Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.531140 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9\") " pod="openstack/glance-default-external-api-0" Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.606925 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.693750 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-74c5b6b6c5-q8djs" Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.748577 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-brdr4" Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.810627 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a6347657-1897-458a-9a02-36d8655525f2-fernet-keys\") pod \"a6347657-1897-458a-9a02-36d8655525f2\" (UID: \"a6347657-1897-458a-9a02-36d8655525f2\") " Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.810777 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6347657-1897-458a-9a02-36d8655525f2-combined-ca-bundle\") pod \"a6347657-1897-458a-9a02-36d8655525f2\" (UID: \"a6347657-1897-458a-9a02-36d8655525f2\") " Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.810851 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a6347657-1897-458a-9a02-36d8655525f2-credential-keys\") pod \"a6347657-1897-458a-9a02-36d8655525f2\" (UID: \"a6347657-1897-458a-9a02-36d8655525f2\") " Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.810897 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-27blw\" (UniqueName: \"kubernetes.io/projected/a6347657-1897-458a-9a02-36d8655525f2-kube-api-access-27blw\") pod \"a6347657-1897-458a-9a02-36d8655525f2\" (UID: \"a6347657-1897-458a-9a02-36d8655525f2\") " Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.810942 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6347657-1897-458a-9a02-36d8655525f2-scripts\") pod \"a6347657-1897-458a-9a02-36d8655525f2\" (UID: \"a6347657-1897-458a-9a02-36d8655525f2\") " Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.810985 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6347657-1897-458a-9a02-36d8655525f2-config-data\") pod \"a6347657-1897-458a-9a02-36d8655525f2\" (UID: \"a6347657-1897-458a-9a02-36d8655525f2\") " Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.818405 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6347657-1897-458a-9a02-36d8655525f2-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "a6347657-1897-458a-9a02-36d8655525f2" (UID: "a6347657-1897-458a-9a02-36d8655525f2"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.818644 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6347657-1897-458a-9a02-36d8655525f2-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "a6347657-1897-458a-9a02-36d8655525f2" (UID: "a6347657-1897-458a-9a02-36d8655525f2"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.821189 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6347657-1897-458a-9a02-36d8655525f2-kube-api-access-27blw" (OuterVolumeSpecName: "kube-api-access-27blw") pod "a6347657-1897-458a-9a02-36d8655525f2" (UID: "a6347657-1897-458a-9a02-36d8655525f2"). InnerVolumeSpecName "kube-api-access-27blw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.821675 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6347657-1897-458a-9a02-36d8655525f2-scripts" (OuterVolumeSpecName: "scripts") pod "a6347657-1897-458a-9a02-36d8655525f2" (UID: "a6347657-1897-458a-9a02-36d8655525f2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.844291 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6347657-1897-458a-9a02-36d8655525f2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a6347657-1897-458a-9a02-36d8655525f2" (UID: "a6347657-1897-458a-9a02-36d8655525f2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.875308 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6347657-1897-458a-9a02-36d8655525f2-config-data" (OuterVolumeSpecName: "config-data") pod "a6347657-1897-458a-9a02-36d8655525f2" (UID: "a6347657-1897-458a-9a02-36d8655525f2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.914881 4556 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6347657-1897-458a-9a02-36d8655525f2-config-data\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.914916 4556 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a6347657-1897-458a-9a02-36d8655525f2-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.914926 4556 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6347657-1897-458a-9a02-36d8655525f2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.914941 4556 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a6347657-1897-458a-9a02-36d8655525f2-credential-keys\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.914950 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-27blw\" (UniqueName: \"kubernetes.io/projected/a6347657-1897-458a-9a02-36d8655525f2-kube-api-access-27blw\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:14 crc kubenswrapper[4556]: I0218 09:18:14.914958 4556 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6347657-1897-458a-9a02-36d8655525f2-scripts\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:15 crc kubenswrapper[4556]: I0218 09:18:15.205292 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"35011d99-3341-4081-9733-a6f42f69a6d2","Type":"ContainerStarted","Data":"e18dd7ef339961839ac3bcbecba5e030be1878f9424536354a1a994260d68983"} Feb 18 09:18:15 crc kubenswrapper[4556]: I0218 09:18:15.208884 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-brdr4" Feb 18 09:18:15 crc kubenswrapper[4556]: I0218 09:18:15.209204 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-brdr4" event={"ID":"a6347657-1897-458a-9a02-36d8655525f2","Type":"ContainerDied","Data":"edbc28809a341505414bc97051fe0667a2952b1a20bf410da5bfbc5c0c2eeb16"} Feb 18 09:18:15 crc kubenswrapper[4556]: I0218 09:18:15.209245 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="edbc28809a341505414bc97051fe0667a2952b1a20bf410da5bfbc5c0c2eeb16" Feb 18 09:18:15 crc kubenswrapper[4556]: I0218 09:18:15.244344 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.244329497 podStartE2EDuration="3.244329497s" podCreationTimestamp="2026-02-18 09:18:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:18:15.23647229 +0000 UTC m=+852.253433280" watchObservedRunningTime="2026-02-18 09:18:15.244329497 +0000 UTC m=+852.261290477" Feb 18 09:18:15 crc kubenswrapper[4556]: I0218 09:18:15.299847 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7dd567df-a88f-4cfd-b465-f949d75ca4a2" path="/var/lib/kubelet/pods/7dd567df-a88f-4cfd-b465-f949d75ca4a2/volumes" Feb 18 09:18:15 crc kubenswrapper[4556]: I0218 09:18:15.300657 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-66f6978747-npdn5"] Feb 18 09:18:15 crc kubenswrapper[4556]: E0218 09:18:15.301046 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6347657-1897-458a-9a02-36d8655525f2" containerName="keystone-bootstrap" Feb 18 09:18:15 crc kubenswrapper[4556]: I0218 09:18:15.301064 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6347657-1897-458a-9a02-36d8655525f2" containerName="keystone-bootstrap" Feb 18 09:18:15 crc kubenswrapper[4556]: I0218 09:18:15.301261 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6347657-1897-458a-9a02-36d8655525f2" containerName="keystone-bootstrap" Feb 18 09:18:15 crc kubenswrapper[4556]: I0218 09:18:15.301905 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-66f6978747-npdn5" Feb 18 09:18:15 crc kubenswrapper[4556]: I0218 09:18:15.316973 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 18 09:18:15 crc kubenswrapper[4556]: I0218 09:18:15.317216 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 18 09:18:15 crc kubenswrapper[4556]: I0218 09:18:15.317273 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Feb 18 09:18:15 crc kubenswrapper[4556]: I0218 09:18:15.317497 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Feb 18 09:18:15 crc kubenswrapper[4556]: I0218 09:18:15.318054 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-dpsn5" Feb 18 09:18:15 crc kubenswrapper[4556]: I0218 09:18:15.318190 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 18 09:18:15 crc kubenswrapper[4556]: I0218 09:18:15.329838 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 18 09:18:15 crc kubenswrapper[4556]: I0218 09:18:15.345072 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-66f6978747-npdn5"] Feb 18 09:18:15 crc kubenswrapper[4556]: I0218 09:18:15.438055 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b87f75f-f2e8-4a26-b59c-1d0bbe9c777d-internal-tls-certs\") pod \"keystone-66f6978747-npdn5\" (UID: \"6b87f75f-f2e8-4a26-b59c-1d0bbe9c777d\") " pod="openstack/keystone-66f6978747-npdn5" Feb 18 09:18:15 crc kubenswrapper[4556]: I0218 09:18:15.438191 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b87f75f-f2e8-4a26-b59c-1d0bbe9c777d-config-data\") pod \"keystone-66f6978747-npdn5\" (UID: \"6b87f75f-f2e8-4a26-b59c-1d0bbe9c777d\") " pod="openstack/keystone-66f6978747-npdn5" Feb 18 09:18:15 crc kubenswrapper[4556]: I0218 09:18:15.438274 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b87f75f-f2e8-4a26-b59c-1d0bbe9c777d-scripts\") pod \"keystone-66f6978747-npdn5\" (UID: \"6b87f75f-f2e8-4a26-b59c-1d0bbe9c777d\") " pod="openstack/keystone-66f6978747-npdn5" Feb 18 09:18:15 crc kubenswrapper[4556]: I0218 09:18:15.438351 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b87f75f-f2e8-4a26-b59c-1d0bbe9c777d-public-tls-certs\") pod \"keystone-66f6978747-npdn5\" (UID: \"6b87f75f-f2e8-4a26-b59c-1d0bbe9c777d\") " pod="openstack/keystone-66f6978747-npdn5" Feb 18 09:18:15 crc kubenswrapper[4556]: I0218 09:18:15.438408 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xp6lr\" (UniqueName: \"kubernetes.io/projected/6b87f75f-f2e8-4a26-b59c-1d0bbe9c777d-kube-api-access-xp6lr\") pod \"keystone-66f6978747-npdn5\" (UID: \"6b87f75f-f2e8-4a26-b59c-1d0bbe9c777d\") " pod="openstack/keystone-66f6978747-npdn5" Feb 18 09:18:15 crc kubenswrapper[4556]: I0218 09:18:15.438444 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6b87f75f-f2e8-4a26-b59c-1d0bbe9c777d-fernet-keys\") pod \"keystone-66f6978747-npdn5\" (UID: \"6b87f75f-f2e8-4a26-b59c-1d0bbe9c777d\") " pod="openstack/keystone-66f6978747-npdn5" Feb 18 09:18:15 crc kubenswrapper[4556]: I0218 09:18:15.438500 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b87f75f-f2e8-4a26-b59c-1d0bbe9c777d-combined-ca-bundle\") pod \"keystone-66f6978747-npdn5\" (UID: \"6b87f75f-f2e8-4a26-b59c-1d0bbe9c777d\") " pod="openstack/keystone-66f6978747-npdn5" Feb 18 09:18:15 crc kubenswrapper[4556]: I0218 09:18:15.438564 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6b87f75f-f2e8-4a26-b59c-1d0bbe9c777d-credential-keys\") pod \"keystone-66f6978747-npdn5\" (UID: \"6b87f75f-f2e8-4a26-b59c-1d0bbe9c777d\") " pod="openstack/keystone-66f6978747-npdn5" Feb 18 09:18:15 crc kubenswrapper[4556]: I0218 09:18:15.540517 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b87f75f-f2e8-4a26-b59c-1d0bbe9c777d-config-data\") pod \"keystone-66f6978747-npdn5\" (UID: \"6b87f75f-f2e8-4a26-b59c-1d0bbe9c777d\") " pod="openstack/keystone-66f6978747-npdn5" Feb 18 09:18:15 crc kubenswrapper[4556]: I0218 09:18:15.540591 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b87f75f-f2e8-4a26-b59c-1d0bbe9c777d-scripts\") pod \"keystone-66f6978747-npdn5\" (UID: \"6b87f75f-f2e8-4a26-b59c-1d0bbe9c777d\") " pod="openstack/keystone-66f6978747-npdn5" Feb 18 09:18:15 crc kubenswrapper[4556]: I0218 09:18:15.540650 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b87f75f-f2e8-4a26-b59c-1d0bbe9c777d-public-tls-certs\") pod \"keystone-66f6978747-npdn5\" (UID: \"6b87f75f-f2e8-4a26-b59c-1d0bbe9c777d\") " pod="openstack/keystone-66f6978747-npdn5" Feb 18 09:18:15 crc kubenswrapper[4556]: I0218 09:18:15.540684 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xp6lr\" (UniqueName: \"kubernetes.io/projected/6b87f75f-f2e8-4a26-b59c-1d0bbe9c777d-kube-api-access-xp6lr\") pod \"keystone-66f6978747-npdn5\" (UID: \"6b87f75f-f2e8-4a26-b59c-1d0bbe9c777d\") " pod="openstack/keystone-66f6978747-npdn5" Feb 18 09:18:15 crc kubenswrapper[4556]: I0218 09:18:15.540713 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6b87f75f-f2e8-4a26-b59c-1d0bbe9c777d-fernet-keys\") pod \"keystone-66f6978747-npdn5\" (UID: \"6b87f75f-f2e8-4a26-b59c-1d0bbe9c777d\") " pod="openstack/keystone-66f6978747-npdn5" Feb 18 09:18:15 crc kubenswrapper[4556]: I0218 09:18:15.540756 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b87f75f-f2e8-4a26-b59c-1d0bbe9c777d-combined-ca-bundle\") pod \"keystone-66f6978747-npdn5\" (UID: \"6b87f75f-f2e8-4a26-b59c-1d0bbe9c777d\") " pod="openstack/keystone-66f6978747-npdn5" Feb 18 09:18:15 crc kubenswrapper[4556]: I0218 09:18:15.540793 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6b87f75f-f2e8-4a26-b59c-1d0bbe9c777d-credential-keys\") pod \"keystone-66f6978747-npdn5\" (UID: \"6b87f75f-f2e8-4a26-b59c-1d0bbe9c777d\") " pod="openstack/keystone-66f6978747-npdn5" Feb 18 09:18:15 crc kubenswrapper[4556]: I0218 09:18:15.540845 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b87f75f-f2e8-4a26-b59c-1d0bbe9c777d-internal-tls-certs\") pod \"keystone-66f6978747-npdn5\" (UID: \"6b87f75f-f2e8-4a26-b59c-1d0bbe9c777d\") " pod="openstack/keystone-66f6978747-npdn5" Feb 18 09:18:15 crc kubenswrapper[4556]: I0218 09:18:15.560738 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b87f75f-f2e8-4a26-b59c-1d0bbe9c777d-public-tls-certs\") pod \"keystone-66f6978747-npdn5\" (UID: \"6b87f75f-f2e8-4a26-b59c-1d0bbe9c777d\") " pod="openstack/keystone-66f6978747-npdn5" Feb 18 09:18:15 crc kubenswrapper[4556]: I0218 09:18:15.561204 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b87f75f-f2e8-4a26-b59c-1d0bbe9c777d-internal-tls-certs\") pod \"keystone-66f6978747-npdn5\" (UID: \"6b87f75f-f2e8-4a26-b59c-1d0bbe9c777d\") " pod="openstack/keystone-66f6978747-npdn5" Feb 18 09:18:15 crc kubenswrapper[4556]: I0218 09:18:15.564502 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6b87f75f-f2e8-4a26-b59c-1d0bbe9c777d-fernet-keys\") pod \"keystone-66f6978747-npdn5\" (UID: \"6b87f75f-f2e8-4a26-b59c-1d0bbe9c777d\") " pod="openstack/keystone-66f6978747-npdn5" Feb 18 09:18:15 crc kubenswrapper[4556]: I0218 09:18:15.566241 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b87f75f-f2e8-4a26-b59c-1d0bbe9c777d-scripts\") pod \"keystone-66f6978747-npdn5\" (UID: \"6b87f75f-f2e8-4a26-b59c-1d0bbe9c777d\") " pod="openstack/keystone-66f6978747-npdn5" Feb 18 09:18:15 crc kubenswrapper[4556]: I0218 09:18:15.566942 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b87f75f-f2e8-4a26-b59c-1d0bbe9c777d-config-data\") pod \"keystone-66f6978747-npdn5\" (UID: \"6b87f75f-f2e8-4a26-b59c-1d0bbe9c777d\") " pod="openstack/keystone-66f6978747-npdn5" Feb 18 09:18:15 crc kubenswrapper[4556]: I0218 09:18:15.572001 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xp6lr\" (UniqueName: \"kubernetes.io/projected/6b87f75f-f2e8-4a26-b59c-1d0bbe9c777d-kube-api-access-xp6lr\") pod \"keystone-66f6978747-npdn5\" (UID: \"6b87f75f-f2e8-4a26-b59c-1d0bbe9c777d\") " pod="openstack/keystone-66f6978747-npdn5" Feb 18 09:18:15 crc kubenswrapper[4556]: I0218 09:18:15.583797 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b87f75f-f2e8-4a26-b59c-1d0bbe9c777d-combined-ca-bundle\") pod \"keystone-66f6978747-npdn5\" (UID: \"6b87f75f-f2e8-4a26-b59c-1d0bbe9c777d\") " pod="openstack/keystone-66f6978747-npdn5" Feb 18 09:18:15 crc kubenswrapper[4556]: I0218 09:18:15.584016 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6b87f75f-f2e8-4a26-b59c-1d0bbe9c777d-credential-keys\") pod \"keystone-66f6978747-npdn5\" (UID: \"6b87f75f-f2e8-4a26-b59c-1d0bbe9c777d\") " pod="openstack/keystone-66f6978747-npdn5" Feb 18 09:18:15 crc kubenswrapper[4556]: I0218 09:18:15.642918 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-66f6978747-npdn5" Feb 18 09:18:16 crc kubenswrapper[4556]: I0218 09:18:16.342481 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-86dd6887f5-kqgml" Feb 18 09:18:16 crc kubenswrapper[4556]: I0218 09:18:16.414904 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f58898b5f-r69gj"] Feb 18 09:18:16 crc kubenswrapper[4556]: I0218 09:18:16.415357 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-f58898b5f-r69gj" podUID="840ae99a-2a91-4c3e-9fab-b05787d3b97b" containerName="dnsmasq-dns" containerID="cri-o://c607ad2bcebbaf1277b72f198c43bdc45e51329ad3f67a2c9ef45c2c139be0fc" gracePeriod=10 Feb 18 09:18:16 crc kubenswrapper[4556]: I0218 09:18:16.806682 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-79dd98bf77-fdbv7" Feb 18 09:18:17 crc kubenswrapper[4556]: I0218 09:18:17.269793 4556 generic.go:334] "Generic (PLEG): container finished" podID="2b111dd2-11a6-47cc-ba9e-2886186c676b" containerID="627b08c9854a12fcffab1aed34bc89bff2f1f8dfd17c4c63a28713fd77d9847b" exitCode=0 Feb 18 09:18:17 crc kubenswrapper[4556]: I0218 09:18:17.269876 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-xhbg8" event={"ID":"2b111dd2-11a6-47cc-ba9e-2886186c676b","Type":"ContainerDied","Data":"627b08c9854a12fcffab1aed34bc89bff2f1f8dfd17c4c63a28713fd77d9847b"} Feb 18 09:18:17 crc kubenswrapper[4556]: I0218 09:18:17.284291 4556 generic.go:334] "Generic (PLEG): container finished" podID="840ae99a-2a91-4c3e-9fab-b05787d3b97b" containerID="c607ad2bcebbaf1277b72f198c43bdc45e51329ad3f67a2c9ef45c2c139be0fc" exitCode=0 Feb 18 09:18:17 crc kubenswrapper[4556]: I0218 09:18:17.307669 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f58898b5f-r69gj" event={"ID":"840ae99a-2a91-4c3e-9fab-b05787d3b97b","Type":"ContainerDied","Data":"c607ad2bcebbaf1277b72f198c43bdc45e51329ad3f67a2c9ef45c2c139be0fc"} Feb 18 09:18:19 crc kubenswrapper[4556]: I0218 09:18:19.961300 4556 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-f58898b5f-r69gj" podUID="840ae99a-2a91-4c3e-9fab-b05787d3b97b" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.140:5353: connect: connection refused" Feb 18 09:18:20 crc kubenswrapper[4556]: W0218 09:18:20.370291 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f3c5d32_5a48_43df_a7e8_c65ae39bf8b9.slice/crio-10752ccd372681e784cd9b6d7a7a940de3084c11a711f90554b2dd4b3a778e5a WatchSource:0}: Error finding container 10752ccd372681e784cd9b6d7a7a940de3084c11a711f90554b2dd4b3a778e5a: Status 404 returned error can't find the container with id 10752ccd372681e784cd9b6d7a7a940de3084c11a711f90554b2dd4b3a778e5a Feb 18 09:18:20 crc kubenswrapper[4556]: I0218 09:18:20.525047 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-xhbg8" Feb 18 09:18:20 crc kubenswrapper[4556]: I0218 09:18:20.675016 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b111dd2-11a6-47cc-ba9e-2886186c676b-config-data\") pod \"2b111dd2-11a6-47cc-ba9e-2886186c676b\" (UID: \"2b111dd2-11a6-47cc-ba9e-2886186c676b\") " Feb 18 09:18:20 crc kubenswrapper[4556]: I0218 09:18:20.675319 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b111dd2-11a6-47cc-ba9e-2886186c676b-combined-ca-bundle\") pod \"2b111dd2-11a6-47cc-ba9e-2886186c676b\" (UID: \"2b111dd2-11a6-47cc-ba9e-2886186c676b\") " Feb 18 09:18:20 crc kubenswrapper[4556]: I0218 09:18:20.675382 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5dj6\" (UniqueName: \"kubernetes.io/projected/2b111dd2-11a6-47cc-ba9e-2886186c676b-kube-api-access-x5dj6\") pod \"2b111dd2-11a6-47cc-ba9e-2886186c676b\" (UID: \"2b111dd2-11a6-47cc-ba9e-2886186c676b\") " Feb 18 09:18:20 crc kubenswrapper[4556]: I0218 09:18:20.675448 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b111dd2-11a6-47cc-ba9e-2886186c676b-logs\") pod \"2b111dd2-11a6-47cc-ba9e-2886186c676b\" (UID: \"2b111dd2-11a6-47cc-ba9e-2886186c676b\") " Feb 18 09:18:20 crc kubenswrapper[4556]: I0218 09:18:20.675501 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b111dd2-11a6-47cc-ba9e-2886186c676b-scripts\") pod \"2b111dd2-11a6-47cc-ba9e-2886186c676b\" (UID: \"2b111dd2-11a6-47cc-ba9e-2886186c676b\") " Feb 18 09:18:20 crc kubenswrapper[4556]: I0218 09:18:20.679844 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b111dd2-11a6-47cc-ba9e-2886186c676b-logs" (OuterVolumeSpecName: "logs") pod "2b111dd2-11a6-47cc-ba9e-2886186c676b" (UID: "2b111dd2-11a6-47cc-ba9e-2886186c676b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:18:20 crc kubenswrapper[4556]: I0218 09:18:20.688281 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b111dd2-11a6-47cc-ba9e-2886186c676b-scripts" (OuterVolumeSpecName: "scripts") pod "2b111dd2-11a6-47cc-ba9e-2886186c676b" (UID: "2b111dd2-11a6-47cc-ba9e-2886186c676b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:18:20 crc kubenswrapper[4556]: I0218 09:18:20.694606 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b111dd2-11a6-47cc-ba9e-2886186c676b-kube-api-access-x5dj6" (OuterVolumeSpecName: "kube-api-access-x5dj6") pod "2b111dd2-11a6-47cc-ba9e-2886186c676b" (UID: "2b111dd2-11a6-47cc-ba9e-2886186c676b"). InnerVolumeSpecName "kube-api-access-x5dj6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:18:20 crc kubenswrapper[4556]: I0218 09:18:20.707777 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f58898b5f-r69gj" Feb 18 09:18:20 crc kubenswrapper[4556]: I0218 09:18:20.746315 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b111dd2-11a6-47cc-ba9e-2886186c676b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2b111dd2-11a6-47cc-ba9e-2886186c676b" (UID: "2b111dd2-11a6-47cc-ba9e-2886186c676b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:18:20 crc kubenswrapper[4556]: I0218 09:18:20.780974 4556 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b111dd2-11a6-47cc-ba9e-2886186c676b-scripts\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:20 crc kubenswrapper[4556]: I0218 09:18:20.781013 4556 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b111dd2-11a6-47cc-ba9e-2886186c676b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:20 crc kubenswrapper[4556]: I0218 09:18:20.781028 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x5dj6\" (UniqueName: \"kubernetes.io/projected/2b111dd2-11a6-47cc-ba9e-2886186c676b-kube-api-access-x5dj6\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:20 crc kubenswrapper[4556]: I0218 09:18:20.781038 4556 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b111dd2-11a6-47cc-ba9e-2886186c676b-logs\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:20 crc kubenswrapper[4556]: I0218 09:18:20.781792 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b111dd2-11a6-47cc-ba9e-2886186c676b-config-data" (OuterVolumeSpecName: "config-data") pod "2b111dd2-11a6-47cc-ba9e-2886186c676b" (UID: "2b111dd2-11a6-47cc-ba9e-2886186c676b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:18:20 crc kubenswrapper[4556]: I0218 09:18:20.882678 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/840ae99a-2a91-4c3e-9fab-b05787d3b97b-dns-svc\") pod \"840ae99a-2a91-4c3e-9fab-b05787d3b97b\" (UID: \"840ae99a-2a91-4c3e-9fab-b05787d3b97b\") " Feb 18 09:18:20 crc kubenswrapper[4556]: I0218 09:18:20.882755 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/840ae99a-2a91-4c3e-9fab-b05787d3b97b-ovsdbserver-sb\") pod \"840ae99a-2a91-4c3e-9fab-b05787d3b97b\" (UID: \"840ae99a-2a91-4c3e-9fab-b05787d3b97b\") " Feb 18 09:18:20 crc kubenswrapper[4556]: I0218 09:18:20.883336 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/840ae99a-2a91-4c3e-9fab-b05787d3b97b-config\") pod \"840ae99a-2a91-4c3e-9fab-b05787d3b97b\" (UID: \"840ae99a-2a91-4c3e-9fab-b05787d3b97b\") " Feb 18 09:18:20 crc kubenswrapper[4556]: I0218 09:18:20.883597 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5nnrt\" (UniqueName: \"kubernetes.io/projected/840ae99a-2a91-4c3e-9fab-b05787d3b97b-kube-api-access-5nnrt\") pod \"840ae99a-2a91-4c3e-9fab-b05787d3b97b\" (UID: \"840ae99a-2a91-4c3e-9fab-b05787d3b97b\") " Feb 18 09:18:20 crc kubenswrapper[4556]: I0218 09:18:20.883988 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/840ae99a-2a91-4c3e-9fab-b05787d3b97b-ovsdbserver-nb\") pod \"840ae99a-2a91-4c3e-9fab-b05787d3b97b\" (UID: \"840ae99a-2a91-4c3e-9fab-b05787d3b97b\") " Feb 18 09:18:20 crc kubenswrapper[4556]: I0218 09:18:20.884055 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/840ae99a-2a91-4c3e-9fab-b05787d3b97b-dns-swift-storage-0\") pod \"840ae99a-2a91-4c3e-9fab-b05787d3b97b\" (UID: \"840ae99a-2a91-4c3e-9fab-b05787d3b97b\") " Feb 18 09:18:20 crc kubenswrapper[4556]: I0218 09:18:20.884575 4556 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b111dd2-11a6-47cc-ba9e-2886186c676b-config-data\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:20 crc kubenswrapper[4556]: I0218 09:18:20.893586 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/840ae99a-2a91-4c3e-9fab-b05787d3b97b-kube-api-access-5nnrt" (OuterVolumeSpecName: "kube-api-access-5nnrt") pod "840ae99a-2a91-4c3e-9fab-b05787d3b97b" (UID: "840ae99a-2a91-4c3e-9fab-b05787d3b97b"). InnerVolumeSpecName "kube-api-access-5nnrt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:18:20 crc kubenswrapper[4556]: I0218 09:18:20.904737 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-66f6978747-npdn5"] Feb 18 09:18:20 crc kubenswrapper[4556]: W0218 09:18:20.913583 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6b87f75f_f2e8_4a26_b59c_1d0bbe9c777d.slice/crio-54c6789739798b39092af4bf8c3768ca31e3c621c18e403a00b0162a066f4177 WatchSource:0}: Error finding container 54c6789739798b39092af4bf8c3768ca31e3c621c18e403a00b0162a066f4177: Status 404 returned error can't find the container with id 54c6789739798b39092af4bf8c3768ca31e3c621c18e403a00b0162a066f4177 Feb 18 09:18:20 crc kubenswrapper[4556]: I0218 09:18:20.946038 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/840ae99a-2a91-4c3e-9fab-b05787d3b97b-config" (OuterVolumeSpecName: "config") pod "840ae99a-2a91-4c3e-9fab-b05787d3b97b" (UID: "840ae99a-2a91-4c3e-9fab-b05787d3b97b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:18:20 crc kubenswrapper[4556]: I0218 09:18:20.961978 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/840ae99a-2a91-4c3e-9fab-b05787d3b97b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "840ae99a-2a91-4c3e-9fab-b05787d3b97b" (UID: "840ae99a-2a91-4c3e-9fab-b05787d3b97b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:18:20 crc kubenswrapper[4556]: I0218 09:18:20.965009 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/840ae99a-2a91-4c3e-9fab-b05787d3b97b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "840ae99a-2a91-4c3e-9fab-b05787d3b97b" (UID: "840ae99a-2a91-4c3e-9fab-b05787d3b97b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:18:20 crc kubenswrapper[4556]: I0218 09:18:20.968953 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/840ae99a-2a91-4c3e-9fab-b05787d3b97b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "840ae99a-2a91-4c3e-9fab-b05787d3b97b" (UID: "840ae99a-2a91-4c3e-9fab-b05787d3b97b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:18:20 crc kubenswrapper[4556]: I0218 09:18:20.974836 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/840ae99a-2a91-4c3e-9fab-b05787d3b97b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "840ae99a-2a91-4c3e-9fab-b05787d3b97b" (UID: "840ae99a-2a91-4c3e-9fab-b05787d3b97b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:18:20 crc kubenswrapper[4556]: I0218 09:18:20.987173 4556 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/840ae99a-2a91-4c3e-9fab-b05787d3b97b-config\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:20 crc kubenswrapper[4556]: I0218 09:18:20.987203 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5nnrt\" (UniqueName: \"kubernetes.io/projected/840ae99a-2a91-4c3e-9fab-b05787d3b97b-kube-api-access-5nnrt\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:20 crc kubenswrapper[4556]: I0218 09:18:20.987214 4556 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/840ae99a-2a91-4c3e-9fab-b05787d3b97b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:20 crc kubenswrapper[4556]: I0218 09:18:20.987226 4556 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/840ae99a-2a91-4c3e-9fab-b05787d3b97b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:20 crc kubenswrapper[4556]: I0218 09:18:20.987234 4556 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/840ae99a-2a91-4c3e-9fab-b05787d3b97b-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:20 crc kubenswrapper[4556]: I0218 09:18:20.987242 4556 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/840ae99a-2a91-4c3e-9fab-b05787d3b97b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:21 crc kubenswrapper[4556]: I0218 09:18:21.327513 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f58898b5f-r69gj" event={"ID":"840ae99a-2a91-4c3e-9fab-b05787d3b97b","Type":"ContainerDied","Data":"1e33760599d03c798ff9357cdc78861f65926769817d4d7440761aaf7945dc91"} Feb 18 09:18:21 crc kubenswrapper[4556]: I0218 09:18:21.327572 4556 scope.go:117] "RemoveContainer" containerID="c607ad2bcebbaf1277b72f198c43bdc45e51329ad3f67a2c9ef45c2c139be0fc" Feb 18 09:18:21 crc kubenswrapper[4556]: I0218 09:18:21.327701 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f58898b5f-r69gj" Feb 18 09:18:21 crc kubenswrapper[4556]: I0218 09:18:21.347075 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9","Type":"ContainerStarted","Data":"c31bb188cd36a64068b7d79dd9886e17c0884ce62c635deb9aaf356133dad0ae"} Feb 18 09:18:21 crc kubenswrapper[4556]: I0218 09:18:21.347123 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9","Type":"ContainerStarted","Data":"10752ccd372681e784cd9b6d7a7a940de3084c11a711f90554b2dd4b3a778e5a"} Feb 18 09:18:21 crc kubenswrapper[4556]: I0218 09:18:21.353142 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e0739f72-891b-46e1-bce5-e17c54a407ab","Type":"ContainerStarted","Data":"f7367e0b2c468791b656052267d5c9e3f713321ffc299c0bae2758bf91f088f7"} Feb 18 09:18:21 crc kubenswrapper[4556]: I0218 09:18:21.354981 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-xhbg8" event={"ID":"2b111dd2-11a6-47cc-ba9e-2886186c676b","Type":"ContainerDied","Data":"1f75efa9811d0d1c1df4aa765d0b73d8685adb22da219fb07a779b3fd46403bb"} Feb 18 09:18:21 crc kubenswrapper[4556]: I0218 09:18:21.355012 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1f75efa9811d0d1c1df4aa765d0b73d8685adb22da219fb07a779b3fd46403bb" Feb 18 09:18:21 crc kubenswrapper[4556]: I0218 09:18:21.355069 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-xhbg8" Feb 18 09:18:21 crc kubenswrapper[4556]: I0218 09:18:21.369677 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f58898b5f-r69gj"] Feb 18 09:18:21 crc kubenswrapper[4556]: I0218 09:18:21.375059 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-r4h4k" event={"ID":"8c46304b-1b60-493b-918f-678451f13057","Type":"ContainerStarted","Data":"51fa6da89b840614647763e3ed2be148353737332fbcbef9091dbbbf265415f3"} Feb 18 09:18:21 crc kubenswrapper[4556]: I0218 09:18:21.402540 4556 scope.go:117] "RemoveContainer" containerID="031d0ab58c1cf6e0f2a156e287886e3fb123b3baf49e0b05f7123414e3a8048f" Feb 18 09:18:21 crc kubenswrapper[4556]: I0218 09:18:21.404978 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-66f6978747-npdn5" event={"ID":"6b87f75f-f2e8-4a26-b59c-1d0bbe9c777d","Type":"ContainerStarted","Data":"2328196ca68d5e4c4250dbdd2fccd296d6d28ebd5b5133ea815381057abfd1f9"} Feb 18 09:18:21 crc kubenswrapper[4556]: I0218 09:18:21.405034 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-66f6978747-npdn5" event={"ID":"6b87f75f-f2e8-4a26-b59c-1d0bbe9c777d","Type":"ContainerStarted","Data":"54c6789739798b39092af4bf8c3768ca31e3c621c18e403a00b0162a066f4177"} Feb 18 09:18:21 crc kubenswrapper[4556]: I0218 09:18:21.407370 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-66f6978747-npdn5" Feb 18 09:18:21 crc kubenswrapper[4556]: I0218 09:18:21.469603 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-f58898b5f-r69gj"] Feb 18 09:18:21 crc kubenswrapper[4556]: I0218 09:18:21.470523 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-r4h4k" podStartSLOduration=2.2641931619999998 podStartE2EDuration="37.470504101s" podCreationTimestamp="2026-02-18 09:17:44 +0000 UTC" firstStartedPulling="2026-02-18 09:17:45.727006583 +0000 UTC m=+822.743967563" lastFinishedPulling="2026-02-18 09:18:20.933317521 +0000 UTC m=+857.950278502" observedRunningTime="2026-02-18 09:18:21.417044443 +0000 UTC m=+858.434005423" watchObservedRunningTime="2026-02-18 09:18:21.470504101 +0000 UTC m=+858.487465081" Feb 18 09:18:21 crc kubenswrapper[4556]: I0218 09:18:21.481438 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-66f6978747-npdn5" podStartSLOduration=6.4814209 podStartE2EDuration="6.4814209s" podCreationTimestamp="2026-02-18 09:18:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:18:21.435778794 +0000 UTC m=+858.452739774" watchObservedRunningTime="2026-02-18 09:18:21.4814209 +0000 UTC m=+858.498381881" Feb 18 09:18:21 crc kubenswrapper[4556]: I0218 09:18:21.642625 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-9855c4778-9n9f2"] Feb 18 09:18:21 crc kubenswrapper[4556]: E0218 09:18:21.643008 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="840ae99a-2a91-4c3e-9fab-b05787d3b97b" containerName="init" Feb 18 09:18:21 crc kubenswrapper[4556]: I0218 09:18:21.643028 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="840ae99a-2a91-4c3e-9fab-b05787d3b97b" containerName="init" Feb 18 09:18:21 crc kubenswrapper[4556]: E0218 09:18:21.643044 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="840ae99a-2a91-4c3e-9fab-b05787d3b97b" containerName="dnsmasq-dns" Feb 18 09:18:21 crc kubenswrapper[4556]: I0218 09:18:21.643051 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="840ae99a-2a91-4c3e-9fab-b05787d3b97b" containerName="dnsmasq-dns" Feb 18 09:18:21 crc kubenswrapper[4556]: E0218 09:18:21.643065 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b111dd2-11a6-47cc-ba9e-2886186c676b" containerName="placement-db-sync" Feb 18 09:18:21 crc kubenswrapper[4556]: I0218 09:18:21.643071 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b111dd2-11a6-47cc-ba9e-2886186c676b" containerName="placement-db-sync" Feb 18 09:18:21 crc kubenswrapper[4556]: I0218 09:18:21.643296 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="840ae99a-2a91-4c3e-9fab-b05787d3b97b" containerName="dnsmasq-dns" Feb 18 09:18:21 crc kubenswrapper[4556]: I0218 09:18:21.643326 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b111dd2-11a6-47cc-ba9e-2886186c676b" containerName="placement-db-sync" Feb 18 09:18:21 crc kubenswrapper[4556]: I0218 09:18:21.644283 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9855c4778-9n9f2" Feb 18 09:18:21 crc kubenswrapper[4556]: I0218 09:18:21.648340 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Feb 18 09:18:21 crc kubenswrapper[4556]: I0218 09:18:21.648800 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-wgbkj" Feb 18 09:18:21 crc kubenswrapper[4556]: I0218 09:18:21.648981 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Feb 18 09:18:21 crc kubenswrapper[4556]: I0218 09:18:21.649116 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Feb 18 09:18:21 crc kubenswrapper[4556]: I0218 09:18:21.649257 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Feb 18 09:18:21 crc kubenswrapper[4556]: I0218 09:18:21.654944 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-9855c4778-9n9f2"] Feb 18 09:18:21 crc kubenswrapper[4556]: I0218 09:18:21.738505 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac2fe5b2-09b6-440f-83db-8555ec304a27-internal-tls-certs\") pod \"placement-9855c4778-9n9f2\" (UID: \"ac2fe5b2-09b6-440f-83db-8555ec304a27\") " pod="openstack/placement-9855c4778-9n9f2" Feb 18 09:18:21 crc kubenswrapper[4556]: I0218 09:18:21.738577 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac2fe5b2-09b6-440f-83db-8555ec304a27-public-tls-certs\") pod \"placement-9855c4778-9n9f2\" (UID: \"ac2fe5b2-09b6-440f-83db-8555ec304a27\") " pod="openstack/placement-9855c4778-9n9f2" Feb 18 09:18:21 crc kubenswrapper[4556]: I0218 09:18:21.738731 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac2fe5b2-09b6-440f-83db-8555ec304a27-config-data\") pod \"placement-9855c4778-9n9f2\" (UID: \"ac2fe5b2-09b6-440f-83db-8555ec304a27\") " pod="openstack/placement-9855c4778-9n9f2" Feb 18 09:18:21 crc kubenswrapper[4556]: I0218 09:18:21.738864 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fmxj\" (UniqueName: \"kubernetes.io/projected/ac2fe5b2-09b6-440f-83db-8555ec304a27-kube-api-access-5fmxj\") pod \"placement-9855c4778-9n9f2\" (UID: \"ac2fe5b2-09b6-440f-83db-8555ec304a27\") " pod="openstack/placement-9855c4778-9n9f2" Feb 18 09:18:21 crc kubenswrapper[4556]: I0218 09:18:21.738942 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac2fe5b2-09b6-440f-83db-8555ec304a27-combined-ca-bundle\") pod \"placement-9855c4778-9n9f2\" (UID: \"ac2fe5b2-09b6-440f-83db-8555ec304a27\") " pod="openstack/placement-9855c4778-9n9f2" Feb 18 09:18:21 crc kubenswrapper[4556]: I0218 09:18:21.739143 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ac2fe5b2-09b6-440f-83db-8555ec304a27-logs\") pod \"placement-9855c4778-9n9f2\" (UID: \"ac2fe5b2-09b6-440f-83db-8555ec304a27\") " pod="openstack/placement-9855c4778-9n9f2" Feb 18 09:18:21 crc kubenswrapper[4556]: I0218 09:18:21.739258 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac2fe5b2-09b6-440f-83db-8555ec304a27-scripts\") pod \"placement-9855c4778-9n9f2\" (UID: \"ac2fe5b2-09b6-440f-83db-8555ec304a27\") " pod="openstack/placement-9855c4778-9n9f2" Feb 18 09:18:21 crc kubenswrapper[4556]: I0218 09:18:21.840910 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ac2fe5b2-09b6-440f-83db-8555ec304a27-logs\") pod \"placement-9855c4778-9n9f2\" (UID: \"ac2fe5b2-09b6-440f-83db-8555ec304a27\") " pod="openstack/placement-9855c4778-9n9f2" Feb 18 09:18:21 crc kubenswrapper[4556]: I0218 09:18:21.840984 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac2fe5b2-09b6-440f-83db-8555ec304a27-scripts\") pod \"placement-9855c4778-9n9f2\" (UID: \"ac2fe5b2-09b6-440f-83db-8555ec304a27\") " pod="openstack/placement-9855c4778-9n9f2" Feb 18 09:18:21 crc kubenswrapper[4556]: I0218 09:18:21.841071 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac2fe5b2-09b6-440f-83db-8555ec304a27-internal-tls-certs\") pod \"placement-9855c4778-9n9f2\" (UID: \"ac2fe5b2-09b6-440f-83db-8555ec304a27\") " pod="openstack/placement-9855c4778-9n9f2" Feb 18 09:18:21 crc kubenswrapper[4556]: I0218 09:18:21.841125 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac2fe5b2-09b6-440f-83db-8555ec304a27-public-tls-certs\") pod \"placement-9855c4778-9n9f2\" (UID: \"ac2fe5b2-09b6-440f-83db-8555ec304a27\") " pod="openstack/placement-9855c4778-9n9f2" Feb 18 09:18:21 crc kubenswrapper[4556]: I0218 09:18:21.841218 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac2fe5b2-09b6-440f-83db-8555ec304a27-config-data\") pod \"placement-9855c4778-9n9f2\" (UID: \"ac2fe5b2-09b6-440f-83db-8555ec304a27\") " pod="openstack/placement-9855c4778-9n9f2" Feb 18 09:18:21 crc kubenswrapper[4556]: I0218 09:18:21.841293 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5fmxj\" (UniqueName: \"kubernetes.io/projected/ac2fe5b2-09b6-440f-83db-8555ec304a27-kube-api-access-5fmxj\") pod \"placement-9855c4778-9n9f2\" (UID: \"ac2fe5b2-09b6-440f-83db-8555ec304a27\") " pod="openstack/placement-9855c4778-9n9f2" Feb 18 09:18:21 crc kubenswrapper[4556]: I0218 09:18:21.841336 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ac2fe5b2-09b6-440f-83db-8555ec304a27-logs\") pod \"placement-9855c4778-9n9f2\" (UID: \"ac2fe5b2-09b6-440f-83db-8555ec304a27\") " pod="openstack/placement-9855c4778-9n9f2" Feb 18 09:18:21 crc kubenswrapper[4556]: I0218 09:18:21.841347 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac2fe5b2-09b6-440f-83db-8555ec304a27-combined-ca-bundle\") pod \"placement-9855c4778-9n9f2\" (UID: \"ac2fe5b2-09b6-440f-83db-8555ec304a27\") " pod="openstack/placement-9855c4778-9n9f2" Feb 18 09:18:21 crc kubenswrapper[4556]: I0218 09:18:21.847448 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac2fe5b2-09b6-440f-83db-8555ec304a27-combined-ca-bundle\") pod \"placement-9855c4778-9n9f2\" (UID: \"ac2fe5b2-09b6-440f-83db-8555ec304a27\") " pod="openstack/placement-9855c4778-9n9f2" Feb 18 09:18:21 crc kubenswrapper[4556]: I0218 09:18:21.848611 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac2fe5b2-09b6-440f-83db-8555ec304a27-public-tls-certs\") pod \"placement-9855c4778-9n9f2\" (UID: \"ac2fe5b2-09b6-440f-83db-8555ec304a27\") " pod="openstack/placement-9855c4778-9n9f2" Feb 18 09:18:21 crc kubenswrapper[4556]: I0218 09:18:21.849175 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac2fe5b2-09b6-440f-83db-8555ec304a27-config-data\") pod \"placement-9855c4778-9n9f2\" (UID: \"ac2fe5b2-09b6-440f-83db-8555ec304a27\") " pod="openstack/placement-9855c4778-9n9f2" Feb 18 09:18:21 crc kubenswrapper[4556]: I0218 09:18:21.860478 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac2fe5b2-09b6-440f-83db-8555ec304a27-internal-tls-certs\") pod \"placement-9855c4778-9n9f2\" (UID: \"ac2fe5b2-09b6-440f-83db-8555ec304a27\") " pod="openstack/placement-9855c4778-9n9f2" Feb 18 09:18:21 crc kubenswrapper[4556]: I0218 09:18:21.861398 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac2fe5b2-09b6-440f-83db-8555ec304a27-scripts\") pod \"placement-9855c4778-9n9f2\" (UID: \"ac2fe5b2-09b6-440f-83db-8555ec304a27\") " pod="openstack/placement-9855c4778-9n9f2" Feb 18 09:18:21 crc kubenswrapper[4556]: I0218 09:18:21.862635 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5fmxj\" (UniqueName: \"kubernetes.io/projected/ac2fe5b2-09b6-440f-83db-8555ec304a27-kube-api-access-5fmxj\") pod \"placement-9855c4778-9n9f2\" (UID: \"ac2fe5b2-09b6-440f-83db-8555ec304a27\") " pod="openstack/placement-9855c4778-9n9f2" Feb 18 09:18:21 crc kubenswrapper[4556]: I0218 09:18:21.960441 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9855c4778-9n9f2" Feb 18 09:18:22 crc kubenswrapper[4556]: I0218 09:18:22.393488 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9","Type":"ContainerStarted","Data":"4f89dee65414ab2d433b8d26869688906b5bc9fad55290401fe98c3fb76403fd"} Feb 18 09:18:22 crc kubenswrapper[4556]: I0218 09:18:22.411906 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=8.411882557 podStartE2EDuration="8.411882557s" podCreationTimestamp="2026-02-18 09:18:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:18:22.410100185 +0000 UTC m=+859.427061164" watchObservedRunningTime="2026-02-18 09:18:22.411882557 +0000 UTC m=+859.428843538" Feb 18 09:18:22 crc kubenswrapper[4556]: I0218 09:18:22.477228 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-9855c4778-9n9f2"] Feb 18 09:18:22 crc kubenswrapper[4556]: W0218 09:18:22.481457 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podac2fe5b2_09b6_440f_83db_8555ec304a27.slice/crio-7383d290783e801838f9496efce1c878958791831dadb2e767aff37632287b72 WatchSource:0}: Error finding container 7383d290783e801838f9496efce1c878958791831dadb2e767aff37632287b72: Status 404 returned error can't find the container with id 7383d290783e801838f9496efce1c878958791831dadb2e767aff37632287b72 Feb 18 09:18:22 crc kubenswrapper[4556]: I0218 09:18:22.554630 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 18 09:18:22 crc kubenswrapper[4556]: I0218 09:18:22.554683 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 18 09:18:22 crc kubenswrapper[4556]: I0218 09:18:22.605988 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 18 09:18:22 crc kubenswrapper[4556]: I0218 09:18:22.652456 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 18 09:18:23 crc kubenswrapper[4556]: I0218 09:18:23.293417 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="840ae99a-2a91-4c3e-9fab-b05787d3b97b" path="/var/lib/kubelet/pods/840ae99a-2a91-4c3e-9fab-b05787d3b97b/volumes" Feb 18 09:18:23 crc kubenswrapper[4556]: I0218 09:18:23.416946 4556 generic.go:334] "Generic (PLEG): container finished" podID="8c46304b-1b60-493b-918f-678451f13057" containerID="51fa6da89b840614647763e3ed2be148353737332fbcbef9091dbbbf265415f3" exitCode=0 Feb 18 09:18:23 crc kubenswrapper[4556]: I0218 09:18:23.417058 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-r4h4k" event={"ID":"8c46304b-1b60-493b-918f-678451f13057","Type":"ContainerDied","Data":"51fa6da89b840614647763e3ed2be148353737332fbcbef9091dbbbf265415f3"} Feb 18 09:18:23 crc kubenswrapper[4556]: I0218 09:18:23.424234 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-9855c4778-9n9f2" event={"ID":"ac2fe5b2-09b6-440f-83db-8555ec304a27","Type":"ContainerStarted","Data":"29a468e226c0efccf508a799815c67bd280c8f7b18e9e103b08ada3882a10f9f"} Feb 18 09:18:23 crc kubenswrapper[4556]: I0218 09:18:23.424285 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-9855c4778-9n9f2" event={"ID":"ac2fe5b2-09b6-440f-83db-8555ec304a27","Type":"ContainerStarted","Data":"bdd5048e11830b1de3aca1ea937dd598df35f31bc490397ebf49a7e32747484e"} Feb 18 09:18:23 crc kubenswrapper[4556]: I0218 09:18:23.424298 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-9855c4778-9n9f2" event={"ID":"ac2fe5b2-09b6-440f-83db-8555ec304a27","Type":"ContainerStarted","Data":"7383d290783e801838f9496efce1c878958791831dadb2e767aff37632287b72"} Feb 18 09:18:23 crc kubenswrapper[4556]: I0218 09:18:23.425233 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 18 09:18:23 crc kubenswrapper[4556]: I0218 09:18:23.425273 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-9855c4778-9n9f2" Feb 18 09:18:23 crc kubenswrapper[4556]: I0218 09:18:23.426335 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-9855c4778-9n9f2" Feb 18 09:18:23 crc kubenswrapper[4556]: I0218 09:18:23.426363 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 18 09:18:23 crc kubenswrapper[4556]: I0218 09:18:23.454230 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-9855c4778-9n9f2" podStartSLOduration=2.454208131 podStartE2EDuration="2.454208131s" podCreationTimestamp="2026-02-18 09:18:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:18:23.44995225 +0000 UTC m=+860.466913230" watchObservedRunningTime="2026-02-18 09:18:23.454208131 +0000 UTC m=+860.471169111" Feb 18 09:18:24 crc kubenswrapper[4556]: I0218 09:18:24.072780 4556 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-85bb8c66f4-fhblh" podUID="3354b930-6b00-4aa7-a4bd-97ec410cc863" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Feb 18 09:18:24 crc kubenswrapper[4556]: I0218 09:18:24.204850 4556 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-595566bb8b-6b5h6" podUID="30fffa0b-72a5-4bbe-a1fb-fa8b26d1decf" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Feb 18 09:18:24 crc kubenswrapper[4556]: I0218 09:18:24.443133 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-dtq7t" event={"ID":"cb11270a-c754-44b2-b0b1-937cff515ddd","Type":"ContainerStarted","Data":"cb6f76579994554cb029ab9574593a5fd90e6979565cbbfb1eacfea7989f8262"} Feb 18 09:18:24 crc kubenswrapper[4556]: I0218 09:18:24.480722 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-dtq7t" podStartSLOduration=2.063878695 podStartE2EDuration="40.480690469s" podCreationTimestamp="2026-02-18 09:17:44 +0000 UTC" firstStartedPulling="2026-02-18 09:17:45.403136478 +0000 UTC m=+822.420097458" lastFinishedPulling="2026-02-18 09:18:23.819948252 +0000 UTC m=+860.836909232" observedRunningTime="2026-02-18 09:18:24.458078921 +0000 UTC m=+861.475039902" watchObservedRunningTime="2026-02-18 09:18:24.480690469 +0000 UTC m=+861.497651449" Feb 18 09:18:24 crc kubenswrapper[4556]: I0218 09:18:24.611396 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 18 09:18:24 crc kubenswrapper[4556]: I0218 09:18:24.614383 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 18 09:18:24 crc kubenswrapper[4556]: I0218 09:18:24.665508 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 18 09:18:24 crc kubenswrapper[4556]: I0218 09:18:24.679577 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 18 09:18:24 crc kubenswrapper[4556]: I0218 09:18:24.816396 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-r4h4k" Feb 18 09:18:24 crc kubenswrapper[4556]: I0218 09:18:24.925301 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8c46304b-1b60-493b-918f-678451f13057-db-sync-config-data\") pod \"8c46304b-1b60-493b-918f-678451f13057\" (UID: \"8c46304b-1b60-493b-918f-678451f13057\") " Feb 18 09:18:24 crc kubenswrapper[4556]: I0218 09:18:24.925395 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mss5k\" (UniqueName: \"kubernetes.io/projected/8c46304b-1b60-493b-918f-678451f13057-kube-api-access-mss5k\") pod \"8c46304b-1b60-493b-918f-678451f13057\" (UID: \"8c46304b-1b60-493b-918f-678451f13057\") " Feb 18 09:18:24 crc kubenswrapper[4556]: I0218 09:18:24.925559 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c46304b-1b60-493b-918f-678451f13057-combined-ca-bundle\") pod \"8c46304b-1b60-493b-918f-678451f13057\" (UID: \"8c46304b-1b60-493b-918f-678451f13057\") " Feb 18 09:18:24 crc kubenswrapper[4556]: I0218 09:18:24.940318 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c46304b-1b60-493b-918f-678451f13057-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "8c46304b-1b60-493b-918f-678451f13057" (UID: "8c46304b-1b60-493b-918f-678451f13057"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:18:24 crc kubenswrapper[4556]: I0218 09:18:24.947792 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c46304b-1b60-493b-918f-678451f13057-kube-api-access-mss5k" (OuterVolumeSpecName: "kube-api-access-mss5k") pod "8c46304b-1b60-493b-918f-678451f13057" (UID: "8c46304b-1b60-493b-918f-678451f13057"). InnerVolumeSpecName "kube-api-access-mss5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:18:24 crc kubenswrapper[4556]: I0218 09:18:24.955710 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c46304b-1b60-493b-918f-678451f13057-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8c46304b-1b60-493b-918f-678451f13057" (UID: "8c46304b-1b60-493b-918f-678451f13057"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:18:25 crc kubenswrapper[4556]: I0218 09:18:25.040071 4556 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c46304b-1b60-493b-918f-678451f13057-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:25 crc kubenswrapper[4556]: I0218 09:18:25.040519 4556 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8c46304b-1b60-493b-918f-678451f13057-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:25 crc kubenswrapper[4556]: I0218 09:18:25.040533 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mss5k\" (UniqueName: \"kubernetes.io/projected/8c46304b-1b60-493b-918f-678451f13057-kube-api-access-mss5k\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:25 crc kubenswrapper[4556]: I0218 09:18:25.076047 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-x6wtm"] Feb 18 09:18:25 crc kubenswrapper[4556]: E0218 09:18:25.076543 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c46304b-1b60-493b-918f-678451f13057" containerName="barbican-db-sync" Feb 18 09:18:25 crc kubenswrapper[4556]: I0218 09:18:25.076565 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c46304b-1b60-493b-918f-678451f13057" containerName="barbican-db-sync" Feb 18 09:18:25 crc kubenswrapper[4556]: I0218 09:18:25.076710 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c46304b-1b60-493b-918f-678451f13057" containerName="barbican-db-sync" Feb 18 09:18:25 crc kubenswrapper[4556]: I0218 09:18:25.077971 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-x6wtm" Feb 18 09:18:25 crc kubenswrapper[4556]: I0218 09:18:25.085340 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-x6wtm"] Feb 18 09:18:25 crc kubenswrapper[4556]: I0218 09:18:25.244686 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1fb56420-2137-43d2-b790-1773daf3f0c2-utilities\") pod \"certified-operators-x6wtm\" (UID: \"1fb56420-2137-43d2-b790-1773daf3f0c2\") " pod="openshift-marketplace/certified-operators-x6wtm" Feb 18 09:18:25 crc kubenswrapper[4556]: I0218 09:18:25.244804 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1fb56420-2137-43d2-b790-1773daf3f0c2-catalog-content\") pod \"certified-operators-x6wtm\" (UID: \"1fb56420-2137-43d2-b790-1773daf3f0c2\") " pod="openshift-marketplace/certified-operators-x6wtm" Feb 18 09:18:25 crc kubenswrapper[4556]: I0218 09:18:25.244960 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xtlr\" (UniqueName: \"kubernetes.io/projected/1fb56420-2137-43d2-b790-1773daf3f0c2-kube-api-access-5xtlr\") pod \"certified-operators-x6wtm\" (UID: \"1fb56420-2137-43d2-b790-1773daf3f0c2\") " pod="openshift-marketplace/certified-operators-x6wtm" Feb 18 09:18:25 crc kubenswrapper[4556]: I0218 09:18:25.347333 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1fb56420-2137-43d2-b790-1773daf3f0c2-utilities\") pod \"certified-operators-x6wtm\" (UID: \"1fb56420-2137-43d2-b790-1773daf3f0c2\") " pod="openshift-marketplace/certified-operators-x6wtm" Feb 18 09:18:25 crc kubenswrapper[4556]: I0218 09:18:25.347469 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1fb56420-2137-43d2-b790-1773daf3f0c2-catalog-content\") pod \"certified-operators-x6wtm\" (UID: \"1fb56420-2137-43d2-b790-1773daf3f0c2\") " pod="openshift-marketplace/certified-operators-x6wtm" Feb 18 09:18:25 crc kubenswrapper[4556]: I0218 09:18:25.347658 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xtlr\" (UniqueName: \"kubernetes.io/projected/1fb56420-2137-43d2-b790-1773daf3f0c2-kube-api-access-5xtlr\") pod \"certified-operators-x6wtm\" (UID: \"1fb56420-2137-43d2-b790-1773daf3f0c2\") " pod="openshift-marketplace/certified-operators-x6wtm" Feb 18 09:18:25 crc kubenswrapper[4556]: I0218 09:18:25.347972 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1fb56420-2137-43d2-b790-1773daf3f0c2-utilities\") pod \"certified-operators-x6wtm\" (UID: \"1fb56420-2137-43d2-b790-1773daf3f0c2\") " pod="openshift-marketplace/certified-operators-x6wtm" Feb 18 09:18:25 crc kubenswrapper[4556]: I0218 09:18:25.351574 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1fb56420-2137-43d2-b790-1773daf3f0c2-catalog-content\") pod \"certified-operators-x6wtm\" (UID: \"1fb56420-2137-43d2-b790-1773daf3f0c2\") " pod="openshift-marketplace/certified-operators-x6wtm" Feb 18 09:18:25 crc kubenswrapper[4556]: I0218 09:18:25.375994 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xtlr\" (UniqueName: \"kubernetes.io/projected/1fb56420-2137-43d2-b790-1773daf3f0c2-kube-api-access-5xtlr\") pod \"certified-operators-x6wtm\" (UID: \"1fb56420-2137-43d2-b790-1773daf3f0c2\") " pod="openshift-marketplace/certified-operators-x6wtm" Feb 18 09:18:25 crc kubenswrapper[4556]: I0218 09:18:25.397306 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 18 09:18:25 crc kubenswrapper[4556]: I0218 09:18:25.397374 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-x6wtm" Feb 18 09:18:25 crc kubenswrapper[4556]: I0218 09:18:25.473978 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-r4h4k" Feb 18 09:18:25 crc kubenswrapper[4556]: I0218 09:18:25.474409 4556 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 18 09:18:25 crc kubenswrapper[4556]: I0218 09:18:25.474933 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-r4h4k" event={"ID":"8c46304b-1b60-493b-918f-678451f13057","Type":"ContainerDied","Data":"2ef9a757b29b8804aea1b53755203b640f5b983a618f22d01e71c2a8661ad43a"} Feb 18 09:18:25 crc kubenswrapper[4556]: I0218 09:18:25.474962 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2ef9a757b29b8804aea1b53755203b640f5b983a618f22d01e71c2a8661ad43a" Feb 18 09:18:25 crc kubenswrapper[4556]: I0218 09:18:25.474979 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 18 09:18:25 crc kubenswrapper[4556]: I0218 09:18:25.474989 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 18 09:18:25 crc kubenswrapper[4556]: I0218 09:18:25.776568 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-5fc745f7dd-68srz"] Feb 18 09:18:25 crc kubenswrapper[4556]: I0218 09:18:25.778108 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5fc745f7dd-68srz" Feb 18 09:18:25 crc kubenswrapper[4556]: I0218 09:18:25.801170 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-xbqch" Feb 18 09:18:25 crc kubenswrapper[4556]: I0218 09:18:25.807183 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-b64d8f579-7sj9h"] Feb 18 09:18:25 crc kubenswrapper[4556]: I0218 09:18:25.808493 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-b64d8f579-7sj9h" Feb 18 09:18:25 crc kubenswrapper[4556]: I0218 09:18:25.814439 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Feb 18 09:18:25 crc kubenswrapper[4556]: I0218 09:18:25.814602 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Feb 18 09:18:25 crc kubenswrapper[4556]: I0218 09:18:25.826399 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Feb 18 09:18:25 crc kubenswrapper[4556]: I0218 09:18:25.890897 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8be9c112-7903-4ec0-a503-fb16c676caf4-combined-ca-bundle\") pod \"barbican-keystone-listener-5fc745f7dd-68srz\" (UID: \"8be9c112-7903-4ec0-a503-fb16c676caf4\") " pod="openstack/barbican-keystone-listener-5fc745f7dd-68srz" Feb 18 09:18:25 crc kubenswrapper[4556]: I0218 09:18:25.891071 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8be9c112-7903-4ec0-a503-fb16c676caf4-logs\") pod \"barbican-keystone-listener-5fc745f7dd-68srz\" (UID: \"8be9c112-7903-4ec0-a503-fb16c676caf4\") " pod="openstack/barbican-keystone-listener-5fc745f7dd-68srz" Feb 18 09:18:25 crc kubenswrapper[4556]: I0218 09:18:25.899240 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8be9c112-7903-4ec0-a503-fb16c676caf4-config-data\") pod \"barbican-keystone-listener-5fc745f7dd-68srz\" (UID: \"8be9c112-7903-4ec0-a503-fb16c676caf4\") " pod="openstack/barbican-keystone-listener-5fc745f7dd-68srz" Feb 18 09:18:25 crc kubenswrapper[4556]: I0218 09:18:25.899336 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7s66b\" (UniqueName: \"kubernetes.io/projected/8be9c112-7903-4ec0-a503-fb16c676caf4-kube-api-access-7s66b\") pod \"barbican-keystone-listener-5fc745f7dd-68srz\" (UID: \"8be9c112-7903-4ec0-a503-fb16c676caf4\") " pod="openstack/barbican-keystone-listener-5fc745f7dd-68srz" Feb 18 09:18:25 crc kubenswrapper[4556]: I0218 09:18:25.899480 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8be9c112-7903-4ec0-a503-fb16c676caf4-config-data-custom\") pod \"barbican-keystone-listener-5fc745f7dd-68srz\" (UID: \"8be9c112-7903-4ec0-a503-fb16c676caf4\") " pod="openstack/barbican-keystone-listener-5fc745f7dd-68srz" Feb 18 09:18:25 crc kubenswrapper[4556]: I0218 09:18:25.920949 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-b64d8f579-7sj9h"] Feb 18 09:18:25 crc kubenswrapper[4556]: I0218 09:18:25.950493 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5fc745f7dd-68srz"] Feb 18 09:18:25 crc kubenswrapper[4556]: I0218 09:18:25.993753 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-54c675984c-gfwjb"] Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.001415 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54c675984c-gfwjb" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.001303 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8be9c112-7903-4ec0-a503-fb16c676caf4-combined-ca-bundle\") pod \"barbican-keystone-listener-5fc745f7dd-68srz\" (UID: \"8be9c112-7903-4ec0-a503-fb16c676caf4\") " pod="openstack/barbican-keystone-listener-5fc745f7dd-68srz" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.001627 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d6ef48e2-d690-4060-b013-5ac2be288161-config-data-custom\") pod \"barbican-worker-b64d8f579-7sj9h\" (UID: \"d6ef48e2-d690-4060-b013-5ac2be288161\") " pod="openstack/barbican-worker-b64d8f579-7sj9h" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.001647 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6ef48e2-d690-4060-b013-5ac2be288161-combined-ca-bundle\") pod \"barbican-worker-b64d8f579-7sj9h\" (UID: \"d6ef48e2-d690-4060-b013-5ac2be288161\") " pod="openstack/barbican-worker-b64d8f579-7sj9h" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.001697 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6ef48e2-d690-4060-b013-5ac2be288161-logs\") pod \"barbican-worker-b64d8f579-7sj9h\" (UID: \"d6ef48e2-d690-4060-b013-5ac2be288161\") " pod="openstack/barbican-worker-b64d8f579-7sj9h" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.001725 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8be9c112-7903-4ec0-a503-fb16c676caf4-logs\") pod \"barbican-keystone-listener-5fc745f7dd-68srz\" (UID: \"8be9c112-7903-4ec0-a503-fb16c676caf4\") " pod="openstack/barbican-keystone-listener-5fc745f7dd-68srz" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.001810 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6ef48e2-d690-4060-b013-5ac2be288161-config-data\") pod \"barbican-worker-b64d8f579-7sj9h\" (UID: \"d6ef48e2-d690-4060-b013-5ac2be288161\") " pod="openstack/barbican-worker-b64d8f579-7sj9h" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.001866 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8be9c112-7903-4ec0-a503-fb16c676caf4-config-data\") pod \"barbican-keystone-listener-5fc745f7dd-68srz\" (UID: \"8be9c112-7903-4ec0-a503-fb16c676caf4\") " pod="openstack/barbican-keystone-listener-5fc745f7dd-68srz" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.001907 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7s66b\" (UniqueName: \"kubernetes.io/projected/8be9c112-7903-4ec0-a503-fb16c676caf4-kube-api-access-7s66b\") pod \"barbican-keystone-listener-5fc745f7dd-68srz\" (UID: \"8be9c112-7903-4ec0-a503-fb16c676caf4\") " pod="openstack/barbican-keystone-listener-5fc745f7dd-68srz" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.001942 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8be9c112-7903-4ec0-a503-fb16c676caf4-config-data-custom\") pod \"barbican-keystone-listener-5fc745f7dd-68srz\" (UID: \"8be9c112-7903-4ec0-a503-fb16c676caf4\") " pod="openstack/barbican-keystone-listener-5fc745f7dd-68srz" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.002031 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nzqc\" (UniqueName: \"kubernetes.io/projected/d6ef48e2-d690-4060-b013-5ac2be288161-kube-api-access-4nzqc\") pod \"barbican-worker-b64d8f579-7sj9h\" (UID: \"d6ef48e2-d690-4060-b013-5ac2be288161\") " pod="openstack/barbican-worker-b64d8f579-7sj9h" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.002908 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8be9c112-7903-4ec0-a503-fb16c676caf4-logs\") pod \"barbican-keystone-listener-5fc745f7dd-68srz\" (UID: \"8be9c112-7903-4ec0-a503-fb16c676caf4\") " pod="openstack/barbican-keystone-listener-5fc745f7dd-68srz" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.010023 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8be9c112-7903-4ec0-a503-fb16c676caf4-combined-ca-bundle\") pod \"barbican-keystone-listener-5fc745f7dd-68srz\" (UID: \"8be9c112-7903-4ec0-a503-fb16c676caf4\") " pod="openstack/barbican-keystone-listener-5fc745f7dd-68srz" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.012771 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8be9c112-7903-4ec0-a503-fb16c676caf4-config-data-custom\") pod \"barbican-keystone-listener-5fc745f7dd-68srz\" (UID: \"8be9c112-7903-4ec0-a503-fb16c676caf4\") " pod="openstack/barbican-keystone-listener-5fc745f7dd-68srz" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.013066 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8be9c112-7903-4ec0-a503-fb16c676caf4-config-data\") pod \"barbican-keystone-listener-5fc745f7dd-68srz\" (UID: \"8be9c112-7903-4ec0-a503-fb16c676caf4\") " pod="openstack/barbican-keystone-listener-5fc745f7dd-68srz" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.026204 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54c675984c-gfwjb"] Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.045882 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7s66b\" (UniqueName: \"kubernetes.io/projected/8be9c112-7903-4ec0-a503-fb16c676caf4-kube-api-access-7s66b\") pod \"barbican-keystone-listener-5fc745f7dd-68srz\" (UID: \"8be9c112-7903-4ec0-a503-fb16c676caf4\") " pod="openstack/barbican-keystone-listener-5fc745f7dd-68srz" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.065352 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-x6wtm"] Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.107708 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/afe9fc26-12ca-41b2-8076-6d8374384041-dns-svc\") pod \"dnsmasq-dns-54c675984c-gfwjb\" (UID: \"afe9fc26-12ca-41b2-8076-6d8374384041\") " pod="openstack/dnsmasq-dns-54c675984c-gfwjb" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.107806 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/afe9fc26-12ca-41b2-8076-6d8374384041-dns-swift-storage-0\") pod \"dnsmasq-dns-54c675984c-gfwjb\" (UID: \"afe9fc26-12ca-41b2-8076-6d8374384041\") " pod="openstack/dnsmasq-dns-54c675984c-gfwjb" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.107866 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nzqc\" (UniqueName: \"kubernetes.io/projected/d6ef48e2-d690-4060-b013-5ac2be288161-kube-api-access-4nzqc\") pod \"barbican-worker-b64d8f579-7sj9h\" (UID: \"d6ef48e2-d690-4060-b013-5ac2be288161\") " pod="openstack/barbican-worker-b64d8f579-7sj9h" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.107988 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d6ef48e2-d690-4060-b013-5ac2be288161-config-data-custom\") pod \"barbican-worker-b64d8f579-7sj9h\" (UID: \"d6ef48e2-d690-4060-b013-5ac2be288161\") " pod="openstack/barbican-worker-b64d8f579-7sj9h" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.108020 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6ef48e2-d690-4060-b013-5ac2be288161-combined-ca-bundle\") pod \"barbican-worker-b64d8f579-7sj9h\" (UID: \"d6ef48e2-d690-4060-b013-5ac2be288161\") " pod="openstack/barbican-worker-b64d8f579-7sj9h" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.108071 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6ef48e2-d690-4060-b013-5ac2be288161-logs\") pod \"barbican-worker-b64d8f579-7sj9h\" (UID: \"d6ef48e2-d690-4060-b013-5ac2be288161\") " pod="openstack/barbican-worker-b64d8f579-7sj9h" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.108099 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/afe9fc26-12ca-41b2-8076-6d8374384041-ovsdbserver-nb\") pod \"dnsmasq-dns-54c675984c-gfwjb\" (UID: \"afe9fc26-12ca-41b2-8076-6d8374384041\") " pod="openstack/dnsmasq-dns-54c675984c-gfwjb" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.108179 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6ef48e2-d690-4060-b013-5ac2be288161-config-data\") pod \"barbican-worker-b64d8f579-7sj9h\" (UID: \"d6ef48e2-d690-4060-b013-5ac2be288161\") " pod="openstack/barbican-worker-b64d8f579-7sj9h" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.108217 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/afe9fc26-12ca-41b2-8076-6d8374384041-config\") pod \"dnsmasq-dns-54c675984c-gfwjb\" (UID: \"afe9fc26-12ca-41b2-8076-6d8374384041\") " pod="openstack/dnsmasq-dns-54c675984c-gfwjb" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.108236 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/afe9fc26-12ca-41b2-8076-6d8374384041-ovsdbserver-sb\") pod \"dnsmasq-dns-54c675984c-gfwjb\" (UID: \"afe9fc26-12ca-41b2-8076-6d8374384041\") " pod="openstack/dnsmasq-dns-54c675984c-gfwjb" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.108258 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgjdk\" (UniqueName: \"kubernetes.io/projected/afe9fc26-12ca-41b2-8076-6d8374384041-kube-api-access-jgjdk\") pod \"dnsmasq-dns-54c675984c-gfwjb\" (UID: \"afe9fc26-12ca-41b2-8076-6d8374384041\") " pod="openstack/dnsmasq-dns-54c675984c-gfwjb" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.108580 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6ef48e2-d690-4060-b013-5ac2be288161-logs\") pod \"barbican-worker-b64d8f579-7sj9h\" (UID: \"d6ef48e2-d690-4060-b013-5ac2be288161\") " pod="openstack/barbican-worker-b64d8f579-7sj9h" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.121668 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d6ef48e2-d690-4060-b013-5ac2be288161-config-data-custom\") pod \"barbican-worker-b64d8f579-7sj9h\" (UID: \"d6ef48e2-d690-4060-b013-5ac2be288161\") " pod="openstack/barbican-worker-b64d8f579-7sj9h" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.132438 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5fc745f7dd-68srz" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.135232 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6ef48e2-d690-4060-b013-5ac2be288161-config-data\") pod \"barbican-worker-b64d8f579-7sj9h\" (UID: \"d6ef48e2-d690-4060-b013-5ac2be288161\") " pod="openstack/barbican-worker-b64d8f579-7sj9h" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.141711 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6ef48e2-d690-4060-b013-5ac2be288161-combined-ca-bundle\") pod \"barbican-worker-b64d8f579-7sj9h\" (UID: \"d6ef48e2-d690-4060-b013-5ac2be288161\") " pod="openstack/barbican-worker-b64d8f579-7sj9h" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.161630 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nzqc\" (UniqueName: \"kubernetes.io/projected/d6ef48e2-d690-4060-b013-5ac2be288161-kube-api-access-4nzqc\") pod \"barbican-worker-b64d8f579-7sj9h\" (UID: \"d6ef48e2-d690-4060-b013-5ac2be288161\") " pod="openstack/barbican-worker-b64d8f579-7sj9h" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.181785 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-c6ccddcd4-hbvfc"] Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.183665 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-c6ccddcd4-hbvfc" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.191329 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-c6ccddcd4-hbvfc"] Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.194904 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.210064 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/afe9fc26-12ca-41b2-8076-6d8374384041-config\") pod \"dnsmasq-dns-54c675984c-gfwjb\" (UID: \"afe9fc26-12ca-41b2-8076-6d8374384041\") " pod="openstack/dnsmasq-dns-54c675984c-gfwjb" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.210101 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/afe9fc26-12ca-41b2-8076-6d8374384041-ovsdbserver-sb\") pod \"dnsmasq-dns-54c675984c-gfwjb\" (UID: \"afe9fc26-12ca-41b2-8076-6d8374384041\") " pod="openstack/dnsmasq-dns-54c675984c-gfwjb" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.210129 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgjdk\" (UniqueName: \"kubernetes.io/projected/afe9fc26-12ca-41b2-8076-6d8374384041-kube-api-access-jgjdk\") pod \"dnsmasq-dns-54c675984c-gfwjb\" (UID: \"afe9fc26-12ca-41b2-8076-6d8374384041\") " pod="openstack/dnsmasq-dns-54c675984c-gfwjb" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.212005 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/afe9fc26-12ca-41b2-8076-6d8374384041-config\") pod \"dnsmasq-dns-54c675984c-gfwjb\" (UID: \"afe9fc26-12ca-41b2-8076-6d8374384041\") " pod="openstack/dnsmasq-dns-54c675984c-gfwjb" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.212187 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/afe9fc26-12ca-41b2-8076-6d8374384041-dns-svc\") pod \"dnsmasq-dns-54c675984c-gfwjb\" (UID: \"afe9fc26-12ca-41b2-8076-6d8374384041\") " pod="openstack/dnsmasq-dns-54c675984c-gfwjb" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.212269 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/afe9fc26-12ca-41b2-8076-6d8374384041-dns-swift-storage-0\") pod \"dnsmasq-dns-54c675984c-gfwjb\" (UID: \"afe9fc26-12ca-41b2-8076-6d8374384041\") " pod="openstack/dnsmasq-dns-54c675984c-gfwjb" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.212508 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/afe9fc26-12ca-41b2-8076-6d8374384041-ovsdbserver-nb\") pod \"dnsmasq-dns-54c675984c-gfwjb\" (UID: \"afe9fc26-12ca-41b2-8076-6d8374384041\") " pod="openstack/dnsmasq-dns-54c675984c-gfwjb" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.213098 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/afe9fc26-12ca-41b2-8076-6d8374384041-ovsdbserver-nb\") pod \"dnsmasq-dns-54c675984c-gfwjb\" (UID: \"afe9fc26-12ca-41b2-8076-6d8374384041\") " pod="openstack/dnsmasq-dns-54c675984c-gfwjb" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.213787 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/afe9fc26-12ca-41b2-8076-6d8374384041-ovsdbserver-sb\") pod \"dnsmasq-dns-54c675984c-gfwjb\" (UID: \"afe9fc26-12ca-41b2-8076-6d8374384041\") " pod="openstack/dnsmasq-dns-54c675984c-gfwjb" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.219048 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/afe9fc26-12ca-41b2-8076-6d8374384041-dns-swift-storage-0\") pod \"dnsmasq-dns-54c675984c-gfwjb\" (UID: \"afe9fc26-12ca-41b2-8076-6d8374384041\") " pod="openstack/dnsmasq-dns-54c675984c-gfwjb" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.219606 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/afe9fc26-12ca-41b2-8076-6d8374384041-dns-svc\") pod \"dnsmasq-dns-54c675984c-gfwjb\" (UID: \"afe9fc26-12ca-41b2-8076-6d8374384041\") " pod="openstack/dnsmasq-dns-54c675984c-gfwjb" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.239635 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-b64d8f579-7sj9h" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.247138 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgjdk\" (UniqueName: \"kubernetes.io/projected/afe9fc26-12ca-41b2-8076-6d8374384041-kube-api-access-jgjdk\") pod \"dnsmasq-dns-54c675984c-gfwjb\" (UID: \"afe9fc26-12ca-41b2-8076-6d8374384041\") " pod="openstack/dnsmasq-dns-54c675984c-gfwjb" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.265719 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.317570 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18385928-f2f7-4f58-9f83-e563af72c8b4-combined-ca-bundle\") pod \"barbican-api-c6ccddcd4-hbvfc\" (UID: \"18385928-f2f7-4f58-9f83-e563af72c8b4\") " pod="openstack/barbican-api-c6ccddcd4-hbvfc" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.317835 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rc7nn\" (UniqueName: \"kubernetes.io/projected/18385928-f2f7-4f58-9f83-e563af72c8b4-kube-api-access-rc7nn\") pod \"barbican-api-c6ccddcd4-hbvfc\" (UID: \"18385928-f2f7-4f58-9f83-e563af72c8b4\") " pod="openstack/barbican-api-c6ccddcd4-hbvfc" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.317866 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/18385928-f2f7-4f58-9f83-e563af72c8b4-config-data-custom\") pod \"barbican-api-c6ccddcd4-hbvfc\" (UID: \"18385928-f2f7-4f58-9f83-e563af72c8b4\") " pod="openstack/barbican-api-c6ccddcd4-hbvfc" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.318050 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18385928-f2f7-4f58-9f83-e563af72c8b4-config-data\") pod \"barbican-api-c6ccddcd4-hbvfc\" (UID: \"18385928-f2f7-4f58-9f83-e563af72c8b4\") " pod="openstack/barbican-api-c6ccddcd4-hbvfc" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.318136 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/18385928-f2f7-4f58-9f83-e563af72c8b4-logs\") pod \"barbican-api-c6ccddcd4-hbvfc\" (UID: \"18385928-f2f7-4f58-9f83-e563af72c8b4\") " pod="openstack/barbican-api-c6ccddcd4-hbvfc" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.415775 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54c675984c-gfwjb" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.420200 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/18385928-f2f7-4f58-9f83-e563af72c8b4-logs\") pod \"barbican-api-c6ccddcd4-hbvfc\" (UID: \"18385928-f2f7-4f58-9f83-e563af72c8b4\") " pod="openstack/barbican-api-c6ccddcd4-hbvfc" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.420314 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18385928-f2f7-4f58-9f83-e563af72c8b4-combined-ca-bundle\") pod \"barbican-api-c6ccddcd4-hbvfc\" (UID: \"18385928-f2f7-4f58-9f83-e563af72c8b4\") " pod="openstack/barbican-api-c6ccddcd4-hbvfc" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.420377 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rc7nn\" (UniqueName: \"kubernetes.io/projected/18385928-f2f7-4f58-9f83-e563af72c8b4-kube-api-access-rc7nn\") pod \"barbican-api-c6ccddcd4-hbvfc\" (UID: \"18385928-f2f7-4f58-9f83-e563af72c8b4\") " pod="openstack/barbican-api-c6ccddcd4-hbvfc" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.420412 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/18385928-f2f7-4f58-9f83-e563af72c8b4-config-data-custom\") pod \"barbican-api-c6ccddcd4-hbvfc\" (UID: \"18385928-f2f7-4f58-9f83-e563af72c8b4\") " pod="openstack/barbican-api-c6ccddcd4-hbvfc" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.420585 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18385928-f2f7-4f58-9f83-e563af72c8b4-config-data\") pod \"barbican-api-c6ccddcd4-hbvfc\" (UID: \"18385928-f2f7-4f58-9f83-e563af72c8b4\") " pod="openstack/barbican-api-c6ccddcd4-hbvfc" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.439623 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18385928-f2f7-4f58-9f83-e563af72c8b4-config-data\") pod \"barbican-api-c6ccddcd4-hbvfc\" (UID: \"18385928-f2f7-4f58-9f83-e563af72c8b4\") " pod="openstack/barbican-api-c6ccddcd4-hbvfc" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.442900 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/18385928-f2f7-4f58-9f83-e563af72c8b4-logs\") pod \"barbican-api-c6ccddcd4-hbvfc\" (UID: \"18385928-f2f7-4f58-9f83-e563af72c8b4\") " pod="openstack/barbican-api-c6ccddcd4-hbvfc" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.454206 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18385928-f2f7-4f58-9f83-e563af72c8b4-combined-ca-bundle\") pod \"barbican-api-c6ccddcd4-hbvfc\" (UID: \"18385928-f2f7-4f58-9f83-e563af72c8b4\") " pod="openstack/barbican-api-c6ccddcd4-hbvfc" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.455255 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/18385928-f2f7-4f58-9f83-e563af72c8b4-config-data-custom\") pod \"barbican-api-c6ccddcd4-hbvfc\" (UID: \"18385928-f2f7-4f58-9f83-e563af72c8b4\") " pod="openstack/barbican-api-c6ccddcd4-hbvfc" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.473596 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rc7nn\" (UniqueName: \"kubernetes.io/projected/18385928-f2f7-4f58-9f83-e563af72c8b4-kube-api-access-rc7nn\") pod \"barbican-api-c6ccddcd4-hbvfc\" (UID: \"18385928-f2f7-4f58-9f83-e563af72c8b4\") " pod="openstack/barbican-api-c6ccddcd4-hbvfc" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.489431 4556 generic.go:334] "Generic (PLEG): container finished" podID="1fb56420-2137-43d2-b790-1773daf3f0c2" containerID="0ed40eaa96898558bbf3a3f10d8fc68051c8f2daee7f7acaaf04c49222f3285c" exitCode=0 Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.491510 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x6wtm" event={"ID":"1fb56420-2137-43d2-b790-1773daf3f0c2","Type":"ContainerDied","Data":"0ed40eaa96898558bbf3a3f10d8fc68051c8f2daee7f7acaaf04c49222f3285c"} Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.491552 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x6wtm" event={"ID":"1fb56420-2137-43d2-b790-1773daf3f0c2","Type":"ContainerStarted","Data":"55bc5251185b214278150623c3ba439b07097ad9a8fa559ea68fa2a6bed0096a"} Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.559275 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-c6ccddcd4-hbvfc" Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.800705 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5fc745f7dd-68srz"] Feb 18 09:18:26 crc kubenswrapper[4556]: W0218 09:18:26.814930 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8be9c112_7903_4ec0_a503_fb16c676caf4.slice/crio-4ab983d5d32a187497ba3398aadf6abd24f4f3d75dfd6fa69d49faa9974d84fc WatchSource:0}: Error finding container 4ab983d5d32a187497ba3398aadf6abd24f4f3d75dfd6fa69d49faa9974d84fc: Status 404 returned error can't find the container with id 4ab983d5d32a187497ba3398aadf6abd24f4f3d75dfd6fa69d49faa9974d84fc Feb 18 09:18:26 crc kubenswrapper[4556]: I0218 09:18:26.978965 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-b64d8f579-7sj9h"] Feb 18 09:18:27 crc kubenswrapper[4556]: W0218 09:18:26.986428 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd6ef48e2_d690_4060_b013_5ac2be288161.slice/crio-0812327f3f03c7119c325084cd1912d8b5e04b0482a00ff7853b40126c80bd42 WatchSource:0}: Error finding container 0812327f3f03c7119c325084cd1912d8b5e04b0482a00ff7853b40126c80bd42: Status 404 returned error can't find the container with id 0812327f3f03c7119c325084cd1912d8b5e04b0482a00ff7853b40126c80bd42 Feb 18 09:18:27 crc kubenswrapper[4556]: I0218 09:18:27.167558 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54c675984c-gfwjb"] Feb 18 09:18:27 crc kubenswrapper[4556]: I0218 09:18:27.338798 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-c6ccddcd4-hbvfc"] Feb 18 09:18:27 crc kubenswrapper[4556]: I0218 09:18:27.513633 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-b64d8f579-7sj9h" event={"ID":"d6ef48e2-d690-4060-b013-5ac2be288161","Type":"ContainerStarted","Data":"0812327f3f03c7119c325084cd1912d8b5e04b0482a00ff7853b40126c80bd42"} Feb 18 09:18:27 crc kubenswrapper[4556]: I0218 09:18:27.515354 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-c6ccddcd4-hbvfc" event={"ID":"18385928-f2f7-4f58-9f83-e563af72c8b4","Type":"ContainerStarted","Data":"4adbdfc316b6a3c5fad94306cb4338ce4ccf3ce88ca906c78af38d1d503be5a6"} Feb 18 09:18:27 crc kubenswrapper[4556]: I0218 09:18:27.517291 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54c675984c-gfwjb" event={"ID":"afe9fc26-12ca-41b2-8076-6d8374384041","Type":"ContainerStarted","Data":"e7b8ac5d5bd77858d1c7aae949f5f21ce98fc0418cbb25c273771cedd2212236"} Feb 18 09:18:27 crc kubenswrapper[4556]: I0218 09:18:27.521189 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5fc745f7dd-68srz" event={"ID":"8be9c112-7903-4ec0-a503-fb16c676caf4","Type":"ContainerStarted","Data":"4ab983d5d32a187497ba3398aadf6abd24f4f3d75dfd6fa69d49faa9974d84fc"} Feb 18 09:18:28 crc kubenswrapper[4556]: I0218 09:18:28.554637 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-c6ccddcd4-hbvfc" event={"ID":"18385928-f2f7-4f58-9f83-e563af72c8b4","Type":"ContainerStarted","Data":"2ce3f2b0b5da033ec3e8d4a3882f0e1ddc3ecafa75e1c88a35839c6ddc867cd4"} Feb 18 09:18:28 crc kubenswrapper[4556]: I0218 09:18:28.559858 4556 generic.go:334] "Generic (PLEG): container finished" podID="1fb56420-2137-43d2-b790-1773daf3f0c2" containerID="f4b9aab0d6c8aa124f35a7d50b97576b617bd928e81f72be3d4072a9ae816a71" exitCode=0 Feb 18 09:18:28 crc kubenswrapper[4556]: I0218 09:18:28.559899 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x6wtm" event={"ID":"1fb56420-2137-43d2-b790-1773daf3f0c2","Type":"ContainerDied","Data":"f4b9aab0d6c8aa124f35a7d50b97576b617bd928e81f72be3d4072a9ae816a71"} Feb 18 09:18:28 crc kubenswrapper[4556]: I0218 09:18:28.565952 4556 generic.go:334] "Generic (PLEG): container finished" podID="afe9fc26-12ca-41b2-8076-6d8374384041" containerID="e8b71066786b1a917aea7665cd2bc13916cae20786320151dc408f88a6dc3b9b" exitCode=0 Feb 18 09:18:28 crc kubenswrapper[4556]: I0218 09:18:28.565982 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54c675984c-gfwjb" event={"ID":"afe9fc26-12ca-41b2-8076-6d8374384041","Type":"ContainerDied","Data":"e8b71066786b1a917aea7665cd2bc13916cae20786320151dc408f88a6dc3b9b"} Feb 18 09:18:28 crc kubenswrapper[4556]: I0218 09:18:28.669652 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-845cb5d46d-z5vhw"] Feb 18 09:18:28 crc kubenswrapper[4556]: I0218 09:18:28.671044 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-845cb5d46d-z5vhw" Feb 18 09:18:28 crc kubenswrapper[4556]: I0218 09:18:28.681919 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Feb 18 09:18:28 crc kubenswrapper[4556]: I0218 09:18:28.689093 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Feb 18 09:18:28 crc kubenswrapper[4556]: I0218 09:18:28.706695 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-845cb5d46d-z5vhw"] Feb 18 09:18:28 crc kubenswrapper[4556]: I0218 09:18:28.742295 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 18 09:18:28 crc kubenswrapper[4556]: I0218 09:18:28.742393 4556 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 18 09:18:28 crc kubenswrapper[4556]: I0218 09:18:28.752766 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 18 09:18:28 crc kubenswrapper[4556]: I0218 09:18:28.799609 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ff67ebd8-094e-4c5b-b164-d6b37fc169de-config-data-custom\") pod \"barbican-api-845cb5d46d-z5vhw\" (UID: \"ff67ebd8-094e-4c5b-b164-d6b37fc169de\") " pod="openstack/barbican-api-845cb5d46d-z5vhw" Feb 18 09:18:28 crc kubenswrapper[4556]: I0218 09:18:28.799674 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff67ebd8-094e-4c5b-b164-d6b37fc169de-config-data\") pod \"barbican-api-845cb5d46d-z5vhw\" (UID: \"ff67ebd8-094e-4c5b-b164-d6b37fc169de\") " pod="openstack/barbican-api-845cb5d46d-z5vhw" Feb 18 09:18:28 crc kubenswrapper[4556]: I0218 09:18:28.799777 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xs8pg\" (UniqueName: \"kubernetes.io/projected/ff67ebd8-094e-4c5b-b164-d6b37fc169de-kube-api-access-xs8pg\") pod \"barbican-api-845cb5d46d-z5vhw\" (UID: \"ff67ebd8-094e-4c5b-b164-d6b37fc169de\") " pod="openstack/barbican-api-845cb5d46d-z5vhw" Feb 18 09:18:28 crc kubenswrapper[4556]: I0218 09:18:28.799850 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff67ebd8-094e-4c5b-b164-d6b37fc169de-combined-ca-bundle\") pod \"barbican-api-845cb5d46d-z5vhw\" (UID: \"ff67ebd8-094e-4c5b-b164-d6b37fc169de\") " pod="openstack/barbican-api-845cb5d46d-z5vhw" Feb 18 09:18:28 crc kubenswrapper[4556]: I0218 09:18:28.799938 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff67ebd8-094e-4c5b-b164-d6b37fc169de-public-tls-certs\") pod \"barbican-api-845cb5d46d-z5vhw\" (UID: \"ff67ebd8-094e-4c5b-b164-d6b37fc169de\") " pod="openstack/barbican-api-845cb5d46d-z5vhw" Feb 18 09:18:28 crc kubenswrapper[4556]: I0218 09:18:28.800105 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff67ebd8-094e-4c5b-b164-d6b37fc169de-internal-tls-certs\") pod \"barbican-api-845cb5d46d-z5vhw\" (UID: \"ff67ebd8-094e-4c5b-b164-d6b37fc169de\") " pod="openstack/barbican-api-845cb5d46d-z5vhw" Feb 18 09:18:28 crc kubenswrapper[4556]: I0218 09:18:28.800145 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff67ebd8-094e-4c5b-b164-d6b37fc169de-logs\") pod \"barbican-api-845cb5d46d-z5vhw\" (UID: \"ff67ebd8-094e-4c5b-b164-d6b37fc169de\") " pod="openstack/barbican-api-845cb5d46d-z5vhw" Feb 18 09:18:28 crc kubenswrapper[4556]: I0218 09:18:28.902520 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ff67ebd8-094e-4c5b-b164-d6b37fc169de-config-data-custom\") pod \"barbican-api-845cb5d46d-z5vhw\" (UID: \"ff67ebd8-094e-4c5b-b164-d6b37fc169de\") " pod="openstack/barbican-api-845cb5d46d-z5vhw" Feb 18 09:18:28 crc kubenswrapper[4556]: I0218 09:18:28.902574 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff67ebd8-094e-4c5b-b164-d6b37fc169de-config-data\") pod \"barbican-api-845cb5d46d-z5vhw\" (UID: \"ff67ebd8-094e-4c5b-b164-d6b37fc169de\") " pod="openstack/barbican-api-845cb5d46d-z5vhw" Feb 18 09:18:28 crc kubenswrapper[4556]: I0218 09:18:28.902665 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xs8pg\" (UniqueName: \"kubernetes.io/projected/ff67ebd8-094e-4c5b-b164-d6b37fc169de-kube-api-access-xs8pg\") pod \"barbican-api-845cb5d46d-z5vhw\" (UID: \"ff67ebd8-094e-4c5b-b164-d6b37fc169de\") " pod="openstack/barbican-api-845cb5d46d-z5vhw" Feb 18 09:18:28 crc kubenswrapper[4556]: I0218 09:18:28.902690 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff67ebd8-094e-4c5b-b164-d6b37fc169de-combined-ca-bundle\") pod \"barbican-api-845cb5d46d-z5vhw\" (UID: \"ff67ebd8-094e-4c5b-b164-d6b37fc169de\") " pod="openstack/barbican-api-845cb5d46d-z5vhw" Feb 18 09:18:28 crc kubenswrapper[4556]: I0218 09:18:28.902729 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff67ebd8-094e-4c5b-b164-d6b37fc169de-public-tls-certs\") pod \"barbican-api-845cb5d46d-z5vhw\" (UID: \"ff67ebd8-094e-4c5b-b164-d6b37fc169de\") " pod="openstack/barbican-api-845cb5d46d-z5vhw" Feb 18 09:18:28 crc kubenswrapper[4556]: I0218 09:18:28.902795 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff67ebd8-094e-4c5b-b164-d6b37fc169de-internal-tls-certs\") pod \"barbican-api-845cb5d46d-z5vhw\" (UID: \"ff67ebd8-094e-4c5b-b164-d6b37fc169de\") " pod="openstack/barbican-api-845cb5d46d-z5vhw" Feb 18 09:18:28 crc kubenswrapper[4556]: I0218 09:18:28.902823 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff67ebd8-094e-4c5b-b164-d6b37fc169de-logs\") pod \"barbican-api-845cb5d46d-z5vhw\" (UID: \"ff67ebd8-094e-4c5b-b164-d6b37fc169de\") " pod="openstack/barbican-api-845cb5d46d-z5vhw" Feb 18 09:18:28 crc kubenswrapper[4556]: I0218 09:18:28.903327 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff67ebd8-094e-4c5b-b164-d6b37fc169de-logs\") pod \"barbican-api-845cb5d46d-z5vhw\" (UID: \"ff67ebd8-094e-4c5b-b164-d6b37fc169de\") " pod="openstack/barbican-api-845cb5d46d-z5vhw" Feb 18 09:18:28 crc kubenswrapper[4556]: I0218 09:18:28.908639 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff67ebd8-094e-4c5b-b164-d6b37fc169de-internal-tls-certs\") pod \"barbican-api-845cb5d46d-z5vhw\" (UID: \"ff67ebd8-094e-4c5b-b164-d6b37fc169de\") " pod="openstack/barbican-api-845cb5d46d-z5vhw" Feb 18 09:18:28 crc kubenswrapper[4556]: I0218 09:18:28.908780 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ff67ebd8-094e-4c5b-b164-d6b37fc169de-config-data-custom\") pod \"barbican-api-845cb5d46d-z5vhw\" (UID: \"ff67ebd8-094e-4c5b-b164-d6b37fc169de\") " pod="openstack/barbican-api-845cb5d46d-z5vhw" Feb 18 09:18:28 crc kubenswrapper[4556]: I0218 09:18:28.909192 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff67ebd8-094e-4c5b-b164-d6b37fc169de-config-data\") pod \"barbican-api-845cb5d46d-z5vhw\" (UID: \"ff67ebd8-094e-4c5b-b164-d6b37fc169de\") " pod="openstack/barbican-api-845cb5d46d-z5vhw" Feb 18 09:18:28 crc kubenswrapper[4556]: I0218 09:18:28.909687 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff67ebd8-094e-4c5b-b164-d6b37fc169de-combined-ca-bundle\") pod \"barbican-api-845cb5d46d-z5vhw\" (UID: \"ff67ebd8-094e-4c5b-b164-d6b37fc169de\") " pod="openstack/barbican-api-845cb5d46d-z5vhw" Feb 18 09:18:28 crc kubenswrapper[4556]: I0218 09:18:28.910546 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff67ebd8-094e-4c5b-b164-d6b37fc169de-public-tls-certs\") pod \"barbican-api-845cb5d46d-z5vhw\" (UID: \"ff67ebd8-094e-4c5b-b164-d6b37fc169de\") " pod="openstack/barbican-api-845cb5d46d-z5vhw" Feb 18 09:18:28 crc kubenswrapper[4556]: I0218 09:18:28.920615 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xs8pg\" (UniqueName: \"kubernetes.io/projected/ff67ebd8-094e-4c5b-b164-d6b37fc169de-kube-api-access-xs8pg\") pod \"barbican-api-845cb5d46d-z5vhw\" (UID: \"ff67ebd8-094e-4c5b-b164-d6b37fc169de\") " pod="openstack/barbican-api-845cb5d46d-z5vhw" Feb 18 09:18:28 crc kubenswrapper[4556]: I0218 09:18:28.998963 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-845cb5d46d-z5vhw" Feb 18 09:18:29 crc kubenswrapper[4556]: I0218 09:18:29.575440 4556 generic.go:334] "Generic (PLEG): container finished" podID="cb11270a-c754-44b2-b0b1-937cff515ddd" containerID="cb6f76579994554cb029ab9574593a5fd90e6979565cbbfb1eacfea7989f8262" exitCode=0 Feb 18 09:18:29 crc kubenswrapper[4556]: I0218 09:18:29.575527 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-dtq7t" event={"ID":"cb11270a-c754-44b2-b0b1-937cff515ddd","Type":"ContainerDied","Data":"cb6f76579994554cb029ab9574593a5fd90e6979565cbbfb1eacfea7989f8262"} Feb 18 09:18:31 crc kubenswrapper[4556]: I0218 09:18:31.602299 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-d8cc75c64-fwzfq" Feb 18 09:18:31 crc kubenswrapper[4556]: I0218 09:18:31.727633 4556 patch_prober.go:28] interesting pod/machine-config-daemon-f76hs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 09:18:31 crc kubenswrapper[4556]: I0218 09:18:31.727695 4556 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 09:18:31 crc kubenswrapper[4556]: I0218 09:18:31.886229 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-dtq7t" Feb 18 09:18:31 crc kubenswrapper[4556]: I0218 09:18:31.907631 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6fd7956ddf-b647f"] Feb 18 09:18:31 crc kubenswrapper[4556]: I0218 09:18:31.908301 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6fd7956ddf-b647f" podUID="b92f250f-e888-4ff5-ae26-9e31c7be3379" containerName="neutron-api" containerID="cri-o://20b200abdb599e8ec5463ea4f53a30e71450839bf07dba61fec5a4eb01b348dd" gracePeriod=30 Feb 18 09:18:31 crc kubenswrapper[4556]: I0218 09:18:31.908380 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6fd7956ddf-b647f" podUID="b92f250f-e888-4ff5-ae26-9e31c7be3379" containerName="neutron-httpd" containerID="cri-o://cc6725c2f27c13f0413ab0e4d8de88e810bf5e3b4b491a0eedf920bb5855de30" gracePeriod=30 Feb 18 09:18:31 crc kubenswrapper[4556]: I0218 09:18:31.916166 4556 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-6fd7956ddf-b647f" podUID="b92f250f-e888-4ff5-ae26-9e31c7be3379" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.151:9696/\": EOF" Feb 18 09:18:31 crc kubenswrapper[4556]: I0218 09:18:31.935448 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-54664bdbc-cxnvs"] Feb 18 09:18:31 crc kubenswrapper[4556]: E0218 09:18:31.935953 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb11270a-c754-44b2-b0b1-937cff515ddd" containerName="cinder-db-sync" Feb 18 09:18:31 crc kubenswrapper[4556]: I0218 09:18:31.935973 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb11270a-c754-44b2-b0b1-937cff515ddd" containerName="cinder-db-sync" Feb 18 09:18:31 crc kubenswrapper[4556]: I0218 09:18:31.936216 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb11270a-c754-44b2-b0b1-937cff515ddd" containerName="cinder-db-sync" Feb 18 09:18:31 crc kubenswrapper[4556]: I0218 09:18:31.937299 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-54664bdbc-cxnvs" Feb 18 09:18:31 crc kubenswrapper[4556]: I0218 09:18:31.943318 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-54664bdbc-cxnvs"] Feb 18 09:18:31 crc kubenswrapper[4556]: I0218 09:18:31.975657 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cb11270a-c754-44b2-b0b1-937cff515ddd-etc-machine-id\") pod \"cb11270a-c754-44b2-b0b1-937cff515ddd\" (UID: \"cb11270a-c754-44b2-b0b1-937cff515ddd\") " Feb 18 09:18:31 crc kubenswrapper[4556]: I0218 09:18:31.975788 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb11270a-c754-44b2-b0b1-937cff515ddd-combined-ca-bundle\") pod \"cb11270a-c754-44b2-b0b1-937cff515ddd\" (UID: \"cb11270a-c754-44b2-b0b1-937cff515ddd\") " Feb 18 09:18:31 crc kubenswrapper[4556]: I0218 09:18:31.975801 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cb11270a-c754-44b2-b0b1-937cff515ddd-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "cb11270a-c754-44b2-b0b1-937cff515ddd" (UID: "cb11270a-c754-44b2-b0b1-937cff515ddd"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 09:18:31 crc kubenswrapper[4556]: I0218 09:18:31.975907 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb11270a-c754-44b2-b0b1-937cff515ddd-config-data\") pod \"cb11270a-c754-44b2-b0b1-937cff515ddd\" (UID: \"cb11270a-c754-44b2-b0b1-937cff515ddd\") " Feb 18 09:18:31 crc kubenswrapper[4556]: I0218 09:18:31.975992 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-46vwv\" (UniqueName: \"kubernetes.io/projected/cb11270a-c754-44b2-b0b1-937cff515ddd-kube-api-access-46vwv\") pod \"cb11270a-c754-44b2-b0b1-937cff515ddd\" (UID: \"cb11270a-c754-44b2-b0b1-937cff515ddd\") " Feb 18 09:18:31 crc kubenswrapper[4556]: I0218 09:18:31.976055 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb11270a-c754-44b2-b0b1-937cff515ddd-scripts\") pod \"cb11270a-c754-44b2-b0b1-937cff515ddd\" (UID: \"cb11270a-c754-44b2-b0b1-937cff515ddd\") " Feb 18 09:18:31 crc kubenswrapper[4556]: I0218 09:18:31.976131 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cb11270a-c754-44b2-b0b1-937cff515ddd-db-sync-config-data\") pod \"cb11270a-c754-44b2-b0b1-937cff515ddd\" (UID: \"cb11270a-c754-44b2-b0b1-937cff515ddd\") " Feb 18 09:18:31 crc kubenswrapper[4556]: I0218 09:18:31.976614 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/57873d95-dd1a-4b5c-99b6-459774c90acc-internal-tls-certs\") pod \"neutron-54664bdbc-cxnvs\" (UID: \"57873d95-dd1a-4b5c-99b6-459774c90acc\") " pod="openstack/neutron-54664bdbc-cxnvs" Feb 18 09:18:31 crc kubenswrapper[4556]: I0218 09:18:31.976699 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/57873d95-dd1a-4b5c-99b6-459774c90acc-public-tls-certs\") pod \"neutron-54664bdbc-cxnvs\" (UID: \"57873d95-dd1a-4b5c-99b6-459774c90acc\") " pod="openstack/neutron-54664bdbc-cxnvs" Feb 18 09:18:31 crc kubenswrapper[4556]: I0218 09:18:31.976759 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9sp7g\" (UniqueName: \"kubernetes.io/projected/57873d95-dd1a-4b5c-99b6-459774c90acc-kube-api-access-9sp7g\") pod \"neutron-54664bdbc-cxnvs\" (UID: \"57873d95-dd1a-4b5c-99b6-459774c90acc\") " pod="openstack/neutron-54664bdbc-cxnvs" Feb 18 09:18:31 crc kubenswrapper[4556]: I0218 09:18:31.976806 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57873d95-dd1a-4b5c-99b6-459774c90acc-combined-ca-bundle\") pod \"neutron-54664bdbc-cxnvs\" (UID: \"57873d95-dd1a-4b5c-99b6-459774c90acc\") " pod="openstack/neutron-54664bdbc-cxnvs" Feb 18 09:18:31 crc kubenswrapper[4556]: I0218 09:18:31.977001 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/57873d95-dd1a-4b5c-99b6-459774c90acc-ovndb-tls-certs\") pod \"neutron-54664bdbc-cxnvs\" (UID: \"57873d95-dd1a-4b5c-99b6-459774c90acc\") " pod="openstack/neutron-54664bdbc-cxnvs" Feb 18 09:18:31 crc kubenswrapper[4556]: I0218 09:18:31.977059 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/57873d95-dd1a-4b5c-99b6-459774c90acc-httpd-config\") pod \"neutron-54664bdbc-cxnvs\" (UID: \"57873d95-dd1a-4b5c-99b6-459774c90acc\") " pod="openstack/neutron-54664bdbc-cxnvs" Feb 18 09:18:31 crc kubenswrapper[4556]: I0218 09:18:31.977085 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/57873d95-dd1a-4b5c-99b6-459774c90acc-config\") pod \"neutron-54664bdbc-cxnvs\" (UID: \"57873d95-dd1a-4b5c-99b6-459774c90acc\") " pod="openstack/neutron-54664bdbc-cxnvs" Feb 18 09:18:31 crc kubenswrapper[4556]: I0218 09:18:31.977223 4556 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cb11270a-c754-44b2-b0b1-937cff515ddd-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:31 crc kubenswrapper[4556]: I0218 09:18:31.981182 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb11270a-c754-44b2-b0b1-937cff515ddd-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "cb11270a-c754-44b2-b0b1-937cff515ddd" (UID: "cb11270a-c754-44b2-b0b1-937cff515ddd"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:18:31 crc kubenswrapper[4556]: I0218 09:18:31.983095 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb11270a-c754-44b2-b0b1-937cff515ddd-scripts" (OuterVolumeSpecName: "scripts") pod "cb11270a-c754-44b2-b0b1-937cff515ddd" (UID: "cb11270a-c754-44b2-b0b1-937cff515ddd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:18:31 crc kubenswrapper[4556]: I0218 09:18:31.984488 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb11270a-c754-44b2-b0b1-937cff515ddd-kube-api-access-46vwv" (OuterVolumeSpecName: "kube-api-access-46vwv") pod "cb11270a-c754-44b2-b0b1-937cff515ddd" (UID: "cb11270a-c754-44b2-b0b1-937cff515ddd"). InnerVolumeSpecName "kube-api-access-46vwv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:18:32 crc kubenswrapper[4556]: I0218 09:18:32.001760 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb11270a-c754-44b2-b0b1-937cff515ddd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cb11270a-c754-44b2-b0b1-937cff515ddd" (UID: "cb11270a-c754-44b2-b0b1-937cff515ddd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:18:32 crc kubenswrapper[4556]: I0218 09:18:32.017988 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb11270a-c754-44b2-b0b1-937cff515ddd-config-data" (OuterVolumeSpecName: "config-data") pod "cb11270a-c754-44b2-b0b1-937cff515ddd" (UID: "cb11270a-c754-44b2-b0b1-937cff515ddd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:18:32 crc kubenswrapper[4556]: I0218 09:18:32.078560 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/57873d95-dd1a-4b5c-99b6-459774c90acc-public-tls-certs\") pod \"neutron-54664bdbc-cxnvs\" (UID: \"57873d95-dd1a-4b5c-99b6-459774c90acc\") " pod="openstack/neutron-54664bdbc-cxnvs" Feb 18 09:18:32 crc kubenswrapper[4556]: I0218 09:18:32.078613 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9sp7g\" (UniqueName: \"kubernetes.io/projected/57873d95-dd1a-4b5c-99b6-459774c90acc-kube-api-access-9sp7g\") pod \"neutron-54664bdbc-cxnvs\" (UID: \"57873d95-dd1a-4b5c-99b6-459774c90acc\") " pod="openstack/neutron-54664bdbc-cxnvs" Feb 18 09:18:32 crc kubenswrapper[4556]: I0218 09:18:32.078638 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57873d95-dd1a-4b5c-99b6-459774c90acc-combined-ca-bundle\") pod \"neutron-54664bdbc-cxnvs\" (UID: \"57873d95-dd1a-4b5c-99b6-459774c90acc\") " pod="openstack/neutron-54664bdbc-cxnvs" Feb 18 09:18:32 crc kubenswrapper[4556]: I0218 09:18:32.078698 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/57873d95-dd1a-4b5c-99b6-459774c90acc-ovndb-tls-certs\") pod \"neutron-54664bdbc-cxnvs\" (UID: \"57873d95-dd1a-4b5c-99b6-459774c90acc\") " pod="openstack/neutron-54664bdbc-cxnvs" Feb 18 09:18:32 crc kubenswrapper[4556]: I0218 09:18:32.078721 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/57873d95-dd1a-4b5c-99b6-459774c90acc-httpd-config\") pod \"neutron-54664bdbc-cxnvs\" (UID: \"57873d95-dd1a-4b5c-99b6-459774c90acc\") " pod="openstack/neutron-54664bdbc-cxnvs" Feb 18 09:18:32 crc kubenswrapper[4556]: I0218 09:18:32.078740 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/57873d95-dd1a-4b5c-99b6-459774c90acc-config\") pod \"neutron-54664bdbc-cxnvs\" (UID: \"57873d95-dd1a-4b5c-99b6-459774c90acc\") " pod="openstack/neutron-54664bdbc-cxnvs" Feb 18 09:18:32 crc kubenswrapper[4556]: I0218 09:18:32.079297 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/57873d95-dd1a-4b5c-99b6-459774c90acc-internal-tls-certs\") pod \"neutron-54664bdbc-cxnvs\" (UID: \"57873d95-dd1a-4b5c-99b6-459774c90acc\") " pod="openstack/neutron-54664bdbc-cxnvs" Feb 18 09:18:32 crc kubenswrapper[4556]: I0218 09:18:32.079358 4556 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb11270a-c754-44b2-b0b1-937cff515ddd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:32 crc kubenswrapper[4556]: I0218 09:18:32.079370 4556 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb11270a-c754-44b2-b0b1-937cff515ddd-config-data\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:32 crc kubenswrapper[4556]: I0218 09:18:32.079379 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-46vwv\" (UniqueName: \"kubernetes.io/projected/cb11270a-c754-44b2-b0b1-937cff515ddd-kube-api-access-46vwv\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:32 crc kubenswrapper[4556]: I0218 09:18:32.079390 4556 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb11270a-c754-44b2-b0b1-937cff515ddd-scripts\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:32 crc kubenswrapper[4556]: I0218 09:18:32.079398 4556 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cb11270a-c754-44b2-b0b1-937cff515ddd-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:32 crc kubenswrapper[4556]: I0218 09:18:32.082188 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/57873d95-dd1a-4b5c-99b6-459774c90acc-public-tls-certs\") pod \"neutron-54664bdbc-cxnvs\" (UID: \"57873d95-dd1a-4b5c-99b6-459774c90acc\") " pod="openstack/neutron-54664bdbc-cxnvs" Feb 18 09:18:32 crc kubenswrapper[4556]: I0218 09:18:32.083066 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/57873d95-dd1a-4b5c-99b6-459774c90acc-httpd-config\") pod \"neutron-54664bdbc-cxnvs\" (UID: \"57873d95-dd1a-4b5c-99b6-459774c90acc\") " pod="openstack/neutron-54664bdbc-cxnvs" Feb 18 09:18:32 crc kubenswrapper[4556]: I0218 09:18:32.083543 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/57873d95-dd1a-4b5c-99b6-459774c90acc-ovndb-tls-certs\") pod \"neutron-54664bdbc-cxnvs\" (UID: \"57873d95-dd1a-4b5c-99b6-459774c90acc\") " pod="openstack/neutron-54664bdbc-cxnvs" Feb 18 09:18:32 crc kubenswrapper[4556]: I0218 09:18:32.083649 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/57873d95-dd1a-4b5c-99b6-459774c90acc-internal-tls-certs\") pod \"neutron-54664bdbc-cxnvs\" (UID: \"57873d95-dd1a-4b5c-99b6-459774c90acc\") " pod="openstack/neutron-54664bdbc-cxnvs" Feb 18 09:18:32 crc kubenswrapper[4556]: I0218 09:18:32.083792 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/57873d95-dd1a-4b5c-99b6-459774c90acc-config\") pod \"neutron-54664bdbc-cxnvs\" (UID: \"57873d95-dd1a-4b5c-99b6-459774c90acc\") " pod="openstack/neutron-54664bdbc-cxnvs" Feb 18 09:18:32 crc kubenswrapper[4556]: I0218 09:18:32.084331 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57873d95-dd1a-4b5c-99b6-459774c90acc-combined-ca-bundle\") pod \"neutron-54664bdbc-cxnvs\" (UID: \"57873d95-dd1a-4b5c-99b6-459774c90acc\") " pod="openstack/neutron-54664bdbc-cxnvs" Feb 18 09:18:32 crc kubenswrapper[4556]: I0218 09:18:32.091889 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9sp7g\" (UniqueName: \"kubernetes.io/projected/57873d95-dd1a-4b5c-99b6-459774c90acc-kube-api-access-9sp7g\") pod \"neutron-54664bdbc-cxnvs\" (UID: \"57873d95-dd1a-4b5c-99b6-459774c90acc\") " pod="openstack/neutron-54664bdbc-cxnvs" Feb 18 09:18:32 crc kubenswrapper[4556]: I0218 09:18:32.259119 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-54664bdbc-cxnvs" Feb 18 09:18:32 crc kubenswrapper[4556]: I0218 09:18:32.612022 4556 generic.go:334] "Generic (PLEG): container finished" podID="b92f250f-e888-4ff5-ae26-9e31c7be3379" containerID="cc6725c2f27c13f0413ab0e4d8de88e810bf5e3b4b491a0eedf920bb5855de30" exitCode=0 Feb 18 09:18:32 crc kubenswrapper[4556]: I0218 09:18:32.612106 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6fd7956ddf-b647f" event={"ID":"b92f250f-e888-4ff5-ae26-9e31c7be3379","Type":"ContainerDied","Data":"cc6725c2f27c13f0413ab0e4d8de88e810bf5e3b4b491a0eedf920bb5855de30"} Feb 18 09:18:32 crc kubenswrapper[4556]: I0218 09:18:32.614619 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-dtq7t" event={"ID":"cb11270a-c754-44b2-b0b1-937cff515ddd","Type":"ContainerDied","Data":"95f6e23c65f73eadf9ca8c1d78cda62edfcc71e1c4a5eeee591612e6bd856108"} Feb 18 09:18:32 crc kubenswrapper[4556]: I0218 09:18:32.614650 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="95f6e23c65f73eadf9ca8c1d78cda62edfcc71e1c4a5eeee591612e6bd856108" Feb 18 09:18:32 crc kubenswrapper[4556]: I0218 09:18:32.614680 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-dtq7t" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.163730 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.167020 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.169608 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-kz8d4" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.171912 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.172193 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.172328 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.192618 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.204260 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhpsl\" (UniqueName: \"kubernetes.io/projected/59767f41-27b9-4146-b723-ea8df508a92e-kube-api-access-vhpsl\") pod \"cinder-scheduler-0\" (UID: \"59767f41-27b9-4146-b723-ea8df508a92e\") " pod="openstack/cinder-scheduler-0" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.204410 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/59767f41-27b9-4146-b723-ea8df508a92e-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"59767f41-27b9-4146-b723-ea8df508a92e\") " pod="openstack/cinder-scheduler-0" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.204447 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/59767f41-27b9-4146-b723-ea8df508a92e-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"59767f41-27b9-4146-b723-ea8df508a92e\") " pod="openstack/cinder-scheduler-0" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.204512 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59767f41-27b9-4146-b723-ea8df508a92e-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"59767f41-27b9-4146-b723-ea8df508a92e\") " pod="openstack/cinder-scheduler-0" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.204543 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59767f41-27b9-4146-b723-ea8df508a92e-scripts\") pod \"cinder-scheduler-0\" (UID: \"59767f41-27b9-4146-b723-ea8df508a92e\") " pod="openstack/cinder-scheduler-0" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.204570 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59767f41-27b9-4146-b723-ea8df508a92e-config-data\") pod \"cinder-scheduler-0\" (UID: \"59767f41-27b9-4146-b723-ea8df508a92e\") " pod="openstack/cinder-scheduler-0" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.306494 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59767f41-27b9-4146-b723-ea8df508a92e-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"59767f41-27b9-4146-b723-ea8df508a92e\") " pod="openstack/cinder-scheduler-0" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.306560 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59767f41-27b9-4146-b723-ea8df508a92e-scripts\") pod \"cinder-scheduler-0\" (UID: \"59767f41-27b9-4146-b723-ea8df508a92e\") " pod="openstack/cinder-scheduler-0" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.306603 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59767f41-27b9-4146-b723-ea8df508a92e-config-data\") pod \"cinder-scheduler-0\" (UID: \"59767f41-27b9-4146-b723-ea8df508a92e\") " pod="openstack/cinder-scheduler-0" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.306705 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhpsl\" (UniqueName: \"kubernetes.io/projected/59767f41-27b9-4146-b723-ea8df508a92e-kube-api-access-vhpsl\") pod \"cinder-scheduler-0\" (UID: \"59767f41-27b9-4146-b723-ea8df508a92e\") " pod="openstack/cinder-scheduler-0" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.306816 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/59767f41-27b9-4146-b723-ea8df508a92e-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"59767f41-27b9-4146-b723-ea8df508a92e\") " pod="openstack/cinder-scheduler-0" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.306858 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/59767f41-27b9-4146-b723-ea8df508a92e-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"59767f41-27b9-4146-b723-ea8df508a92e\") " pod="openstack/cinder-scheduler-0" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.307028 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/59767f41-27b9-4146-b723-ea8df508a92e-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"59767f41-27b9-4146-b723-ea8df508a92e\") " pod="openstack/cinder-scheduler-0" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.316667 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54c675984c-gfwjb"] Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.317624 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59767f41-27b9-4146-b723-ea8df508a92e-config-data\") pod \"cinder-scheduler-0\" (UID: \"59767f41-27b9-4146-b723-ea8df508a92e\") " pod="openstack/cinder-scheduler-0" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.319768 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/59767f41-27b9-4146-b723-ea8df508a92e-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"59767f41-27b9-4146-b723-ea8df508a92e\") " pod="openstack/cinder-scheduler-0" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.320243 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59767f41-27b9-4146-b723-ea8df508a92e-scripts\") pod \"cinder-scheduler-0\" (UID: \"59767f41-27b9-4146-b723-ea8df508a92e\") " pod="openstack/cinder-scheduler-0" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.321026 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59767f41-27b9-4146-b723-ea8df508a92e-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"59767f41-27b9-4146-b723-ea8df508a92e\") " pod="openstack/cinder-scheduler-0" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.322214 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-849fd69845-bdbc8"] Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.323622 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-849fd69845-bdbc8" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.341487 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhpsl\" (UniqueName: \"kubernetes.io/projected/59767f41-27b9-4146-b723-ea8df508a92e-kube-api-access-vhpsl\") pod \"cinder-scheduler-0\" (UID: \"59767f41-27b9-4146-b723-ea8df508a92e\") " pod="openstack/cinder-scheduler-0" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.369494 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-849fd69845-bdbc8"] Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.409488 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/129baf51-56c6-488a-88e2-90afcbb3d880-ovsdbserver-nb\") pod \"dnsmasq-dns-849fd69845-bdbc8\" (UID: \"129baf51-56c6-488a-88e2-90afcbb3d880\") " pod="openstack/dnsmasq-dns-849fd69845-bdbc8" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.409566 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zccq\" (UniqueName: \"kubernetes.io/projected/129baf51-56c6-488a-88e2-90afcbb3d880-kube-api-access-6zccq\") pod \"dnsmasq-dns-849fd69845-bdbc8\" (UID: \"129baf51-56c6-488a-88e2-90afcbb3d880\") " pod="openstack/dnsmasq-dns-849fd69845-bdbc8" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.409670 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/129baf51-56c6-488a-88e2-90afcbb3d880-ovsdbserver-sb\") pod \"dnsmasq-dns-849fd69845-bdbc8\" (UID: \"129baf51-56c6-488a-88e2-90afcbb3d880\") " pod="openstack/dnsmasq-dns-849fd69845-bdbc8" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.409771 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/129baf51-56c6-488a-88e2-90afcbb3d880-dns-svc\") pod \"dnsmasq-dns-849fd69845-bdbc8\" (UID: \"129baf51-56c6-488a-88e2-90afcbb3d880\") " pod="openstack/dnsmasq-dns-849fd69845-bdbc8" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.409980 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/129baf51-56c6-488a-88e2-90afcbb3d880-dns-swift-storage-0\") pod \"dnsmasq-dns-849fd69845-bdbc8\" (UID: \"129baf51-56c6-488a-88e2-90afcbb3d880\") " pod="openstack/dnsmasq-dns-849fd69845-bdbc8" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.410034 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/129baf51-56c6-488a-88e2-90afcbb3d880-config\") pod \"dnsmasq-dns-849fd69845-bdbc8\" (UID: \"129baf51-56c6-488a-88e2-90afcbb3d880\") " pod="openstack/dnsmasq-dns-849fd69845-bdbc8" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.495707 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.529082 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.531679 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.536385 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.542170 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/129baf51-56c6-488a-88e2-90afcbb3d880-dns-swift-storage-0\") pod \"dnsmasq-dns-849fd69845-bdbc8\" (UID: \"129baf51-56c6-488a-88e2-90afcbb3d880\") " pod="openstack/dnsmasq-dns-849fd69845-bdbc8" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.542225 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/129baf51-56c6-488a-88e2-90afcbb3d880-config\") pod \"dnsmasq-dns-849fd69845-bdbc8\" (UID: \"129baf51-56c6-488a-88e2-90afcbb3d880\") " pod="openstack/dnsmasq-dns-849fd69845-bdbc8" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.542306 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/129baf51-56c6-488a-88e2-90afcbb3d880-ovsdbserver-nb\") pod \"dnsmasq-dns-849fd69845-bdbc8\" (UID: \"129baf51-56c6-488a-88e2-90afcbb3d880\") " pod="openstack/dnsmasq-dns-849fd69845-bdbc8" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.542348 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zccq\" (UniqueName: \"kubernetes.io/projected/129baf51-56c6-488a-88e2-90afcbb3d880-kube-api-access-6zccq\") pod \"dnsmasq-dns-849fd69845-bdbc8\" (UID: \"129baf51-56c6-488a-88e2-90afcbb3d880\") " pod="openstack/dnsmasq-dns-849fd69845-bdbc8" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.542401 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/129baf51-56c6-488a-88e2-90afcbb3d880-ovsdbserver-sb\") pod \"dnsmasq-dns-849fd69845-bdbc8\" (UID: \"129baf51-56c6-488a-88e2-90afcbb3d880\") " pod="openstack/dnsmasq-dns-849fd69845-bdbc8" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.542459 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/129baf51-56c6-488a-88e2-90afcbb3d880-dns-svc\") pod \"dnsmasq-dns-849fd69845-bdbc8\" (UID: \"129baf51-56c6-488a-88e2-90afcbb3d880\") " pod="openstack/dnsmasq-dns-849fd69845-bdbc8" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.543177 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/129baf51-56c6-488a-88e2-90afcbb3d880-dns-swift-storage-0\") pod \"dnsmasq-dns-849fd69845-bdbc8\" (UID: \"129baf51-56c6-488a-88e2-90afcbb3d880\") " pod="openstack/dnsmasq-dns-849fd69845-bdbc8" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.543368 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/129baf51-56c6-488a-88e2-90afcbb3d880-dns-svc\") pod \"dnsmasq-dns-849fd69845-bdbc8\" (UID: \"129baf51-56c6-488a-88e2-90afcbb3d880\") " pod="openstack/dnsmasq-dns-849fd69845-bdbc8" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.543549 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/129baf51-56c6-488a-88e2-90afcbb3d880-ovsdbserver-nb\") pod \"dnsmasq-dns-849fd69845-bdbc8\" (UID: \"129baf51-56c6-488a-88e2-90afcbb3d880\") " pod="openstack/dnsmasq-dns-849fd69845-bdbc8" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.544034 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/129baf51-56c6-488a-88e2-90afcbb3d880-config\") pod \"dnsmasq-dns-849fd69845-bdbc8\" (UID: \"129baf51-56c6-488a-88e2-90afcbb3d880\") " pod="openstack/dnsmasq-dns-849fd69845-bdbc8" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.544363 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/129baf51-56c6-488a-88e2-90afcbb3d880-ovsdbserver-sb\") pod \"dnsmasq-dns-849fd69845-bdbc8\" (UID: \"129baf51-56c6-488a-88e2-90afcbb3d880\") " pod="openstack/dnsmasq-dns-849fd69845-bdbc8" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.549213 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.561339 4556 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-6fd7956ddf-b647f" podUID="b92f250f-e888-4ff5-ae26-9e31c7be3379" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.151:9696/\": dial tcp 10.217.0.151:9696: connect: connection refused" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.561916 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zccq\" (UniqueName: \"kubernetes.io/projected/129baf51-56c6-488a-88e2-90afcbb3d880-kube-api-access-6zccq\") pod \"dnsmasq-dns-849fd69845-bdbc8\" (UID: \"129baf51-56c6-488a-88e2-90afcbb3d880\") " pod="openstack/dnsmasq-dns-849fd69845-bdbc8" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.648257 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e15b6829-420b-493f-a31c-3bde1c53035f-config-data\") pod \"cinder-api-0\" (UID: \"e15b6829-420b-493f-a31c-3bde1c53035f\") " pod="openstack/cinder-api-0" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.649176 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e15b6829-420b-493f-a31c-3bde1c53035f-config-data-custom\") pod \"cinder-api-0\" (UID: \"e15b6829-420b-493f-a31c-3bde1c53035f\") " pod="openstack/cinder-api-0" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.649353 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e15b6829-420b-493f-a31c-3bde1c53035f-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e15b6829-420b-493f-a31c-3bde1c53035f\") " pod="openstack/cinder-api-0" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.649552 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e15b6829-420b-493f-a31c-3bde1c53035f-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e15b6829-420b-493f-a31c-3bde1c53035f\") " pod="openstack/cinder-api-0" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.649613 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e15b6829-420b-493f-a31c-3bde1c53035f-logs\") pod \"cinder-api-0\" (UID: \"e15b6829-420b-493f-a31c-3bde1c53035f\") " pod="openstack/cinder-api-0" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.649702 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrswm\" (UniqueName: \"kubernetes.io/projected/e15b6829-420b-493f-a31c-3bde1c53035f-kube-api-access-qrswm\") pod \"cinder-api-0\" (UID: \"e15b6829-420b-493f-a31c-3bde1c53035f\") " pod="openstack/cinder-api-0" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.650051 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e15b6829-420b-493f-a31c-3bde1c53035f-scripts\") pod \"cinder-api-0\" (UID: \"e15b6829-420b-493f-a31c-3bde1c53035f\") " pod="openstack/cinder-api-0" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.719560 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-849fd69845-bdbc8" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.753023 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e15b6829-420b-493f-a31c-3bde1c53035f-config-data\") pod \"cinder-api-0\" (UID: \"e15b6829-420b-493f-a31c-3bde1c53035f\") " pod="openstack/cinder-api-0" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.753148 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e15b6829-420b-493f-a31c-3bde1c53035f-config-data-custom\") pod \"cinder-api-0\" (UID: \"e15b6829-420b-493f-a31c-3bde1c53035f\") " pod="openstack/cinder-api-0" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.753887 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e15b6829-420b-493f-a31c-3bde1c53035f-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e15b6829-420b-493f-a31c-3bde1c53035f\") " pod="openstack/cinder-api-0" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.754028 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e15b6829-420b-493f-a31c-3bde1c53035f-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e15b6829-420b-493f-a31c-3bde1c53035f\") " pod="openstack/cinder-api-0" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.754068 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e15b6829-420b-493f-a31c-3bde1c53035f-logs\") pod \"cinder-api-0\" (UID: \"e15b6829-420b-493f-a31c-3bde1c53035f\") " pod="openstack/cinder-api-0" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.754091 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrswm\" (UniqueName: \"kubernetes.io/projected/e15b6829-420b-493f-a31c-3bde1c53035f-kube-api-access-qrswm\") pod \"cinder-api-0\" (UID: \"e15b6829-420b-493f-a31c-3bde1c53035f\") " pod="openstack/cinder-api-0" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.754117 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e15b6829-420b-493f-a31c-3bde1c53035f-scripts\") pod \"cinder-api-0\" (UID: \"e15b6829-420b-493f-a31c-3bde1c53035f\") " pod="openstack/cinder-api-0" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.754685 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e15b6829-420b-493f-a31c-3bde1c53035f-logs\") pod \"cinder-api-0\" (UID: \"e15b6829-420b-493f-a31c-3bde1c53035f\") " pod="openstack/cinder-api-0" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.754737 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e15b6829-420b-493f-a31c-3bde1c53035f-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e15b6829-420b-493f-a31c-3bde1c53035f\") " pod="openstack/cinder-api-0" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.759100 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e15b6829-420b-493f-a31c-3bde1c53035f-config-data-custom\") pod \"cinder-api-0\" (UID: \"e15b6829-420b-493f-a31c-3bde1c53035f\") " pod="openstack/cinder-api-0" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.760224 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e15b6829-420b-493f-a31c-3bde1c53035f-config-data\") pod \"cinder-api-0\" (UID: \"e15b6829-420b-493f-a31c-3bde1c53035f\") " pod="openstack/cinder-api-0" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.760382 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e15b6829-420b-493f-a31c-3bde1c53035f-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e15b6829-420b-493f-a31c-3bde1c53035f\") " pod="openstack/cinder-api-0" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.762571 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e15b6829-420b-493f-a31c-3bde1c53035f-scripts\") pod \"cinder-api-0\" (UID: \"e15b6829-420b-493f-a31c-3bde1c53035f\") " pod="openstack/cinder-api-0" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.770975 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrswm\" (UniqueName: \"kubernetes.io/projected/e15b6829-420b-493f-a31c-3bde1c53035f-kube-api-access-qrswm\") pod \"cinder-api-0\" (UID: \"e15b6829-420b-493f-a31c-3bde1c53035f\") " pod="openstack/cinder-api-0" Feb 18 09:18:33 crc kubenswrapper[4556]: I0218 09:18:33.849906 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 18 09:18:34 crc kubenswrapper[4556]: I0218 09:18:34.083291 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-54664bdbc-cxnvs"] Feb 18 09:18:34 crc kubenswrapper[4556]: W0218 09:18:34.116941 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod57873d95_dd1a_4b5c_99b6_459774c90acc.slice/crio-732c8f513925049f6ad36e9b1df02cd1424d1d5c1830a3122d9662777e2da30e WatchSource:0}: Error finding container 732c8f513925049f6ad36e9b1df02cd1424d1d5c1830a3122d9662777e2da30e: Status 404 returned error can't find the container with id 732c8f513925049f6ad36e9b1df02cd1424d1d5c1830a3122d9662777e2da30e Feb 18 09:18:34 crc kubenswrapper[4556]: I0218 09:18:34.122101 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-845cb5d46d-z5vhw"] Feb 18 09:18:34 crc kubenswrapper[4556]: I0218 09:18:34.531910 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-849fd69845-bdbc8"] Feb 18 09:18:34 crc kubenswrapper[4556]: W0218 09:18:34.553100 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod129baf51_56c6_488a_88e2_90afcbb3d880.slice/crio-3a83c3f218b20ae33af533dcd9f74f5c2e0878811b25c91c0b1e9c44aa7886e1 WatchSource:0}: Error finding container 3a83c3f218b20ae33af533dcd9f74f5c2e0878811b25c91c0b1e9c44aa7886e1: Status 404 returned error can't find the container with id 3a83c3f218b20ae33af533dcd9f74f5c2e0878811b25c91c0b1e9c44aa7886e1 Feb 18 09:18:34 crc kubenswrapper[4556]: I0218 09:18:34.639378 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54c675984c-gfwjb" event={"ID":"afe9fc26-12ca-41b2-8076-6d8374384041","Type":"ContainerStarted","Data":"32f2ece3ab754602a4b4eafc0d9033b7869f1b07a1b24c3730af4ff56d68d50a"} Feb 18 09:18:34 crc kubenswrapper[4556]: I0218 09:18:34.639482 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-54c675984c-gfwjb" podUID="afe9fc26-12ca-41b2-8076-6d8374384041" containerName="dnsmasq-dns" containerID="cri-o://32f2ece3ab754602a4b4eafc0d9033b7869f1b07a1b24c3730af4ff56d68d50a" gracePeriod=10 Feb 18 09:18:34 crc kubenswrapper[4556]: I0218 09:18:34.639613 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-54c675984c-gfwjb" Feb 18 09:18:34 crc kubenswrapper[4556]: I0218 09:18:34.680471 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e0739f72-891b-46e1-bce5-e17c54a407ab","Type":"ContainerStarted","Data":"c7b35b462aab32a2bd6aa0f4634c9496147ec2c072365494e87a9b8128d8030b"} Feb 18 09:18:34 crc kubenswrapper[4556]: I0218 09:18:34.680633 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e0739f72-891b-46e1-bce5-e17c54a407ab" containerName="ceilometer-central-agent" containerID="cri-o://5efdef38fecec015f3f0924c79231a8c7d9c34252e406c92aefa44ef006d639b" gracePeriod=30 Feb 18 09:18:34 crc kubenswrapper[4556]: I0218 09:18:34.680703 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 18 09:18:34 crc kubenswrapper[4556]: I0218 09:18:34.680741 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e0739f72-891b-46e1-bce5-e17c54a407ab" containerName="proxy-httpd" containerID="cri-o://c7b35b462aab32a2bd6aa0f4634c9496147ec2c072365494e87a9b8128d8030b" gracePeriod=30 Feb 18 09:18:34 crc kubenswrapper[4556]: I0218 09:18:34.680780 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e0739f72-891b-46e1-bce5-e17c54a407ab" containerName="sg-core" containerID="cri-o://f7367e0b2c468791b656052267d5c9e3f713321ffc299c0bae2758bf91f088f7" gracePeriod=30 Feb 18 09:18:34 crc kubenswrapper[4556]: I0218 09:18:34.680811 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e0739f72-891b-46e1-bce5-e17c54a407ab" containerName="ceilometer-notification-agent" containerID="cri-o://f644379761b7c18c34aaa29cf6ae6f5433a7bb623d8ff1f7d0fd70f9539f4a9f" gracePeriod=30 Feb 18 09:18:34 crc kubenswrapper[4556]: I0218 09:18:34.689797 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-54c675984c-gfwjb" podStartSLOduration=9.689785976 podStartE2EDuration="9.689785976s" podCreationTimestamp="2026-02-18 09:18:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:18:34.684045834 +0000 UTC m=+871.701006815" watchObservedRunningTime="2026-02-18 09:18:34.689785976 +0000 UTC m=+871.706746956" Feb 18 09:18:34 crc kubenswrapper[4556]: I0218 09:18:34.707671 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 18 09:18:34 crc kubenswrapper[4556]: I0218 09:18:34.729574 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-b64d8f579-7sj9h" event={"ID":"d6ef48e2-d690-4060-b013-5ac2be288161","Type":"ContainerStarted","Data":"c0fc7b2cf7d038a0ab840dac5bcfd09b855728aa91c5f554d186b7494c38c3ef"} Feb 18 09:18:34 crc kubenswrapper[4556]: I0218 09:18:34.731553 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.049184372 podStartE2EDuration="50.731528915s" podCreationTimestamp="2026-02-18 09:17:44 +0000 UTC" firstStartedPulling="2026-02-18 09:17:45.388688609 +0000 UTC m=+822.405649590" lastFinishedPulling="2026-02-18 09:18:34.071033153 +0000 UTC m=+871.087994133" observedRunningTime="2026-02-18 09:18:34.714238128 +0000 UTC m=+871.731199108" watchObservedRunningTime="2026-02-18 09:18:34.731528915 +0000 UTC m=+871.748489896" Feb 18 09:18:34 crc kubenswrapper[4556]: I0218 09:18:34.757724 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-54664bdbc-cxnvs" event={"ID":"57873d95-dd1a-4b5c-99b6-459774c90acc","Type":"ContainerStarted","Data":"732c8f513925049f6ad36e9b1df02cd1424d1d5c1830a3122d9662777e2da30e"} Feb 18 09:18:34 crc kubenswrapper[4556]: I0218 09:18:34.782395 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x6wtm" event={"ID":"1fb56420-2137-43d2-b790-1773daf3f0c2","Type":"ContainerStarted","Data":"83cf21b516d05fcd6e7f07cf698f60fce5b1a1ce77935ae3c4b37e86e37158bc"} Feb 18 09:18:34 crc kubenswrapper[4556]: I0218 09:18:34.790565 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-849fd69845-bdbc8" event={"ID":"129baf51-56c6-488a-88e2-90afcbb3d880","Type":"ContainerStarted","Data":"3a83c3f218b20ae33af533dcd9f74f5c2e0878811b25c91c0b1e9c44aa7886e1"} Feb 18 09:18:34 crc kubenswrapper[4556]: I0218 09:18:34.812410 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 18 09:18:34 crc kubenswrapper[4556]: I0218 09:18:34.813817 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-x6wtm" podStartSLOduration=2.167213566 podStartE2EDuration="9.813802464s" podCreationTimestamp="2026-02-18 09:18:25 +0000 UTC" firstStartedPulling="2026-02-18 09:18:26.496269938 +0000 UTC m=+863.513230919" lastFinishedPulling="2026-02-18 09:18:34.142858837 +0000 UTC m=+871.159819817" observedRunningTime="2026-02-18 09:18:34.802550403 +0000 UTC m=+871.819511382" watchObservedRunningTime="2026-02-18 09:18:34.813802464 +0000 UTC m=+871.830763444" Feb 18 09:18:34 crc kubenswrapper[4556]: I0218 09:18:34.818404 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5fc745f7dd-68srz" event={"ID":"8be9c112-7903-4ec0-a503-fb16c676caf4","Type":"ContainerStarted","Data":"56bc789666e226b39dfc16091ac2192d82a826917c359f1a03b7242d7928ee03"} Feb 18 09:18:34 crc kubenswrapper[4556]: I0218 09:18:34.844085 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-845cb5d46d-z5vhw" event={"ID":"ff67ebd8-094e-4c5b-b164-d6b37fc169de","Type":"ContainerStarted","Data":"9bd16494cd1ae5d3d9b34f142392dfa33ccccdead12384a30a1c4ae1590f9dd2"} Feb 18 09:18:34 crc kubenswrapper[4556]: I0218 09:18:34.844119 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-845cb5d46d-z5vhw" event={"ID":"ff67ebd8-094e-4c5b-b164-d6b37fc169de","Type":"ContainerStarted","Data":"5a5bea611ebc0177d1f037060ed794681686ac58299833793b657bfdb2d9238d"} Feb 18 09:18:34 crc kubenswrapper[4556]: I0218 09:18:34.868743 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-c6ccddcd4-hbvfc" event={"ID":"18385928-f2f7-4f58-9f83-e563af72c8b4","Type":"ContainerStarted","Data":"c50d1bd7e8cebf034249d811b5a1c2d743431d8dd4f4c67266c76f71686f721e"} Feb 18 09:18:34 crc kubenswrapper[4556]: I0218 09:18:34.869256 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-c6ccddcd4-hbvfc" Feb 18 09:18:34 crc kubenswrapper[4556]: I0218 09:18:34.869597 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-c6ccddcd4-hbvfc" Feb 18 09:18:34 crc kubenswrapper[4556]: I0218 09:18:34.881250 4556 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-c6ccddcd4-hbvfc" podUID="18385928-f2f7-4f58-9f83-e563af72c8b4" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": dial tcp 10.217.0.160:9311: connect: connection refused" Feb 18 09:18:34 crc kubenswrapper[4556]: I0218 09:18:34.908437 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-c6ccddcd4-hbvfc" podStartSLOduration=8.908423985 podStartE2EDuration="8.908423985s" podCreationTimestamp="2026-02-18 09:18:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:18:34.90729779 +0000 UTC m=+871.924258770" watchObservedRunningTime="2026-02-18 09:18:34.908423985 +0000 UTC m=+871.925384965" Feb 18 09:18:34 crc kubenswrapper[4556]: W0218 09:18:34.991544 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode15b6829_420b_493f_a31c_3bde1c53035f.slice/crio-b6353f56ef8b6ec4206c8a95be662cb4da864e7e58416f7d69100aa872d0c181 WatchSource:0}: Error finding container b6353f56ef8b6ec4206c8a95be662cb4da864e7e58416f7d69100aa872d0c181: Status 404 returned error can't find the container with id b6353f56ef8b6ec4206c8a95be662cb4da864e7e58416f7d69100aa872d0c181 Feb 18 09:18:35 crc kubenswrapper[4556]: I0218 09:18:35.101414 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Feb 18 09:18:35 crc kubenswrapper[4556]: I0218 09:18:35.348908 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54c675984c-gfwjb" Feb 18 09:18:35 crc kubenswrapper[4556]: I0218 09:18:35.415927 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-x6wtm" Feb 18 09:18:35 crc kubenswrapper[4556]: I0218 09:18:35.416079 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-x6wtm" Feb 18 09:18:35 crc kubenswrapper[4556]: I0218 09:18:35.516815 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/afe9fc26-12ca-41b2-8076-6d8374384041-ovsdbserver-sb\") pod \"afe9fc26-12ca-41b2-8076-6d8374384041\" (UID: \"afe9fc26-12ca-41b2-8076-6d8374384041\") " Feb 18 09:18:35 crc kubenswrapper[4556]: I0218 09:18:35.516886 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jgjdk\" (UniqueName: \"kubernetes.io/projected/afe9fc26-12ca-41b2-8076-6d8374384041-kube-api-access-jgjdk\") pod \"afe9fc26-12ca-41b2-8076-6d8374384041\" (UID: \"afe9fc26-12ca-41b2-8076-6d8374384041\") " Feb 18 09:18:35 crc kubenswrapper[4556]: I0218 09:18:35.516950 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/afe9fc26-12ca-41b2-8076-6d8374384041-dns-swift-storage-0\") pod \"afe9fc26-12ca-41b2-8076-6d8374384041\" (UID: \"afe9fc26-12ca-41b2-8076-6d8374384041\") " Feb 18 09:18:35 crc kubenswrapper[4556]: I0218 09:18:35.516966 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/afe9fc26-12ca-41b2-8076-6d8374384041-ovsdbserver-nb\") pod \"afe9fc26-12ca-41b2-8076-6d8374384041\" (UID: \"afe9fc26-12ca-41b2-8076-6d8374384041\") " Feb 18 09:18:35 crc kubenswrapper[4556]: I0218 09:18:35.517087 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/afe9fc26-12ca-41b2-8076-6d8374384041-dns-svc\") pod \"afe9fc26-12ca-41b2-8076-6d8374384041\" (UID: \"afe9fc26-12ca-41b2-8076-6d8374384041\") " Feb 18 09:18:35 crc kubenswrapper[4556]: I0218 09:18:35.517121 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/afe9fc26-12ca-41b2-8076-6d8374384041-config\") pod \"afe9fc26-12ca-41b2-8076-6d8374384041\" (UID: \"afe9fc26-12ca-41b2-8076-6d8374384041\") " Feb 18 09:18:35 crc kubenswrapper[4556]: I0218 09:18:35.529324 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afe9fc26-12ca-41b2-8076-6d8374384041-kube-api-access-jgjdk" (OuterVolumeSpecName: "kube-api-access-jgjdk") pod "afe9fc26-12ca-41b2-8076-6d8374384041" (UID: "afe9fc26-12ca-41b2-8076-6d8374384041"). InnerVolumeSpecName "kube-api-access-jgjdk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:18:35 crc kubenswrapper[4556]: I0218 09:18:35.620007 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jgjdk\" (UniqueName: \"kubernetes.io/projected/afe9fc26-12ca-41b2-8076-6d8374384041-kube-api-access-jgjdk\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:35 crc kubenswrapper[4556]: I0218 09:18:35.797423 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/afe9fc26-12ca-41b2-8076-6d8374384041-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "afe9fc26-12ca-41b2-8076-6d8374384041" (UID: "afe9fc26-12ca-41b2-8076-6d8374384041"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:18:35 crc kubenswrapper[4556]: I0218 09:18:35.836711 4556 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/afe9fc26-12ca-41b2-8076-6d8374384041-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:35 crc kubenswrapper[4556]: I0218 09:18:35.939521 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-54664bdbc-cxnvs" event={"ID":"57873d95-dd1a-4b5c-99b6-459774c90acc","Type":"ContainerStarted","Data":"8b4a391388cae2fec54af8781371efe905a0a211d25cb8fd69046934c9cb9fa5"} Feb 18 09:18:35 crc kubenswrapper[4556]: I0218 09:18:35.939566 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-54664bdbc-cxnvs" event={"ID":"57873d95-dd1a-4b5c-99b6-459774c90acc","Type":"ContainerStarted","Data":"b0231d1f039ea6a2f6b359d336460563f778b60dbda311b7619c37eea0227937"} Feb 18 09:18:35 crc kubenswrapper[4556]: I0218 09:18:35.939739 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-54664bdbc-cxnvs" Feb 18 09:18:35 crc kubenswrapper[4556]: I0218 09:18:35.970204 4556 generic.go:334] "Generic (PLEG): container finished" podID="129baf51-56c6-488a-88e2-90afcbb3d880" containerID="20742bc1bf20986ce09b83f67db4ef2c5890e69a52e453bf65c6c8a6b4b3a88b" exitCode=0 Feb 18 09:18:35 crc kubenswrapper[4556]: I0218 09:18:35.970290 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-849fd69845-bdbc8" event={"ID":"129baf51-56c6-488a-88e2-90afcbb3d880","Type":"ContainerDied","Data":"20742bc1bf20986ce09b83f67db4ef2c5890e69a52e453bf65c6c8a6b4b3a88b"} Feb 18 09:18:35 crc kubenswrapper[4556]: I0218 09:18:35.971521 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/afe9fc26-12ca-41b2-8076-6d8374384041-config" (OuterVolumeSpecName: "config") pod "afe9fc26-12ca-41b2-8076-6d8374384041" (UID: "afe9fc26-12ca-41b2-8076-6d8374384041"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:18:35 crc kubenswrapper[4556]: I0218 09:18:35.972764 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/afe9fc26-12ca-41b2-8076-6d8374384041-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "afe9fc26-12ca-41b2-8076-6d8374384041" (UID: "afe9fc26-12ca-41b2-8076-6d8374384041"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:18:35 crc kubenswrapper[4556]: I0218 09:18:35.978099 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-54664bdbc-cxnvs" podStartSLOduration=4.978072818 podStartE2EDuration="4.978072818s" podCreationTimestamp="2026-02-18 09:18:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:18:35.971498883 +0000 UTC m=+872.988459863" watchObservedRunningTime="2026-02-18 09:18:35.978072818 +0000 UTC m=+872.995033798" Feb 18 09:18:35 crc kubenswrapper[4556]: I0218 09:18:35.978365 4556 generic.go:334] "Generic (PLEG): container finished" podID="afe9fc26-12ca-41b2-8076-6d8374384041" containerID="32f2ece3ab754602a4b4eafc0d9033b7869f1b07a1b24c3730af4ff56d68d50a" exitCode=0 Feb 18 09:18:35 crc kubenswrapper[4556]: I0218 09:18:35.978441 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54c675984c-gfwjb" event={"ID":"afe9fc26-12ca-41b2-8076-6d8374384041","Type":"ContainerDied","Data":"32f2ece3ab754602a4b4eafc0d9033b7869f1b07a1b24c3730af4ff56d68d50a"} Feb 18 09:18:35 crc kubenswrapper[4556]: I0218 09:18:35.978478 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54c675984c-gfwjb" event={"ID":"afe9fc26-12ca-41b2-8076-6d8374384041","Type":"ContainerDied","Data":"e7b8ac5d5bd77858d1c7aae949f5f21ce98fc0418cbb25c273771cedd2212236"} Feb 18 09:18:35 crc kubenswrapper[4556]: I0218 09:18:35.978499 4556 scope.go:117] "RemoveContainer" containerID="32f2ece3ab754602a4b4eafc0d9033b7869f1b07a1b24c3730af4ff56d68d50a" Feb 18 09:18:35 crc kubenswrapper[4556]: I0218 09:18:35.978645 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54c675984c-gfwjb" Feb 18 09:18:35 crc kubenswrapper[4556]: I0218 09:18:35.988442 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5fc745f7dd-68srz" event={"ID":"8be9c112-7903-4ec0-a503-fb16c676caf4","Type":"ContainerStarted","Data":"77acf9fb7c028d109e636d0193e94c6864525f924fab4212e6303a149e8481ea"} Feb 18 09:18:36 crc kubenswrapper[4556]: I0218 09:18:36.021659 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/afe9fc26-12ca-41b2-8076-6d8374384041-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "afe9fc26-12ca-41b2-8076-6d8374384041" (UID: "afe9fc26-12ca-41b2-8076-6d8374384041"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:18:36 crc kubenswrapper[4556]: I0218 09:18:36.022280 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/afe9fc26-12ca-41b2-8076-6d8374384041-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "afe9fc26-12ca-41b2-8076-6d8374384041" (UID: "afe9fc26-12ca-41b2-8076-6d8374384041"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:18:36 crc kubenswrapper[4556]: I0218 09:18:36.025981 4556 generic.go:334] "Generic (PLEG): container finished" podID="e0739f72-891b-46e1-bce5-e17c54a407ab" containerID="f7367e0b2c468791b656052267d5c9e3f713321ffc299c0bae2758bf91f088f7" exitCode=2 Feb 18 09:18:36 crc kubenswrapper[4556]: I0218 09:18:36.026015 4556 generic.go:334] "Generic (PLEG): container finished" podID="e0739f72-891b-46e1-bce5-e17c54a407ab" containerID="5efdef38fecec015f3f0924c79231a8c7d9c34252e406c92aefa44ef006d639b" exitCode=0 Feb 18 09:18:36 crc kubenswrapper[4556]: I0218 09:18:36.026066 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e0739f72-891b-46e1-bce5-e17c54a407ab","Type":"ContainerDied","Data":"f7367e0b2c468791b656052267d5c9e3f713321ffc299c0bae2758bf91f088f7"} Feb 18 09:18:36 crc kubenswrapper[4556]: I0218 09:18:36.026091 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e0739f72-891b-46e1-bce5-e17c54a407ab","Type":"ContainerDied","Data":"5efdef38fecec015f3f0924c79231a8c7d9c34252e406c92aefa44ef006d639b"} Feb 18 09:18:36 crc kubenswrapper[4556]: I0218 09:18:36.028268 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-5fc745f7dd-68srz" podStartSLOduration=4.307606117 podStartE2EDuration="11.028251705s" podCreationTimestamp="2026-02-18 09:18:25 +0000 UTC" firstStartedPulling="2026-02-18 09:18:26.833476295 +0000 UTC m=+863.850437275" lastFinishedPulling="2026-02-18 09:18:33.554121883 +0000 UTC m=+870.571082863" observedRunningTime="2026-02-18 09:18:36.021184429 +0000 UTC m=+873.038145409" watchObservedRunningTime="2026-02-18 09:18:36.028251705 +0000 UTC m=+873.045212685" Feb 18 09:18:36 crc kubenswrapper[4556]: I0218 09:18:36.033014 4556 scope.go:117] "RemoveContainer" containerID="e8b71066786b1a917aea7665cd2bc13916cae20786320151dc408f88a6dc3b9b" Feb 18 09:18:36 crc kubenswrapper[4556]: I0218 09:18:36.038781 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-b64d8f579-7sj9h" event={"ID":"d6ef48e2-d690-4060-b013-5ac2be288161","Type":"ContainerStarted","Data":"e601d06ff5a8e883f3184cafdafe57654ae5befce0f5cb7c22a44cc1fce2f7f1"} Feb 18 09:18:36 crc kubenswrapper[4556]: I0218 09:18:36.043438 4556 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/afe9fc26-12ca-41b2-8076-6d8374384041-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:36 crc kubenswrapper[4556]: I0218 09:18:36.043465 4556 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/afe9fc26-12ca-41b2-8076-6d8374384041-config\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:36 crc kubenswrapper[4556]: I0218 09:18:36.043476 4556 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/afe9fc26-12ca-41b2-8076-6d8374384041-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:36 crc kubenswrapper[4556]: I0218 09:18:36.043486 4556 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/afe9fc26-12ca-41b2-8076-6d8374384041-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:36 crc kubenswrapper[4556]: I0218 09:18:36.052229 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e15b6829-420b-493f-a31c-3bde1c53035f","Type":"ContainerStarted","Data":"b6353f56ef8b6ec4206c8a95be662cb4da864e7e58416f7d69100aa872d0c181"} Feb 18 09:18:36 crc kubenswrapper[4556]: I0218 09:18:36.062753 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-b64d8f579-7sj9h" podStartSLOduration=3.92531761 podStartE2EDuration="11.062738412s" podCreationTimestamp="2026-02-18 09:18:25 +0000 UTC" firstStartedPulling="2026-02-18 09:18:26.997312954 +0000 UTC m=+864.014273933" lastFinishedPulling="2026-02-18 09:18:34.134733755 +0000 UTC m=+871.151694735" observedRunningTime="2026-02-18 09:18:36.054310898 +0000 UTC m=+873.071271879" watchObservedRunningTime="2026-02-18 09:18:36.062738412 +0000 UTC m=+873.079699391" Feb 18 09:18:36 crc kubenswrapper[4556]: I0218 09:18:36.068661 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"59767f41-27b9-4146-b723-ea8df508a92e","Type":"ContainerStarted","Data":"ea8e05987e69fa9665be060865cd530276a85b3405d4d7496e36384c8ea914e4"} Feb 18 09:18:36 crc kubenswrapper[4556]: I0218 09:18:36.227317 4556 scope.go:117] "RemoveContainer" containerID="32f2ece3ab754602a4b4eafc0d9033b7869f1b07a1b24c3730af4ff56d68d50a" Feb 18 09:18:36 crc kubenswrapper[4556]: E0218 09:18:36.230679 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32f2ece3ab754602a4b4eafc0d9033b7869f1b07a1b24c3730af4ff56d68d50a\": container with ID starting with 32f2ece3ab754602a4b4eafc0d9033b7869f1b07a1b24c3730af4ff56d68d50a not found: ID does not exist" containerID="32f2ece3ab754602a4b4eafc0d9033b7869f1b07a1b24c3730af4ff56d68d50a" Feb 18 09:18:36 crc kubenswrapper[4556]: I0218 09:18:36.230709 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32f2ece3ab754602a4b4eafc0d9033b7869f1b07a1b24c3730af4ff56d68d50a"} err="failed to get container status \"32f2ece3ab754602a4b4eafc0d9033b7869f1b07a1b24c3730af4ff56d68d50a\": rpc error: code = NotFound desc = could not find container \"32f2ece3ab754602a4b4eafc0d9033b7869f1b07a1b24c3730af4ff56d68d50a\": container with ID starting with 32f2ece3ab754602a4b4eafc0d9033b7869f1b07a1b24c3730af4ff56d68d50a not found: ID does not exist" Feb 18 09:18:36 crc kubenswrapper[4556]: I0218 09:18:36.230731 4556 scope.go:117] "RemoveContainer" containerID="e8b71066786b1a917aea7665cd2bc13916cae20786320151dc408f88a6dc3b9b" Feb 18 09:18:36 crc kubenswrapper[4556]: E0218 09:18:36.231649 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8b71066786b1a917aea7665cd2bc13916cae20786320151dc408f88a6dc3b9b\": container with ID starting with e8b71066786b1a917aea7665cd2bc13916cae20786320151dc408f88a6dc3b9b not found: ID does not exist" containerID="e8b71066786b1a917aea7665cd2bc13916cae20786320151dc408f88a6dc3b9b" Feb 18 09:18:36 crc kubenswrapper[4556]: I0218 09:18:36.231671 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8b71066786b1a917aea7665cd2bc13916cae20786320151dc408f88a6dc3b9b"} err="failed to get container status \"e8b71066786b1a917aea7665cd2bc13916cae20786320151dc408f88a6dc3b9b\": rpc error: code = NotFound desc = could not find container \"e8b71066786b1a917aea7665cd2bc13916cae20786320151dc408f88a6dc3b9b\": container with ID starting with e8b71066786b1a917aea7665cd2bc13916cae20786320151dc408f88a6dc3b9b not found: ID does not exist" Feb 18 09:18:36 crc kubenswrapper[4556]: I0218 09:18:36.310075 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54c675984c-gfwjb"] Feb 18 09:18:36 crc kubenswrapper[4556]: I0218 09:18:36.331322 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-54c675984c-gfwjb"] Feb 18 09:18:36 crc kubenswrapper[4556]: I0218 09:18:36.431545 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-85bb8c66f4-fhblh" Feb 18 09:18:36 crc kubenswrapper[4556]: I0218 09:18:36.567768 4556 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-x6wtm" podUID="1fb56420-2137-43d2-b790-1773daf3f0c2" containerName="registry-server" probeResult="failure" output=< Feb 18 09:18:36 crc kubenswrapper[4556]: timeout: failed to connect service ":50051" within 1s Feb 18 09:18:36 crc kubenswrapper[4556]: > Feb 18 09:18:37 crc kubenswrapper[4556]: I0218 09:18:37.080467 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-849fd69845-bdbc8" event={"ID":"129baf51-56c6-488a-88e2-90afcbb3d880","Type":"ContainerStarted","Data":"6cbad85d5780b1b4e977194f87ef2f9451600f0524aecf9002beb6201ffcc69f"} Feb 18 09:18:37 crc kubenswrapper[4556]: I0218 09:18:37.099660 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-845cb5d46d-z5vhw" event={"ID":"ff67ebd8-094e-4c5b-b164-d6b37fc169de","Type":"ContainerStarted","Data":"cca6de5c4eee478603a34ef76359d94e5d4e40d05c2bd769a864907612a179b6"} Feb 18 09:18:37 crc kubenswrapper[4556]: I0218 09:18:37.099913 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-845cb5d46d-z5vhw" Feb 18 09:18:37 crc kubenswrapper[4556]: I0218 09:18:37.108804 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e15b6829-420b-493f-a31c-3bde1c53035f","Type":"ContainerStarted","Data":"9d80e4b26b777401843d9382d2ebfd2f637f602a3747547b652ad119072af823"} Feb 18 09:18:37 crc kubenswrapper[4556]: I0218 09:18:37.108855 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e15b6829-420b-493f-a31c-3bde1c53035f","Type":"ContainerStarted","Data":"d67852ff68b4610cf54e2b45a423c4b0b67eec6e82b39b44a3cf5c7922d54615"} Feb 18 09:18:37 crc kubenswrapper[4556]: I0218 09:18:37.108968 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="e15b6829-420b-493f-a31c-3bde1c53035f" containerName="cinder-api-log" containerID="cri-o://d67852ff68b4610cf54e2b45a423c4b0b67eec6e82b39b44a3cf5c7922d54615" gracePeriod=30 Feb 18 09:18:37 crc kubenswrapper[4556]: I0218 09:18:37.109255 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Feb 18 09:18:37 crc kubenswrapper[4556]: I0218 09:18:37.109296 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="e15b6829-420b-493f-a31c-3bde1c53035f" containerName="cinder-api" containerID="cri-o://9d80e4b26b777401843d9382d2ebfd2f637f602a3747547b652ad119072af823" gracePeriod=30 Feb 18 09:18:37 crc kubenswrapper[4556]: I0218 09:18:37.114239 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"59767f41-27b9-4146-b723-ea8df508a92e","Type":"ContainerStarted","Data":"ed41c2e1779bc2ff90abc289c7c836a56a67c4f7e881c290aa073046688b2f99"} Feb 18 09:18:37 crc kubenswrapper[4556]: I0218 09:18:37.119614 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-849fd69845-bdbc8" podStartSLOduration=4.119598722 podStartE2EDuration="4.119598722s" podCreationTimestamp="2026-02-18 09:18:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:18:37.100164189 +0000 UTC m=+874.117125168" watchObservedRunningTime="2026-02-18 09:18:37.119598722 +0000 UTC m=+874.136559702" Feb 18 09:18:37 crc kubenswrapper[4556]: I0218 09:18:37.120002 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-845cb5d46d-z5vhw" podStartSLOduration=9.119997475 podStartE2EDuration="9.119997475s" podCreationTimestamp="2026-02-18 09:18:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:18:37.118467768 +0000 UTC m=+874.135428748" watchObservedRunningTime="2026-02-18 09:18:37.119997475 +0000 UTC m=+874.136958455" Feb 18 09:18:37 crc kubenswrapper[4556]: I0218 09:18:37.138234 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.13821999 podStartE2EDuration="4.13821999s" podCreationTimestamp="2026-02-18 09:18:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:18:37.135120421 +0000 UTC m=+874.152081402" watchObservedRunningTime="2026-02-18 09:18:37.13821999 +0000 UTC m=+874.155180970" Feb 18 09:18:37 crc kubenswrapper[4556]: I0218 09:18:37.201180 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-595566bb8b-6b5h6" Feb 18 09:18:37 crc kubenswrapper[4556]: I0218 09:18:37.323085 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="afe9fc26-12ca-41b2-8076-6d8374384041" path="/var/lib/kubelet/pods/afe9fc26-12ca-41b2-8076-6d8374384041/volumes" Feb 18 09:18:38 crc kubenswrapper[4556]: I0218 09:18:38.123478 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"59767f41-27b9-4146-b723-ea8df508a92e","Type":"ContainerStarted","Data":"5bcd75d3aae13e3e94b1fa31eb135277d4a4343b64ada8a1ab708f2ff72fad1e"} Feb 18 09:18:38 crc kubenswrapper[4556]: I0218 09:18:38.126871 4556 generic.go:334] "Generic (PLEG): container finished" podID="e0739f72-891b-46e1-bce5-e17c54a407ab" containerID="f644379761b7c18c34aaa29cf6ae6f5433a7bb623d8ff1f7d0fd70f9539f4a9f" exitCode=0 Feb 18 09:18:38 crc kubenswrapper[4556]: I0218 09:18:38.126922 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e0739f72-891b-46e1-bce5-e17c54a407ab","Type":"ContainerDied","Data":"f644379761b7c18c34aaa29cf6ae6f5433a7bb623d8ff1f7d0fd70f9539f4a9f"} Feb 18 09:18:38 crc kubenswrapper[4556]: I0218 09:18:38.129034 4556 generic.go:334] "Generic (PLEG): container finished" podID="e15b6829-420b-493f-a31c-3bde1c53035f" containerID="d67852ff68b4610cf54e2b45a423c4b0b67eec6e82b39b44a3cf5c7922d54615" exitCode=143 Feb 18 09:18:38 crc kubenswrapper[4556]: I0218 09:18:38.129275 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e15b6829-420b-493f-a31c-3bde1c53035f","Type":"ContainerDied","Data":"d67852ff68b4610cf54e2b45a423c4b0b67eec6e82b39b44a3cf5c7922d54615"} Feb 18 09:18:38 crc kubenswrapper[4556]: I0218 09:18:38.129573 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-845cb5d46d-z5vhw" Feb 18 09:18:38 crc kubenswrapper[4556]: I0218 09:18:38.129597 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-849fd69845-bdbc8" Feb 18 09:18:38 crc kubenswrapper[4556]: I0218 09:18:38.146264 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.8601075639999998 podStartE2EDuration="5.146253195s" podCreationTimestamp="2026-02-18 09:18:33 +0000 UTC" firstStartedPulling="2026-02-18 09:18:34.782423302 +0000 UTC m=+871.799384282" lastFinishedPulling="2026-02-18 09:18:36.068568934 +0000 UTC m=+873.085529913" observedRunningTime="2026-02-18 09:18:38.144052892 +0000 UTC m=+875.161013872" watchObservedRunningTime="2026-02-18 09:18:38.146253195 +0000 UTC m=+875.163214174" Feb 18 09:18:38 crc kubenswrapper[4556]: I0218 09:18:38.269620 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-c6ccddcd4-hbvfc" Feb 18 09:18:38 crc kubenswrapper[4556]: I0218 09:18:38.408533 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-85bb8c66f4-fhblh" Feb 18 09:18:38 crc kubenswrapper[4556]: I0218 09:18:38.497208 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Feb 18 09:18:38 crc kubenswrapper[4556]: I0218 09:18:38.815353 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-595566bb8b-6b5h6" Feb 18 09:18:38 crc kubenswrapper[4556]: I0218 09:18:38.886098 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-85bb8c66f4-fhblh"] Feb 18 09:18:39 crc kubenswrapper[4556]: I0218 09:18:39.136531 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-85bb8c66f4-fhblh" podUID="3354b930-6b00-4aa7-a4bd-97ec410cc863" containerName="horizon-log" containerID="cri-o://c1df47a378b25fad3361d8a29971e5ef0d47d8783acc451f619234849683c0f1" gracePeriod=30 Feb 18 09:18:39 crc kubenswrapper[4556]: I0218 09:18:39.136560 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-85bb8c66f4-fhblh" podUID="3354b930-6b00-4aa7-a4bd-97ec410cc863" containerName="horizon" containerID="cri-o://4423fda867205940e352d08e4c633d2caee4bc72b78e6280dd5fd52606caa4dd" gracePeriod=30 Feb 18 09:18:39 crc kubenswrapper[4556]: I0218 09:18:39.978184 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6fd7956ddf-b647f" Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.147826 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b92f250f-e888-4ff5-ae26-9e31c7be3379-internal-tls-certs\") pod \"b92f250f-e888-4ff5-ae26-9e31c7be3379\" (UID: \"b92f250f-e888-4ff5-ae26-9e31c7be3379\") " Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.148121 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b92f250f-e888-4ff5-ae26-9e31c7be3379-ovndb-tls-certs\") pod \"b92f250f-e888-4ff5-ae26-9e31c7be3379\" (UID: \"b92f250f-e888-4ff5-ae26-9e31c7be3379\") " Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.148199 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b92f250f-e888-4ff5-ae26-9e31c7be3379-public-tls-certs\") pod \"b92f250f-e888-4ff5-ae26-9e31c7be3379\" (UID: \"b92f250f-e888-4ff5-ae26-9e31c7be3379\") " Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.148224 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b92f250f-e888-4ff5-ae26-9e31c7be3379-config\") pod \"b92f250f-e888-4ff5-ae26-9e31c7be3379\" (UID: \"b92f250f-e888-4ff5-ae26-9e31c7be3379\") " Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.148309 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-czjtk\" (UniqueName: \"kubernetes.io/projected/b92f250f-e888-4ff5-ae26-9e31c7be3379-kube-api-access-czjtk\") pod \"b92f250f-e888-4ff5-ae26-9e31c7be3379\" (UID: \"b92f250f-e888-4ff5-ae26-9e31c7be3379\") " Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.148348 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b92f250f-e888-4ff5-ae26-9e31c7be3379-combined-ca-bundle\") pod \"b92f250f-e888-4ff5-ae26-9e31c7be3379\" (UID: \"b92f250f-e888-4ff5-ae26-9e31c7be3379\") " Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.148417 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b92f250f-e888-4ff5-ae26-9e31c7be3379-httpd-config\") pod \"b92f250f-e888-4ff5-ae26-9e31c7be3379\" (UID: \"b92f250f-e888-4ff5-ae26-9e31c7be3379\") " Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.158739 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b92f250f-e888-4ff5-ae26-9e31c7be3379-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "b92f250f-e888-4ff5-ae26-9e31c7be3379" (UID: "b92f250f-e888-4ff5-ae26-9e31c7be3379"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.158879 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b92f250f-e888-4ff5-ae26-9e31c7be3379-kube-api-access-czjtk" (OuterVolumeSpecName: "kube-api-access-czjtk") pod "b92f250f-e888-4ff5-ae26-9e31c7be3379" (UID: "b92f250f-e888-4ff5-ae26-9e31c7be3379"). InnerVolumeSpecName "kube-api-access-czjtk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.160740 4556 generic.go:334] "Generic (PLEG): container finished" podID="66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98" containerID="2973401c6ebe4b90a1471e53a5013aae7d75364f2830c2e5cd584b5f74812ad2" exitCode=137 Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.160757 4556 generic.go:334] "Generic (PLEG): container finished" podID="66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98" containerID="ace36a61edb3ea0808af8dd243149192bdb820e0dff28b907b17b501b9e6c790" exitCode=137 Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.160790 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-77cf5996cc-dj8s4" event={"ID":"66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98","Type":"ContainerDied","Data":"2973401c6ebe4b90a1471e53a5013aae7d75364f2830c2e5cd584b5f74812ad2"} Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.160814 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-77cf5996cc-dj8s4" event={"ID":"66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98","Type":"ContainerDied","Data":"ace36a61edb3ea0808af8dd243149192bdb820e0dff28b907b17b501b9e6c790"} Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.162182 4556 generic.go:334] "Generic (PLEG): container finished" podID="c5fec43c-7b17-45d5-91c4-609b6d2ca1b9" containerID="a6e43edd5803076f4c76dc543e9ef0bdba7eda69af36b2cf7df94a0200eacf09" exitCode=137 Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.162194 4556 generic.go:334] "Generic (PLEG): container finished" podID="c5fec43c-7b17-45d5-91c4-609b6d2ca1b9" containerID="2fed1042675c8bc639ff1bf40020b12a7f5f553be9bb216173f3fbe9c18aa7dd" exitCode=137 Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.162221 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79dd98bf77-fdbv7" event={"ID":"c5fec43c-7b17-45d5-91c4-609b6d2ca1b9","Type":"ContainerDied","Data":"a6e43edd5803076f4c76dc543e9ef0bdba7eda69af36b2cf7df94a0200eacf09"} Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.162235 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79dd98bf77-fdbv7" event={"ID":"c5fec43c-7b17-45d5-91c4-609b6d2ca1b9","Type":"ContainerDied","Data":"2fed1042675c8bc639ff1bf40020b12a7f5f553be9bb216173f3fbe9c18aa7dd"} Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.163382 4556 generic.go:334] "Generic (PLEG): container finished" podID="50228011-f908-465e-a754-d4b6b626dda9" containerID="c354f6bc02d120ffecda388ac1244fdc7228debcaed10748429b9072f8e42df2" exitCode=137 Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.163393 4556 generic.go:334] "Generic (PLEG): container finished" podID="50228011-f908-465e-a754-d4b6b626dda9" containerID="7147437f196486d78c3232e03d579055aba6c510b6715c9264fd4d2b04e9089e" exitCode=137 Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.163418 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-74c5b6b6c5-q8djs" event={"ID":"50228011-f908-465e-a754-d4b6b626dda9","Type":"ContainerDied","Data":"c354f6bc02d120ffecda388ac1244fdc7228debcaed10748429b9072f8e42df2"} Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.163435 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-74c5b6b6c5-q8djs" event={"ID":"50228011-f908-465e-a754-d4b6b626dda9","Type":"ContainerDied","Data":"7147437f196486d78c3232e03d579055aba6c510b6715c9264fd4d2b04e9089e"} Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.170068 4556 generic.go:334] "Generic (PLEG): container finished" podID="b92f250f-e888-4ff5-ae26-9e31c7be3379" containerID="20b200abdb599e8ec5463ea4f53a30e71450839bf07dba61fec5a4eb01b348dd" exitCode=0 Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.170518 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6fd7956ddf-b647f" Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.171467 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6fd7956ddf-b647f" event={"ID":"b92f250f-e888-4ff5-ae26-9e31c7be3379","Type":"ContainerDied","Data":"20b200abdb599e8ec5463ea4f53a30e71450839bf07dba61fec5a4eb01b348dd"} Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.171522 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6fd7956ddf-b647f" event={"ID":"b92f250f-e888-4ff5-ae26-9e31c7be3379","Type":"ContainerDied","Data":"4cf48eebb92b1e8d95ce8b8111c230c3a42d8ec80542347be535fc1a082eee1b"} Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.171542 4556 scope.go:117] "RemoveContainer" containerID="cc6725c2f27c13f0413ab0e4d8de88e810bf5e3b4b491a0eedf920bb5855de30" Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.225409 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b92f250f-e888-4ff5-ae26-9e31c7be3379-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b92f250f-e888-4ff5-ae26-9e31c7be3379" (UID: "b92f250f-e888-4ff5-ae26-9e31c7be3379"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.230249 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b92f250f-e888-4ff5-ae26-9e31c7be3379-config" (OuterVolumeSpecName: "config") pod "b92f250f-e888-4ff5-ae26-9e31c7be3379" (UID: "b92f250f-e888-4ff5-ae26-9e31c7be3379"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.230804 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b92f250f-e888-4ff5-ae26-9e31c7be3379-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "b92f250f-e888-4ff5-ae26-9e31c7be3379" (UID: "b92f250f-e888-4ff5-ae26-9e31c7be3379"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.250646 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-czjtk\" (UniqueName: \"kubernetes.io/projected/b92f250f-e888-4ff5-ae26-9e31c7be3379-kube-api-access-czjtk\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.250800 4556 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b92f250f-e888-4ff5-ae26-9e31c7be3379-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.250869 4556 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b92f250f-e888-4ff5-ae26-9e31c7be3379-httpd-config\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.250928 4556 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b92f250f-e888-4ff5-ae26-9e31c7be3379-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.250979 4556 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/b92f250f-e888-4ff5-ae26-9e31c7be3379-config\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.257486 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b92f250f-e888-4ff5-ae26-9e31c7be3379-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "b92f250f-e888-4ff5-ae26-9e31c7be3379" (UID: "b92f250f-e888-4ff5-ae26-9e31c7be3379"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.259024 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b92f250f-e888-4ff5-ae26-9e31c7be3379-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "b92f250f-e888-4ff5-ae26-9e31c7be3379" (UID: "b92f250f-e888-4ff5-ae26-9e31c7be3379"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.313742 4556 scope.go:117] "RemoveContainer" containerID="20b200abdb599e8ec5463ea4f53a30e71450839bf07dba61fec5a4eb01b348dd" Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.338335 4556 scope.go:117] "RemoveContainer" containerID="cc6725c2f27c13f0413ab0e4d8de88e810bf5e3b4b491a0eedf920bb5855de30" Feb 18 09:18:40 crc kubenswrapper[4556]: E0218 09:18:40.340394 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc6725c2f27c13f0413ab0e4d8de88e810bf5e3b4b491a0eedf920bb5855de30\": container with ID starting with cc6725c2f27c13f0413ab0e4d8de88e810bf5e3b4b491a0eedf920bb5855de30 not found: ID does not exist" containerID="cc6725c2f27c13f0413ab0e4d8de88e810bf5e3b4b491a0eedf920bb5855de30" Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.340432 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc6725c2f27c13f0413ab0e4d8de88e810bf5e3b4b491a0eedf920bb5855de30"} err="failed to get container status \"cc6725c2f27c13f0413ab0e4d8de88e810bf5e3b4b491a0eedf920bb5855de30\": rpc error: code = NotFound desc = could not find container \"cc6725c2f27c13f0413ab0e4d8de88e810bf5e3b4b491a0eedf920bb5855de30\": container with ID starting with cc6725c2f27c13f0413ab0e4d8de88e810bf5e3b4b491a0eedf920bb5855de30 not found: ID does not exist" Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.340491 4556 scope.go:117] "RemoveContainer" containerID="20b200abdb599e8ec5463ea4f53a30e71450839bf07dba61fec5a4eb01b348dd" Feb 18 09:18:40 crc kubenswrapper[4556]: E0218 09:18:40.343374 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20b200abdb599e8ec5463ea4f53a30e71450839bf07dba61fec5a4eb01b348dd\": container with ID starting with 20b200abdb599e8ec5463ea4f53a30e71450839bf07dba61fec5a4eb01b348dd not found: ID does not exist" containerID="20b200abdb599e8ec5463ea4f53a30e71450839bf07dba61fec5a4eb01b348dd" Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.343411 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20b200abdb599e8ec5463ea4f53a30e71450839bf07dba61fec5a4eb01b348dd"} err="failed to get container status \"20b200abdb599e8ec5463ea4f53a30e71450839bf07dba61fec5a4eb01b348dd\": rpc error: code = NotFound desc = could not find container \"20b200abdb599e8ec5463ea4f53a30e71450839bf07dba61fec5a4eb01b348dd\": container with ID starting with 20b200abdb599e8ec5463ea4f53a30e71450839bf07dba61fec5a4eb01b348dd not found: ID does not exist" Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.352676 4556 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b92f250f-e888-4ff5-ae26-9e31c7be3379-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.352704 4556 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b92f250f-e888-4ff5-ae26-9e31c7be3379-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.470564 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-74c5b6b6c5-q8djs" Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.530263 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6fd7956ddf-b647f"] Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.537831 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-6fd7956ddf-b647f"] Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.555742 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/50228011-f908-465e-a754-d4b6b626dda9-scripts\") pod \"50228011-f908-465e-a754-d4b6b626dda9\" (UID: \"50228011-f908-465e-a754-d4b6b626dda9\") " Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.556116 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/50228011-f908-465e-a754-d4b6b626dda9-horizon-secret-key\") pod \"50228011-f908-465e-a754-d4b6b626dda9\" (UID: \"50228011-f908-465e-a754-d4b6b626dda9\") " Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.556224 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b8mr2\" (UniqueName: \"kubernetes.io/projected/50228011-f908-465e-a754-d4b6b626dda9-kube-api-access-b8mr2\") pod \"50228011-f908-465e-a754-d4b6b626dda9\" (UID: \"50228011-f908-465e-a754-d4b6b626dda9\") " Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.556287 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50228011-f908-465e-a754-d4b6b626dda9-logs\") pod \"50228011-f908-465e-a754-d4b6b626dda9\" (UID: \"50228011-f908-465e-a754-d4b6b626dda9\") " Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.556322 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/50228011-f908-465e-a754-d4b6b626dda9-config-data\") pod \"50228011-f908-465e-a754-d4b6b626dda9\" (UID: \"50228011-f908-465e-a754-d4b6b626dda9\") " Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.565311 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/50228011-f908-465e-a754-d4b6b626dda9-logs" (OuterVolumeSpecName: "logs") pod "50228011-f908-465e-a754-d4b6b626dda9" (UID: "50228011-f908-465e-a754-d4b6b626dda9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.567231 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50228011-f908-465e-a754-d4b6b626dda9-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "50228011-f908-465e-a754-d4b6b626dda9" (UID: "50228011-f908-465e-a754-d4b6b626dda9"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.571099 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50228011-f908-465e-a754-d4b6b626dda9-kube-api-access-b8mr2" (OuterVolumeSpecName: "kube-api-access-b8mr2") pod "50228011-f908-465e-a754-d4b6b626dda9" (UID: "50228011-f908-465e-a754-d4b6b626dda9"). InnerVolumeSpecName "kube-api-access-b8mr2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.578771 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/50228011-f908-465e-a754-d4b6b626dda9-config-data" (OuterVolumeSpecName: "config-data") pod "50228011-f908-465e-a754-d4b6b626dda9" (UID: "50228011-f908-465e-a754-d4b6b626dda9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.589241 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/50228011-f908-465e-a754-d4b6b626dda9-scripts" (OuterVolumeSpecName: "scripts") pod "50228011-f908-465e-a754-d4b6b626dda9" (UID: "50228011-f908-465e-a754-d4b6b626dda9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.658038 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b8mr2\" (UniqueName: \"kubernetes.io/projected/50228011-f908-465e-a754-d4b6b626dda9-kube-api-access-b8mr2\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.658064 4556 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50228011-f908-465e-a754-d4b6b626dda9-logs\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.658075 4556 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/50228011-f908-465e-a754-d4b6b626dda9-config-data\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.658083 4556 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/50228011-f908-465e-a754-d4b6b626dda9-scripts\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.658091 4556 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/50228011-f908-465e-a754-d4b6b626dda9-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.670220 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-77cf5996cc-dj8s4" Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.679898 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-79dd98bf77-fdbv7" Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.758962 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2plk8\" (UniqueName: \"kubernetes.io/projected/66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98-kube-api-access-2plk8\") pod \"66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98\" (UID: \"66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98\") " Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.759005 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98-logs\") pod \"66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98\" (UID: \"66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98\") " Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.759095 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98-scripts\") pod \"66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98\" (UID: \"66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98\") " Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.759127 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c5fec43c-7b17-45d5-91c4-609b6d2ca1b9-horizon-secret-key\") pod \"c5fec43c-7b17-45d5-91c4-609b6d2ca1b9\" (UID: \"c5fec43c-7b17-45d5-91c4-609b6d2ca1b9\") " Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.759174 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98-horizon-secret-key\") pod \"66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98\" (UID: \"66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98\") " Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.759194 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c5fec43c-7b17-45d5-91c4-609b6d2ca1b9-config-data\") pod \"c5fec43c-7b17-45d5-91c4-609b6d2ca1b9\" (UID: \"c5fec43c-7b17-45d5-91c4-609b6d2ca1b9\") " Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.759314 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c5fec43c-7b17-45d5-91c4-609b6d2ca1b9-scripts\") pod \"c5fec43c-7b17-45d5-91c4-609b6d2ca1b9\" (UID: \"c5fec43c-7b17-45d5-91c4-609b6d2ca1b9\") " Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.759366 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5fec43c-7b17-45d5-91c4-609b6d2ca1b9-logs\") pod \"c5fec43c-7b17-45d5-91c4-609b6d2ca1b9\" (UID: \"c5fec43c-7b17-45d5-91c4-609b6d2ca1b9\") " Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.759449 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tpm2\" (UniqueName: \"kubernetes.io/projected/c5fec43c-7b17-45d5-91c4-609b6d2ca1b9-kube-api-access-8tpm2\") pod \"c5fec43c-7b17-45d5-91c4-609b6d2ca1b9\" (UID: \"c5fec43c-7b17-45d5-91c4-609b6d2ca1b9\") " Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.759479 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98-config-data\") pod \"66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98\" (UID: \"66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98\") " Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.760295 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98-logs" (OuterVolumeSpecName: "logs") pod "66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98" (UID: "66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.760689 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5fec43c-7b17-45d5-91c4-609b6d2ca1b9-logs" (OuterVolumeSpecName: "logs") pod "c5fec43c-7b17-45d5-91c4-609b6d2ca1b9" (UID: "c5fec43c-7b17-45d5-91c4-609b6d2ca1b9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.763609 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98-kube-api-access-2plk8" (OuterVolumeSpecName: "kube-api-access-2plk8") pod "66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98" (UID: "66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98"). InnerVolumeSpecName "kube-api-access-2plk8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.763816 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5fec43c-7b17-45d5-91c4-609b6d2ca1b9-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "c5fec43c-7b17-45d5-91c4-609b6d2ca1b9" (UID: "c5fec43c-7b17-45d5-91c4-609b6d2ca1b9"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.768517 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98" (UID: "66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.769737 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5fec43c-7b17-45d5-91c4-609b6d2ca1b9-kube-api-access-8tpm2" (OuterVolumeSpecName: "kube-api-access-8tpm2") pod "c5fec43c-7b17-45d5-91c4-609b6d2ca1b9" (UID: "c5fec43c-7b17-45d5-91c4-609b6d2ca1b9"). InnerVolumeSpecName "kube-api-access-8tpm2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.779342 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98-scripts" (OuterVolumeSpecName: "scripts") pod "66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98" (UID: "66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.780707 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98-config-data" (OuterVolumeSpecName: "config-data") pod "66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98" (UID: "66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.780733 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5fec43c-7b17-45d5-91c4-609b6d2ca1b9-config-data" (OuterVolumeSpecName: "config-data") pod "c5fec43c-7b17-45d5-91c4-609b6d2ca1b9" (UID: "c5fec43c-7b17-45d5-91c4-609b6d2ca1b9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.793738 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5fec43c-7b17-45d5-91c4-609b6d2ca1b9-scripts" (OuterVolumeSpecName: "scripts") pod "c5fec43c-7b17-45d5-91c4-609b6d2ca1b9" (UID: "c5fec43c-7b17-45d5-91c4-609b6d2ca1b9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.863005 4556 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c5fec43c-7b17-45d5-91c4-609b6d2ca1b9-scripts\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.863042 4556 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5fec43c-7b17-45d5-91c4-609b6d2ca1b9-logs\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.863052 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tpm2\" (UniqueName: \"kubernetes.io/projected/c5fec43c-7b17-45d5-91c4-609b6d2ca1b9-kube-api-access-8tpm2\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.863065 4556 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98-config-data\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.863078 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2plk8\" (UniqueName: \"kubernetes.io/projected/66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98-kube-api-access-2plk8\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.863086 4556 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98-logs\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.863094 4556 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98-scripts\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.863102 4556 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c5fec43c-7b17-45d5-91c4-609b6d2ca1b9-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.863112 4556 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.863120 4556 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c5fec43c-7b17-45d5-91c4-609b6d2ca1b9-config-data\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:40 crc kubenswrapper[4556]: I0218 09:18:40.944958 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-845cb5d46d-z5vhw" Feb 18 09:18:41 crc kubenswrapper[4556]: I0218 09:18:41.183535 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-77cf5996cc-dj8s4" event={"ID":"66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98","Type":"ContainerDied","Data":"0f49d8a03148f93f5cf7a0e584b50845c6d3288e42f4a093af888c34aea04a84"} Feb 18 09:18:41 crc kubenswrapper[4556]: I0218 09:18:41.188239 4556 scope.go:117] "RemoveContainer" containerID="2973401c6ebe4b90a1471e53a5013aae7d75364f2830c2e5cd584b5f74812ad2" Feb 18 09:18:41 crc kubenswrapper[4556]: I0218 09:18:41.188441 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-77cf5996cc-dj8s4" Feb 18 09:18:41 crc kubenswrapper[4556]: I0218 09:18:41.201315 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79dd98bf77-fdbv7" event={"ID":"c5fec43c-7b17-45d5-91c4-609b6d2ca1b9","Type":"ContainerDied","Data":"6db6e7d4ac741e11b00a64d7861d1d97bfb5f5b1197764f1bac8b0fb6ff4df54"} Feb 18 09:18:41 crc kubenswrapper[4556]: I0218 09:18:41.201328 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-79dd98bf77-fdbv7" Feb 18 09:18:41 crc kubenswrapper[4556]: I0218 09:18:41.209495 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-74c5b6b6c5-q8djs" event={"ID":"50228011-f908-465e-a754-d4b6b626dda9","Type":"ContainerDied","Data":"d7824966b854b4e05a496fcf7987339d08df598f934fcdf6099cd480c21cd503"} Feb 18 09:18:41 crc kubenswrapper[4556]: I0218 09:18:41.209566 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-74c5b6b6c5-q8djs" Feb 18 09:18:41 crc kubenswrapper[4556]: I0218 09:18:41.231028 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-77cf5996cc-dj8s4"] Feb 18 09:18:41 crc kubenswrapper[4556]: I0218 09:18:41.240392 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-77cf5996cc-dj8s4"] Feb 18 09:18:41 crc kubenswrapper[4556]: I0218 09:18:41.264219 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-79dd98bf77-fdbv7"] Feb 18 09:18:41 crc kubenswrapper[4556]: I0218 09:18:41.276797 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-79dd98bf77-fdbv7"] Feb 18 09:18:41 crc kubenswrapper[4556]: I0218 09:18:41.298075 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98" path="/var/lib/kubelet/pods/66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98/volumes" Feb 18 09:18:41 crc kubenswrapper[4556]: I0218 09:18:41.298764 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b92f250f-e888-4ff5-ae26-9e31c7be3379" path="/var/lib/kubelet/pods/b92f250f-e888-4ff5-ae26-9e31c7be3379/volumes" Feb 18 09:18:41 crc kubenswrapper[4556]: I0218 09:18:41.299455 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5fec43c-7b17-45d5-91c4-609b6d2ca1b9" path="/var/lib/kubelet/pods/c5fec43c-7b17-45d5-91c4-609b6d2ca1b9/volumes" Feb 18 09:18:41 crc kubenswrapper[4556]: I0218 09:18:41.300813 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-74c5b6b6c5-q8djs"] Feb 18 09:18:41 crc kubenswrapper[4556]: I0218 09:18:41.300858 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-74c5b6b6c5-q8djs"] Feb 18 09:18:41 crc kubenswrapper[4556]: I0218 09:18:41.347061 4556 scope.go:117] "RemoveContainer" containerID="ace36a61edb3ea0808af8dd243149192bdb820e0dff28b907b17b501b9e6c790" Feb 18 09:18:41 crc kubenswrapper[4556]: I0218 09:18:41.373049 4556 scope.go:117] "RemoveContainer" containerID="a6e43edd5803076f4c76dc543e9ef0bdba7eda69af36b2cf7df94a0200eacf09" Feb 18 09:18:41 crc kubenswrapper[4556]: I0218 09:18:41.529228 4556 scope.go:117] "RemoveContainer" containerID="2fed1042675c8bc639ff1bf40020b12a7f5f553be9bb216173f3fbe9c18aa7dd" Feb 18 09:18:41 crc kubenswrapper[4556]: I0218 09:18:41.553111 4556 scope.go:117] "RemoveContainer" containerID="c354f6bc02d120ffecda388ac1244fdc7228debcaed10748429b9072f8e42df2" Feb 18 09:18:41 crc kubenswrapper[4556]: I0218 09:18:41.687307 4556 scope.go:117] "RemoveContainer" containerID="7147437f196486d78c3232e03d579055aba6c510b6715c9264fd4d2b04e9089e" Feb 18 09:18:42 crc kubenswrapper[4556]: I0218 09:18:42.179511 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-845cb5d46d-z5vhw" Feb 18 09:18:42 crc kubenswrapper[4556]: I0218 09:18:42.229986 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-c6ccddcd4-hbvfc"] Feb 18 09:18:42 crc kubenswrapper[4556]: I0218 09:18:42.230173 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-c6ccddcd4-hbvfc" podUID="18385928-f2f7-4f58-9f83-e563af72c8b4" containerName="barbican-api-log" containerID="cri-o://2ce3f2b0b5da033ec3e8d4a3882f0e1ddc3ecafa75e1c88a35839c6ddc867cd4" gracePeriod=30 Feb 18 09:18:42 crc kubenswrapper[4556]: I0218 09:18:42.230448 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-c6ccddcd4-hbvfc" podUID="18385928-f2f7-4f58-9f83-e563af72c8b4" containerName="barbican-api" containerID="cri-o://c50d1bd7e8cebf034249d811b5a1c2d743431d8dd4f4c67266c76f71686f721e" gracePeriod=30 Feb 18 09:18:42 crc kubenswrapper[4556]: I0218 09:18:42.245258 4556 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-c6ccddcd4-hbvfc" podUID="18385928-f2f7-4f58-9f83-e563af72c8b4" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": EOF" Feb 18 09:18:42 crc kubenswrapper[4556]: I0218 09:18:42.245300 4556 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-c6ccddcd4-hbvfc" podUID="18385928-f2f7-4f58-9f83-e563af72c8b4" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": EOF" Feb 18 09:18:42 crc kubenswrapper[4556]: I0218 09:18:42.255335 4556 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-c6ccddcd4-hbvfc" podUID="18385928-f2f7-4f58-9f83-e563af72c8b4" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": EOF" Feb 18 09:18:42 crc kubenswrapper[4556]: I0218 09:18:42.255361 4556 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-c6ccddcd4-hbvfc" podUID="18385928-f2f7-4f58-9f83-e563af72c8b4" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": EOF" Feb 18 09:18:43 crc kubenswrapper[4556]: I0218 09:18:43.284204 4556 generic.go:334] "Generic (PLEG): container finished" podID="18385928-f2f7-4f58-9f83-e563af72c8b4" containerID="2ce3f2b0b5da033ec3e8d4a3882f0e1ddc3ecafa75e1c88a35839c6ddc867cd4" exitCode=143 Feb 18 09:18:43 crc kubenswrapper[4556]: I0218 09:18:43.287885 4556 generic.go:334] "Generic (PLEG): container finished" podID="3354b930-6b00-4aa7-a4bd-97ec410cc863" containerID="4423fda867205940e352d08e4c633d2caee4bc72b78e6280dd5fd52606caa4dd" exitCode=0 Feb 18 09:18:43 crc kubenswrapper[4556]: I0218 09:18:43.292321 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50228011-f908-465e-a754-d4b6b626dda9" path="/var/lib/kubelet/pods/50228011-f908-465e-a754-d4b6b626dda9/volumes" Feb 18 09:18:43 crc kubenswrapper[4556]: I0218 09:18:43.296021 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-c6ccddcd4-hbvfc" event={"ID":"18385928-f2f7-4f58-9f83-e563af72c8b4","Type":"ContainerDied","Data":"2ce3f2b0b5da033ec3e8d4a3882f0e1ddc3ecafa75e1c88a35839c6ddc867cd4"} Feb 18 09:18:43 crc kubenswrapper[4556]: I0218 09:18:43.296444 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-85bb8c66f4-fhblh" event={"ID":"3354b930-6b00-4aa7-a4bd-97ec410cc863","Type":"ContainerDied","Data":"4423fda867205940e352d08e4c633d2caee4bc72b78e6280dd5fd52606caa4dd"} Feb 18 09:18:43 crc kubenswrapper[4556]: I0218 09:18:43.695421 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Feb 18 09:18:43 crc kubenswrapper[4556]: I0218 09:18:43.722307 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-849fd69845-bdbc8" Feb 18 09:18:43 crc kubenswrapper[4556]: I0218 09:18:43.725278 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 18 09:18:43 crc kubenswrapper[4556]: I0218 09:18:43.779209 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86dd6887f5-kqgml"] Feb 18 09:18:43 crc kubenswrapper[4556]: I0218 09:18:43.779685 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86dd6887f5-kqgml" podUID="02b5c258-0253-4d75-852b-30c2b26b5598" containerName="dnsmasq-dns" containerID="cri-o://0da42c5457683b62830f1b6e7cd6f3910bea6f7d2bc6851bc77ed7aa339d0f0d" gracePeriod=10 Feb 18 09:18:44 crc kubenswrapper[4556]: I0218 09:18:44.070518 4556 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-85bb8c66f4-fhblh" podUID="3354b930-6b00-4aa7-a4bd-97ec410cc863" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Feb 18 09:18:44 crc kubenswrapper[4556]: I0218 09:18:44.229533 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86dd6887f5-kqgml" Feb 18 09:18:44 crc kubenswrapper[4556]: I0218 09:18:44.304288 4556 generic.go:334] "Generic (PLEG): container finished" podID="02b5c258-0253-4d75-852b-30c2b26b5598" containerID="0da42c5457683b62830f1b6e7cd6f3910bea6f7d2bc6851bc77ed7aa339d0f0d" exitCode=0 Feb 18 09:18:44 crc kubenswrapper[4556]: I0218 09:18:44.304369 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86dd6887f5-kqgml" event={"ID":"02b5c258-0253-4d75-852b-30c2b26b5598","Type":"ContainerDied","Data":"0da42c5457683b62830f1b6e7cd6f3910bea6f7d2bc6851bc77ed7aa339d0f0d"} Feb 18 09:18:44 crc kubenswrapper[4556]: I0218 09:18:44.304418 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86dd6887f5-kqgml" event={"ID":"02b5c258-0253-4d75-852b-30c2b26b5598","Type":"ContainerDied","Data":"b642158754e733cbe4e6d6ca65904149495445382adadd8f335ec5f528572058"} Feb 18 09:18:44 crc kubenswrapper[4556]: I0218 09:18:44.304415 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86dd6887f5-kqgml" Feb 18 09:18:44 crc kubenswrapper[4556]: I0218 09:18:44.304445 4556 scope.go:117] "RemoveContainer" containerID="0da42c5457683b62830f1b6e7cd6f3910bea6f7d2bc6851bc77ed7aa339d0f0d" Feb 18 09:18:44 crc kubenswrapper[4556]: I0218 09:18:44.304606 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="59767f41-27b9-4146-b723-ea8df508a92e" containerName="cinder-scheduler" containerID="cri-o://ed41c2e1779bc2ff90abc289c7c836a56a67c4f7e881c290aa073046688b2f99" gracePeriod=30 Feb 18 09:18:44 crc kubenswrapper[4556]: I0218 09:18:44.304643 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="59767f41-27b9-4146-b723-ea8df508a92e" containerName="probe" containerID="cri-o://5bcd75d3aae13e3e94b1fa31eb135277d4a4343b64ada8a1ab708f2ff72fad1e" gracePeriod=30 Feb 18 09:18:44 crc kubenswrapper[4556]: I0218 09:18:44.327067 4556 scope.go:117] "RemoveContainer" containerID="1888148871a332a7d249fe69193657e3adb1614a874e247e986fe92585d9afa7" Feb 18 09:18:44 crc kubenswrapper[4556]: I0218 09:18:44.345048 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02b5c258-0253-4d75-852b-30c2b26b5598-config\") pod \"02b5c258-0253-4d75-852b-30c2b26b5598\" (UID: \"02b5c258-0253-4d75-852b-30c2b26b5598\") " Feb 18 09:18:44 crc kubenswrapper[4556]: I0218 09:18:44.345087 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/02b5c258-0253-4d75-852b-30c2b26b5598-dns-svc\") pod \"02b5c258-0253-4d75-852b-30c2b26b5598\" (UID: \"02b5c258-0253-4d75-852b-30c2b26b5598\") " Feb 18 09:18:44 crc kubenswrapper[4556]: I0218 09:18:44.345226 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/02b5c258-0253-4d75-852b-30c2b26b5598-ovsdbserver-sb\") pod \"02b5c258-0253-4d75-852b-30c2b26b5598\" (UID: \"02b5c258-0253-4d75-852b-30c2b26b5598\") " Feb 18 09:18:44 crc kubenswrapper[4556]: I0218 09:18:44.345362 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/02b5c258-0253-4d75-852b-30c2b26b5598-dns-swift-storage-0\") pod \"02b5c258-0253-4d75-852b-30c2b26b5598\" (UID: \"02b5c258-0253-4d75-852b-30c2b26b5598\") " Feb 18 09:18:44 crc kubenswrapper[4556]: I0218 09:18:44.345390 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c4hph\" (UniqueName: \"kubernetes.io/projected/02b5c258-0253-4d75-852b-30c2b26b5598-kube-api-access-c4hph\") pod \"02b5c258-0253-4d75-852b-30c2b26b5598\" (UID: \"02b5c258-0253-4d75-852b-30c2b26b5598\") " Feb 18 09:18:44 crc kubenswrapper[4556]: I0218 09:18:44.345414 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/02b5c258-0253-4d75-852b-30c2b26b5598-ovsdbserver-nb\") pod \"02b5c258-0253-4d75-852b-30c2b26b5598\" (UID: \"02b5c258-0253-4d75-852b-30c2b26b5598\") " Feb 18 09:18:44 crc kubenswrapper[4556]: I0218 09:18:44.355452 4556 scope.go:117] "RemoveContainer" containerID="0da42c5457683b62830f1b6e7cd6f3910bea6f7d2bc6851bc77ed7aa339d0f0d" Feb 18 09:18:44 crc kubenswrapper[4556]: E0218 09:18:44.360850 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0da42c5457683b62830f1b6e7cd6f3910bea6f7d2bc6851bc77ed7aa339d0f0d\": container with ID starting with 0da42c5457683b62830f1b6e7cd6f3910bea6f7d2bc6851bc77ed7aa339d0f0d not found: ID does not exist" containerID="0da42c5457683b62830f1b6e7cd6f3910bea6f7d2bc6851bc77ed7aa339d0f0d" Feb 18 09:18:44 crc kubenswrapper[4556]: I0218 09:18:44.360915 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0da42c5457683b62830f1b6e7cd6f3910bea6f7d2bc6851bc77ed7aa339d0f0d"} err="failed to get container status \"0da42c5457683b62830f1b6e7cd6f3910bea6f7d2bc6851bc77ed7aa339d0f0d\": rpc error: code = NotFound desc = could not find container \"0da42c5457683b62830f1b6e7cd6f3910bea6f7d2bc6851bc77ed7aa339d0f0d\": container with ID starting with 0da42c5457683b62830f1b6e7cd6f3910bea6f7d2bc6851bc77ed7aa339d0f0d not found: ID does not exist" Feb 18 09:18:44 crc kubenswrapper[4556]: I0218 09:18:44.360949 4556 scope.go:117] "RemoveContainer" containerID="1888148871a332a7d249fe69193657e3adb1614a874e247e986fe92585d9afa7" Feb 18 09:18:44 crc kubenswrapper[4556]: I0218 09:18:44.375952 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02b5c258-0253-4d75-852b-30c2b26b5598-kube-api-access-c4hph" (OuterVolumeSpecName: "kube-api-access-c4hph") pod "02b5c258-0253-4d75-852b-30c2b26b5598" (UID: "02b5c258-0253-4d75-852b-30c2b26b5598"). InnerVolumeSpecName "kube-api-access-c4hph". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:18:44 crc kubenswrapper[4556]: E0218 09:18:44.393946 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1888148871a332a7d249fe69193657e3adb1614a874e247e986fe92585d9afa7\": container with ID starting with 1888148871a332a7d249fe69193657e3adb1614a874e247e986fe92585d9afa7 not found: ID does not exist" containerID="1888148871a332a7d249fe69193657e3adb1614a874e247e986fe92585d9afa7" Feb 18 09:18:44 crc kubenswrapper[4556]: I0218 09:18:44.393991 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1888148871a332a7d249fe69193657e3adb1614a874e247e986fe92585d9afa7"} err="failed to get container status \"1888148871a332a7d249fe69193657e3adb1614a874e247e986fe92585d9afa7\": rpc error: code = NotFound desc = could not find container \"1888148871a332a7d249fe69193657e3adb1614a874e247e986fe92585d9afa7\": container with ID starting with 1888148871a332a7d249fe69193657e3adb1614a874e247e986fe92585d9afa7 not found: ID does not exist" Feb 18 09:18:44 crc kubenswrapper[4556]: I0218 09:18:44.426495 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02b5c258-0253-4d75-852b-30c2b26b5598-config" (OuterVolumeSpecName: "config") pod "02b5c258-0253-4d75-852b-30c2b26b5598" (UID: "02b5c258-0253-4d75-852b-30c2b26b5598"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:18:44 crc kubenswrapper[4556]: I0218 09:18:44.426686 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02b5c258-0253-4d75-852b-30c2b26b5598-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "02b5c258-0253-4d75-852b-30c2b26b5598" (UID: "02b5c258-0253-4d75-852b-30c2b26b5598"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:18:44 crc kubenswrapper[4556]: I0218 09:18:44.438987 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02b5c258-0253-4d75-852b-30c2b26b5598-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "02b5c258-0253-4d75-852b-30c2b26b5598" (UID: "02b5c258-0253-4d75-852b-30c2b26b5598"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:18:44 crc kubenswrapper[4556]: I0218 09:18:44.441115 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02b5c258-0253-4d75-852b-30c2b26b5598-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "02b5c258-0253-4d75-852b-30c2b26b5598" (UID: "02b5c258-0253-4d75-852b-30c2b26b5598"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:18:44 crc kubenswrapper[4556]: I0218 09:18:44.448442 4556 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02b5c258-0253-4d75-852b-30c2b26b5598-config\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:44 crc kubenswrapper[4556]: I0218 09:18:44.448470 4556 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/02b5c258-0253-4d75-852b-30c2b26b5598-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:44 crc kubenswrapper[4556]: I0218 09:18:44.448480 4556 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/02b5c258-0253-4d75-852b-30c2b26b5598-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:44 crc kubenswrapper[4556]: I0218 09:18:44.448491 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c4hph\" (UniqueName: \"kubernetes.io/projected/02b5c258-0253-4d75-852b-30c2b26b5598-kube-api-access-c4hph\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:44 crc kubenswrapper[4556]: I0218 09:18:44.448500 4556 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/02b5c258-0253-4d75-852b-30c2b26b5598-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:44 crc kubenswrapper[4556]: I0218 09:18:44.472541 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02b5c258-0253-4d75-852b-30c2b26b5598-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "02b5c258-0253-4d75-852b-30c2b26b5598" (UID: "02b5c258-0253-4d75-852b-30c2b26b5598"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:18:44 crc kubenswrapper[4556]: I0218 09:18:44.550034 4556 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/02b5c258-0253-4d75-852b-30c2b26b5598-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:44 crc kubenswrapper[4556]: I0218 09:18:44.632056 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86dd6887f5-kqgml"] Feb 18 09:18:44 crc kubenswrapper[4556]: I0218 09:18:44.640010 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86dd6887f5-kqgml"] Feb 18 09:18:44 crc kubenswrapper[4556]: I0218 09:18:44.730853 4556 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="e0739f72-891b-46e1-bce5-e17c54a407ab" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Feb 18 09:18:45 crc kubenswrapper[4556]: I0218 09:18:45.296686 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02b5c258-0253-4d75-852b-30c2b26b5598" path="/var/lib/kubelet/pods/02b5c258-0253-4d75-852b-30c2b26b5598/volumes" Feb 18 09:18:45 crc kubenswrapper[4556]: I0218 09:18:45.315730 4556 generic.go:334] "Generic (PLEG): container finished" podID="59767f41-27b9-4146-b723-ea8df508a92e" containerID="5bcd75d3aae13e3e94b1fa31eb135277d4a4343b64ada8a1ab708f2ff72fad1e" exitCode=0 Feb 18 09:18:45 crc kubenswrapper[4556]: I0218 09:18:45.315766 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"59767f41-27b9-4146-b723-ea8df508a92e","Type":"ContainerDied","Data":"5bcd75d3aae13e3e94b1fa31eb135277d4a4343b64ada8a1ab708f2ff72fad1e"} Feb 18 09:18:45 crc kubenswrapper[4556]: I0218 09:18:45.432211 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-x6wtm" Feb 18 09:18:45 crc kubenswrapper[4556]: I0218 09:18:45.472375 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-x6wtm" Feb 18 09:18:45 crc kubenswrapper[4556]: I0218 09:18:45.666326 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-x6wtm"] Feb 18 09:18:46 crc kubenswrapper[4556]: I0218 09:18:46.045387 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Feb 18 09:18:46 crc kubenswrapper[4556]: I0218 09:18:46.669686 4556 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-c6ccddcd4-hbvfc" podUID="18385928-f2f7-4f58-9f83-e563af72c8b4" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": read tcp 10.217.0.2:52334->10.217.0.160:9311: read: connection reset by peer" Feb 18 09:18:46 crc kubenswrapper[4556]: I0218 09:18:46.670032 4556 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-c6ccddcd4-hbvfc" podUID="18385928-f2f7-4f58-9f83-e563af72c8b4" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": dial tcp 10.217.0.160:9311: connect: connection refused" Feb 18 09:18:46 crc kubenswrapper[4556]: I0218 09:18:46.670093 4556 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-c6ccddcd4-hbvfc" podUID="18385928-f2f7-4f58-9f83-e563af72c8b4" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": read tcp 10.217.0.2:56414->10.217.0.160:9311: read: connection reset by peer" Feb 18 09:18:47 crc kubenswrapper[4556]: I0218 09:18:47.087530 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-c6ccddcd4-hbvfc" Feb 18 09:18:47 crc kubenswrapper[4556]: I0218 09:18:47.157225 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-66f6978747-npdn5" Feb 18 09:18:47 crc kubenswrapper[4556]: I0218 09:18:47.211368 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18385928-f2f7-4f58-9f83-e563af72c8b4-config-data\") pod \"18385928-f2f7-4f58-9f83-e563af72c8b4\" (UID: \"18385928-f2f7-4f58-9f83-e563af72c8b4\") " Feb 18 09:18:47 crc kubenswrapper[4556]: I0218 09:18:47.211509 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/18385928-f2f7-4f58-9f83-e563af72c8b4-config-data-custom\") pod \"18385928-f2f7-4f58-9f83-e563af72c8b4\" (UID: \"18385928-f2f7-4f58-9f83-e563af72c8b4\") " Feb 18 09:18:47 crc kubenswrapper[4556]: I0218 09:18:47.211561 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rc7nn\" (UniqueName: \"kubernetes.io/projected/18385928-f2f7-4f58-9f83-e563af72c8b4-kube-api-access-rc7nn\") pod \"18385928-f2f7-4f58-9f83-e563af72c8b4\" (UID: \"18385928-f2f7-4f58-9f83-e563af72c8b4\") " Feb 18 09:18:47 crc kubenswrapper[4556]: I0218 09:18:47.211734 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18385928-f2f7-4f58-9f83-e563af72c8b4-combined-ca-bundle\") pod \"18385928-f2f7-4f58-9f83-e563af72c8b4\" (UID: \"18385928-f2f7-4f58-9f83-e563af72c8b4\") " Feb 18 09:18:47 crc kubenswrapper[4556]: I0218 09:18:47.211770 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/18385928-f2f7-4f58-9f83-e563af72c8b4-logs\") pod \"18385928-f2f7-4f58-9f83-e563af72c8b4\" (UID: \"18385928-f2f7-4f58-9f83-e563af72c8b4\") " Feb 18 09:18:47 crc kubenswrapper[4556]: I0218 09:18:47.212812 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18385928-f2f7-4f58-9f83-e563af72c8b4-logs" (OuterVolumeSpecName: "logs") pod "18385928-f2f7-4f58-9f83-e563af72c8b4" (UID: "18385928-f2f7-4f58-9f83-e563af72c8b4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:18:47 crc kubenswrapper[4556]: I0218 09:18:47.217284 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18385928-f2f7-4f58-9f83-e563af72c8b4-kube-api-access-rc7nn" (OuterVolumeSpecName: "kube-api-access-rc7nn") pod "18385928-f2f7-4f58-9f83-e563af72c8b4" (UID: "18385928-f2f7-4f58-9f83-e563af72c8b4"). InnerVolumeSpecName "kube-api-access-rc7nn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:18:47 crc kubenswrapper[4556]: I0218 09:18:47.225086 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18385928-f2f7-4f58-9f83-e563af72c8b4-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "18385928-f2f7-4f58-9f83-e563af72c8b4" (UID: "18385928-f2f7-4f58-9f83-e563af72c8b4"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:18:47 crc kubenswrapper[4556]: I0218 09:18:47.237398 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18385928-f2f7-4f58-9f83-e563af72c8b4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "18385928-f2f7-4f58-9f83-e563af72c8b4" (UID: "18385928-f2f7-4f58-9f83-e563af72c8b4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:18:47 crc kubenswrapper[4556]: I0218 09:18:47.265499 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18385928-f2f7-4f58-9f83-e563af72c8b4-config-data" (OuterVolumeSpecName: "config-data") pod "18385928-f2f7-4f58-9f83-e563af72c8b4" (UID: "18385928-f2f7-4f58-9f83-e563af72c8b4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:18:47 crc kubenswrapper[4556]: I0218 09:18:47.315646 4556 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/18385928-f2f7-4f58-9f83-e563af72c8b4-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:47 crc kubenswrapper[4556]: I0218 09:18:47.315680 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rc7nn\" (UniqueName: \"kubernetes.io/projected/18385928-f2f7-4f58-9f83-e563af72c8b4-kube-api-access-rc7nn\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:47 crc kubenswrapper[4556]: I0218 09:18:47.315693 4556 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18385928-f2f7-4f58-9f83-e563af72c8b4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:47 crc kubenswrapper[4556]: I0218 09:18:47.315706 4556 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/18385928-f2f7-4f58-9f83-e563af72c8b4-logs\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:47 crc kubenswrapper[4556]: I0218 09:18:47.315715 4556 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18385928-f2f7-4f58-9f83-e563af72c8b4-config-data\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:47 crc kubenswrapper[4556]: I0218 09:18:47.337069 4556 generic.go:334] "Generic (PLEG): container finished" podID="18385928-f2f7-4f58-9f83-e563af72c8b4" containerID="c50d1bd7e8cebf034249d811b5a1c2d743431d8dd4f4c67266c76f71686f721e" exitCode=0 Feb 18 09:18:47 crc kubenswrapper[4556]: I0218 09:18:47.337133 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-c6ccddcd4-hbvfc" event={"ID":"18385928-f2f7-4f58-9f83-e563af72c8b4","Type":"ContainerDied","Data":"c50d1bd7e8cebf034249d811b5a1c2d743431d8dd4f4c67266c76f71686f721e"} Feb 18 09:18:47 crc kubenswrapper[4556]: I0218 09:18:47.337207 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-c6ccddcd4-hbvfc" Feb 18 09:18:47 crc kubenswrapper[4556]: I0218 09:18:47.337244 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-c6ccddcd4-hbvfc" event={"ID":"18385928-f2f7-4f58-9f83-e563af72c8b4","Type":"ContainerDied","Data":"4adbdfc316b6a3c5fad94306cb4338ce4ccf3ce88ca906c78af38d1d503be5a6"} Feb 18 09:18:47 crc kubenswrapper[4556]: I0218 09:18:47.337280 4556 scope.go:117] "RemoveContainer" containerID="c50d1bd7e8cebf034249d811b5a1c2d743431d8dd4f4c67266c76f71686f721e" Feb 18 09:18:47 crc kubenswrapper[4556]: I0218 09:18:47.337350 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-x6wtm" podUID="1fb56420-2137-43d2-b790-1773daf3f0c2" containerName="registry-server" containerID="cri-o://83cf21b516d05fcd6e7f07cf698f60fce5b1a1ce77935ae3c4b37e86e37158bc" gracePeriod=2 Feb 18 09:18:47 crc kubenswrapper[4556]: I0218 09:18:47.355583 4556 scope.go:117] "RemoveContainer" containerID="2ce3f2b0b5da033ec3e8d4a3882f0e1ddc3ecafa75e1c88a35839c6ddc867cd4" Feb 18 09:18:47 crc kubenswrapper[4556]: I0218 09:18:47.367987 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-c6ccddcd4-hbvfc"] Feb 18 09:18:47 crc kubenswrapper[4556]: I0218 09:18:47.374181 4556 scope.go:117] "RemoveContainer" containerID="c50d1bd7e8cebf034249d811b5a1c2d743431d8dd4f4c67266c76f71686f721e" Feb 18 09:18:47 crc kubenswrapper[4556]: E0218 09:18:47.374623 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c50d1bd7e8cebf034249d811b5a1c2d743431d8dd4f4c67266c76f71686f721e\": container with ID starting with c50d1bd7e8cebf034249d811b5a1c2d743431d8dd4f4c67266c76f71686f721e not found: ID does not exist" containerID="c50d1bd7e8cebf034249d811b5a1c2d743431d8dd4f4c67266c76f71686f721e" Feb 18 09:18:47 crc kubenswrapper[4556]: I0218 09:18:47.374667 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c50d1bd7e8cebf034249d811b5a1c2d743431d8dd4f4c67266c76f71686f721e"} err="failed to get container status \"c50d1bd7e8cebf034249d811b5a1c2d743431d8dd4f4c67266c76f71686f721e\": rpc error: code = NotFound desc = could not find container \"c50d1bd7e8cebf034249d811b5a1c2d743431d8dd4f4c67266c76f71686f721e\": container with ID starting with c50d1bd7e8cebf034249d811b5a1c2d743431d8dd4f4c67266c76f71686f721e not found: ID does not exist" Feb 18 09:18:47 crc kubenswrapper[4556]: I0218 09:18:47.374697 4556 scope.go:117] "RemoveContainer" containerID="2ce3f2b0b5da033ec3e8d4a3882f0e1ddc3ecafa75e1c88a35839c6ddc867cd4" Feb 18 09:18:47 crc kubenswrapper[4556]: I0218 09:18:47.374937 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-c6ccddcd4-hbvfc"] Feb 18 09:18:47 crc kubenswrapper[4556]: E0218 09:18:47.375020 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ce3f2b0b5da033ec3e8d4a3882f0e1ddc3ecafa75e1c88a35839c6ddc867cd4\": container with ID starting with 2ce3f2b0b5da033ec3e8d4a3882f0e1ddc3ecafa75e1c88a35839c6ddc867cd4 not found: ID does not exist" containerID="2ce3f2b0b5da033ec3e8d4a3882f0e1ddc3ecafa75e1c88a35839c6ddc867cd4" Feb 18 09:18:47 crc kubenswrapper[4556]: I0218 09:18:47.375052 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ce3f2b0b5da033ec3e8d4a3882f0e1ddc3ecafa75e1c88a35839c6ddc867cd4"} err="failed to get container status \"2ce3f2b0b5da033ec3e8d4a3882f0e1ddc3ecafa75e1c88a35839c6ddc867cd4\": rpc error: code = NotFound desc = could not find container \"2ce3f2b0b5da033ec3e8d4a3882f0e1ddc3ecafa75e1c88a35839c6ddc867cd4\": container with ID starting with 2ce3f2b0b5da033ec3e8d4a3882f0e1ddc3ecafa75e1c88a35839c6ddc867cd4 not found: ID does not exist" Feb 18 09:18:47 crc kubenswrapper[4556]: I0218 09:18:47.735354 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-x6wtm" Feb 18 09:18:47 crc kubenswrapper[4556]: I0218 09:18:47.834515 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5xtlr\" (UniqueName: \"kubernetes.io/projected/1fb56420-2137-43d2-b790-1773daf3f0c2-kube-api-access-5xtlr\") pod \"1fb56420-2137-43d2-b790-1773daf3f0c2\" (UID: \"1fb56420-2137-43d2-b790-1773daf3f0c2\") " Feb 18 09:18:47 crc kubenswrapper[4556]: I0218 09:18:47.834630 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1fb56420-2137-43d2-b790-1773daf3f0c2-utilities\") pod \"1fb56420-2137-43d2-b790-1773daf3f0c2\" (UID: \"1fb56420-2137-43d2-b790-1773daf3f0c2\") " Feb 18 09:18:47 crc kubenswrapper[4556]: I0218 09:18:47.834739 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1fb56420-2137-43d2-b790-1773daf3f0c2-catalog-content\") pod \"1fb56420-2137-43d2-b790-1773daf3f0c2\" (UID: \"1fb56420-2137-43d2-b790-1773daf3f0c2\") " Feb 18 09:18:47 crc kubenswrapper[4556]: I0218 09:18:47.835280 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1fb56420-2137-43d2-b790-1773daf3f0c2-utilities" (OuterVolumeSpecName: "utilities") pod "1fb56420-2137-43d2-b790-1773daf3f0c2" (UID: "1fb56420-2137-43d2-b790-1773daf3f0c2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:18:47 crc kubenswrapper[4556]: I0218 09:18:47.836019 4556 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1fb56420-2137-43d2-b790-1773daf3f0c2-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:47 crc kubenswrapper[4556]: I0218 09:18:47.840490 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1fb56420-2137-43d2-b790-1773daf3f0c2-kube-api-access-5xtlr" (OuterVolumeSpecName: "kube-api-access-5xtlr") pod "1fb56420-2137-43d2-b790-1773daf3f0c2" (UID: "1fb56420-2137-43d2-b790-1773daf3f0c2"). InnerVolumeSpecName "kube-api-access-5xtlr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:18:47 crc kubenswrapper[4556]: I0218 09:18:47.882581 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1fb56420-2137-43d2-b790-1773daf3f0c2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1fb56420-2137-43d2-b790-1773daf3f0c2" (UID: "1fb56420-2137-43d2-b790-1773daf3f0c2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:18:47 crc kubenswrapper[4556]: I0218 09:18:47.938133 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5xtlr\" (UniqueName: \"kubernetes.io/projected/1fb56420-2137-43d2-b790-1773daf3f0c2-kube-api-access-5xtlr\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:47 crc kubenswrapper[4556]: I0218 09:18:47.938181 4556 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1fb56420-2137-43d2-b790-1773daf3f0c2-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:48 crc kubenswrapper[4556]: I0218 09:18:48.347514 4556 generic.go:334] "Generic (PLEG): container finished" podID="59767f41-27b9-4146-b723-ea8df508a92e" containerID="ed41c2e1779bc2ff90abc289c7c836a56a67c4f7e881c290aa073046688b2f99" exitCode=0 Feb 18 09:18:48 crc kubenswrapper[4556]: I0218 09:18:48.347576 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"59767f41-27b9-4146-b723-ea8df508a92e","Type":"ContainerDied","Data":"ed41c2e1779bc2ff90abc289c7c836a56a67c4f7e881c290aa073046688b2f99"} Feb 18 09:18:48 crc kubenswrapper[4556]: I0218 09:18:48.351038 4556 generic.go:334] "Generic (PLEG): container finished" podID="1fb56420-2137-43d2-b790-1773daf3f0c2" containerID="83cf21b516d05fcd6e7f07cf698f60fce5b1a1ce77935ae3c4b37e86e37158bc" exitCode=0 Feb 18 09:18:48 crc kubenswrapper[4556]: I0218 09:18:48.351081 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x6wtm" event={"ID":"1fb56420-2137-43d2-b790-1773daf3f0c2","Type":"ContainerDied","Data":"83cf21b516d05fcd6e7f07cf698f60fce5b1a1ce77935ae3c4b37e86e37158bc"} Feb 18 09:18:48 crc kubenswrapper[4556]: I0218 09:18:48.351105 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-x6wtm" Feb 18 09:18:48 crc kubenswrapper[4556]: I0218 09:18:48.351111 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x6wtm" event={"ID":"1fb56420-2137-43d2-b790-1773daf3f0c2","Type":"ContainerDied","Data":"55bc5251185b214278150623c3ba439b07097ad9a8fa559ea68fa2a6bed0096a"} Feb 18 09:18:48 crc kubenswrapper[4556]: I0218 09:18:48.351134 4556 scope.go:117] "RemoveContainer" containerID="83cf21b516d05fcd6e7f07cf698f60fce5b1a1ce77935ae3c4b37e86e37158bc" Feb 18 09:18:48 crc kubenswrapper[4556]: I0218 09:18:48.374535 4556 scope.go:117] "RemoveContainer" containerID="f4b9aab0d6c8aa124f35a7d50b97576b617bd928e81f72be3d4072a9ae816a71" Feb 18 09:18:48 crc kubenswrapper[4556]: I0218 09:18:48.383266 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-x6wtm"] Feb 18 09:18:48 crc kubenswrapper[4556]: I0218 09:18:48.395191 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-x6wtm"] Feb 18 09:18:48 crc kubenswrapper[4556]: I0218 09:18:48.410252 4556 scope.go:117] "RemoveContainer" containerID="0ed40eaa96898558bbf3a3f10d8fc68051c8f2daee7f7acaaf04c49222f3285c" Feb 18 09:18:48 crc kubenswrapper[4556]: I0218 09:18:48.438346 4556 scope.go:117] "RemoveContainer" containerID="83cf21b516d05fcd6e7f07cf698f60fce5b1a1ce77935ae3c4b37e86e37158bc" Feb 18 09:18:48 crc kubenswrapper[4556]: E0218 09:18:48.438746 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83cf21b516d05fcd6e7f07cf698f60fce5b1a1ce77935ae3c4b37e86e37158bc\": container with ID starting with 83cf21b516d05fcd6e7f07cf698f60fce5b1a1ce77935ae3c4b37e86e37158bc not found: ID does not exist" containerID="83cf21b516d05fcd6e7f07cf698f60fce5b1a1ce77935ae3c4b37e86e37158bc" Feb 18 09:18:48 crc kubenswrapper[4556]: I0218 09:18:48.438784 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83cf21b516d05fcd6e7f07cf698f60fce5b1a1ce77935ae3c4b37e86e37158bc"} err="failed to get container status \"83cf21b516d05fcd6e7f07cf698f60fce5b1a1ce77935ae3c4b37e86e37158bc\": rpc error: code = NotFound desc = could not find container \"83cf21b516d05fcd6e7f07cf698f60fce5b1a1ce77935ae3c4b37e86e37158bc\": container with ID starting with 83cf21b516d05fcd6e7f07cf698f60fce5b1a1ce77935ae3c4b37e86e37158bc not found: ID does not exist" Feb 18 09:18:48 crc kubenswrapper[4556]: I0218 09:18:48.438809 4556 scope.go:117] "RemoveContainer" containerID="f4b9aab0d6c8aa124f35a7d50b97576b617bd928e81f72be3d4072a9ae816a71" Feb 18 09:18:48 crc kubenswrapper[4556]: E0218 09:18:48.439235 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4b9aab0d6c8aa124f35a7d50b97576b617bd928e81f72be3d4072a9ae816a71\": container with ID starting with f4b9aab0d6c8aa124f35a7d50b97576b617bd928e81f72be3d4072a9ae816a71 not found: ID does not exist" containerID="f4b9aab0d6c8aa124f35a7d50b97576b617bd928e81f72be3d4072a9ae816a71" Feb 18 09:18:48 crc kubenswrapper[4556]: I0218 09:18:48.439262 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4b9aab0d6c8aa124f35a7d50b97576b617bd928e81f72be3d4072a9ae816a71"} err="failed to get container status \"f4b9aab0d6c8aa124f35a7d50b97576b617bd928e81f72be3d4072a9ae816a71\": rpc error: code = NotFound desc = could not find container \"f4b9aab0d6c8aa124f35a7d50b97576b617bd928e81f72be3d4072a9ae816a71\": container with ID starting with f4b9aab0d6c8aa124f35a7d50b97576b617bd928e81f72be3d4072a9ae816a71 not found: ID does not exist" Feb 18 09:18:48 crc kubenswrapper[4556]: I0218 09:18:48.439278 4556 scope.go:117] "RemoveContainer" containerID="0ed40eaa96898558bbf3a3f10d8fc68051c8f2daee7f7acaaf04c49222f3285c" Feb 18 09:18:48 crc kubenswrapper[4556]: E0218 09:18:48.439566 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ed40eaa96898558bbf3a3f10d8fc68051c8f2daee7f7acaaf04c49222f3285c\": container with ID starting with 0ed40eaa96898558bbf3a3f10d8fc68051c8f2daee7f7acaaf04c49222f3285c not found: ID does not exist" containerID="0ed40eaa96898558bbf3a3f10d8fc68051c8f2daee7f7acaaf04c49222f3285c" Feb 18 09:18:48 crc kubenswrapper[4556]: I0218 09:18:48.439613 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ed40eaa96898558bbf3a3f10d8fc68051c8f2daee7f7acaaf04c49222f3285c"} err="failed to get container status \"0ed40eaa96898558bbf3a3f10d8fc68051c8f2daee7f7acaaf04c49222f3285c\": rpc error: code = NotFound desc = could not find container \"0ed40eaa96898558bbf3a3f10d8fc68051c8f2daee7f7acaaf04c49222f3285c\": container with ID starting with 0ed40eaa96898558bbf3a3f10d8fc68051c8f2daee7f7acaaf04c49222f3285c not found: ID does not exist" Feb 18 09:18:48 crc kubenswrapper[4556]: I0218 09:18:48.680699 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 18 09:18:48 crc kubenswrapper[4556]: I0218 09:18:48.751559 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59767f41-27b9-4146-b723-ea8df508a92e-scripts\") pod \"59767f41-27b9-4146-b723-ea8df508a92e\" (UID: \"59767f41-27b9-4146-b723-ea8df508a92e\") " Feb 18 09:18:48 crc kubenswrapper[4556]: I0218 09:18:48.751678 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/59767f41-27b9-4146-b723-ea8df508a92e-config-data-custom\") pod \"59767f41-27b9-4146-b723-ea8df508a92e\" (UID: \"59767f41-27b9-4146-b723-ea8df508a92e\") " Feb 18 09:18:48 crc kubenswrapper[4556]: I0218 09:18:48.751729 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/59767f41-27b9-4146-b723-ea8df508a92e-etc-machine-id\") pod \"59767f41-27b9-4146-b723-ea8df508a92e\" (UID: \"59767f41-27b9-4146-b723-ea8df508a92e\") " Feb 18 09:18:48 crc kubenswrapper[4556]: I0218 09:18:48.751956 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vhpsl\" (UniqueName: \"kubernetes.io/projected/59767f41-27b9-4146-b723-ea8df508a92e-kube-api-access-vhpsl\") pod \"59767f41-27b9-4146-b723-ea8df508a92e\" (UID: \"59767f41-27b9-4146-b723-ea8df508a92e\") " Feb 18 09:18:48 crc kubenswrapper[4556]: I0218 09:18:48.752022 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59767f41-27b9-4146-b723-ea8df508a92e-config-data\") pod \"59767f41-27b9-4146-b723-ea8df508a92e\" (UID: \"59767f41-27b9-4146-b723-ea8df508a92e\") " Feb 18 09:18:48 crc kubenswrapper[4556]: I0218 09:18:48.752068 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59767f41-27b9-4146-b723-ea8df508a92e-combined-ca-bundle\") pod \"59767f41-27b9-4146-b723-ea8df508a92e\" (UID: \"59767f41-27b9-4146-b723-ea8df508a92e\") " Feb 18 09:18:48 crc kubenswrapper[4556]: I0218 09:18:48.752007 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/59767f41-27b9-4146-b723-ea8df508a92e-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "59767f41-27b9-4146-b723-ea8df508a92e" (UID: "59767f41-27b9-4146-b723-ea8df508a92e"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 09:18:48 crc kubenswrapper[4556]: I0218 09:18:48.752842 4556 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/59767f41-27b9-4146-b723-ea8df508a92e-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:48 crc kubenswrapper[4556]: I0218 09:18:48.757343 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59767f41-27b9-4146-b723-ea8df508a92e-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "59767f41-27b9-4146-b723-ea8df508a92e" (UID: "59767f41-27b9-4146-b723-ea8df508a92e"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:18:48 crc kubenswrapper[4556]: I0218 09:18:48.759418 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59767f41-27b9-4146-b723-ea8df508a92e-kube-api-access-vhpsl" (OuterVolumeSpecName: "kube-api-access-vhpsl") pod "59767f41-27b9-4146-b723-ea8df508a92e" (UID: "59767f41-27b9-4146-b723-ea8df508a92e"). InnerVolumeSpecName "kube-api-access-vhpsl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:18:48 crc kubenswrapper[4556]: I0218 09:18:48.761533 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59767f41-27b9-4146-b723-ea8df508a92e-scripts" (OuterVolumeSpecName: "scripts") pod "59767f41-27b9-4146-b723-ea8df508a92e" (UID: "59767f41-27b9-4146-b723-ea8df508a92e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:18:48 crc kubenswrapper[4556]: I0218 09:18:48.816688 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59767f41-27b9-4146-b723-ea8df508a92e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "59767f41-27b9-4146-b723-ea8df508a92e" (UID: "59767f41-27b9-4146-b723-ea8df508a92e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:18:48 crc kubenswrapper[4556]: I0218 09:18:48.854605 4556 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59767f41-27b9-4146-b723-ea8df508a92e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:48 crc kubenswrapper[4556]: I0218 09:18:48.854650 4556 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59767f41-27b9-4146-b723-ea8df508a92e-scripts\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:48 crc kubenswrapper[4556]: I0218 09:18:48.854661 4556 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/59767f41-27b9-4146-b723-ea8df508a92e-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:48 crc kubenswrapper[4556]: I0218 09:18:48.854670 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vhpsl\" (UniqueName: \"kubernetes.io/projected/59767f41-27b9-4146-b723-ea8df508a92e-kube-api-access-vhpsl\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:48 crc kubenswrapper[4556]: I0218 09:18:48.866967 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59767f41-27b9-4146-b723-ea8df508a92e-config-data" (OuterVolumeSpecName: "config-data") pod "59767f41-27b9-4146-b723-ea8df508a92e" (UID: "59767f41-27b9-4146-b723-ea8df508a92e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:18:48 crc kubenswrapper[4556]: I0218 09:18:48.956774 4556 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59767f41-27b9-4146-b723-ea8df508a92e-config-data\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.326373 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18385928-f2f7-4f58-9f83-e563af72c8b4" path="/var/lib/kubelet/pods/18385928-f2f7-4f58-9f83-e563af72c8b4/volumes" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.327029 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1fb56420-2137-43d2-b790-1773daf3f0c2" path="/var/lib/kubelet/pods/1fb56420-2137-43d2-b790-1773daf3f0c2/volumes" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.374828 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"59767f41-27b9-4146-b723-ea8df508a92e","Type":"ContainerDied","Data":"ea8e05987e69fa9665be060865cd530276a85b3405d4d7496e36384c8ea914e4"} Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.374898 4556 scope.go:117] "RemoveContainer" containerID="5bcd75d3aae13e3e94b1fa31eb135277d4a4343b64ada8a1ab708f2ff72fad1e" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.375051 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.397641 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.399741 4556 scope.go:117] "RemoveContainer" containerID="ed41c2e1779bc2ff90abc289c7c836a56a67c4f7e881c290aa073046688b2f99" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.402990 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.427561 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Feb 18 09:18:49 crc kubenswrapper[4556]: E0218 09:18:49.427906 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fb56420-2137-43d2-b790-1773daf3f0c2" containerName="extract-utilities" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.427924 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fb56420-2137-43d2-b790-1773daf3f0c2" containerName="extract-utilities" Feb 18 09:18:49 crc kubenswrapper[4556]: E0218 09:18:49.427945 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b92f250f-e888-4ff5-ae26-9e31c7be3379" containerName="neutron-httpd" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.427952 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="b92f250f-e888-4ff5-ae26-9e31c7be3379" containerName="neutron-httpd" Feb 18 09:18:49 crc kubenswrapper[4556]: E0218 09:18:49.427965 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afe9fc26-12ca-41b2-8076-6d8374384041" containerName="init" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.427970 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="afe9fc26-12ca-41b2-8076-6d8374384041" containerName="init" Feb 18 09:18:49 crc kubenswrapper[4556]: E0218 09:18:49.427981 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50228011-f908-465e-a754-d4b6b626dda9" containerName="horizon" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.427988 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="50228011-f908-465e-a754-d4b6b626dda9" containerName="horizon" Feb 18 09:18:49 crc kubenswrapper[4556]: E0218 09:18:49.427995 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5fec43c-7b17-45d5-91c4-609b6d2ca1b9" containerName="horizon" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.428000 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5fec43c-7b17-45d5-91c4-609b6d2ca1b9" containerName="horizon" Feb 18 09:18:49 crc kubenswrapper[4556]: E0218 09:18:49.428008 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98" containerName="horizon-log" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.428014 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98" containerName="horizon-log" Feb 18 09:18:49 crc kubenswrapper[4556]: E0218 09:18:49.428023 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02b5c258-0253-4d75-852b-30c2b26b5598" containerName="dnsmasq-dns" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.428030 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="02b5c258-0253-4d75-852b-30c2b26b5598" containerName="dnsmasq-dns" Feb 18 09:18:49 crc kubenswrapper[4556]: E0218 09:18:49.428039 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59767f41-27b9-4146-b723-ea8df508a92e" containerName="probe" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.428045 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="59767f41-27b9-4146-b723-ea8df508a92e" containerName="probe" Feb 18 09:18:49 crc kubenswrapper[4556]: E0218 09:18:49.428050 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b92f250f-e888-4ff5-ae26-9e31c7be3379" containerName="neutron-api" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.428056 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="b92f250f-e888-4ff5-ae26-9e31c7be3379" containerName="neutron-api" Feb 18 09:18:49 crc kubenswrapper[4556]: E0218 09:18:49.428064 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fb56420-2137-43d2-b790-1773daf3f0c2" containerName="extract-content" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.428070 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fb56420-2137-43d2-b790-1773daf3f0c2" containerName="extract-content" Feb 18 09:18:49 crc kubenswrapper[4556]: E0218 09:18:49.428079 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59767f41-27b9-4146-b723-ea8df508a92e" containerName="cinder-scheduler" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.428085 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="59767f41-27b9-4146-b723-ea8df508a92e" containerName="cinder-scheduler" Feb 18 09:18:49 crc kubenswrapper[4556]: E0218 09:18:49.428093 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98" containerName="horizon" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.428099 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98" containerName="horizon" Feb 18 09:18:49 crc kubenswrapper[4556]: E0218 09:18:49.428106 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18385928-f2f7-4f58-9f83-e563af72c8b4" containerName="barbican-api" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.428112 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="18385928-f2f7-4f58-9f83-e563af72c8b4" containerName="barbican-api" Feb 18 09:18:49 crc kubenswrapper[4556]: E0218 09:18:49.428118 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5fec43c-7b17-45d5-91c4-609b6d2ca1b9" containerName="horizon-log" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.428123 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5fec43c-7b17-45d5-91c4-609b6d2ca1b9" containerName="horizon-log" Feb 18 09:18:49 crc kubenswrapper[4556]: E0218 09:18:49.428133 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afe9fc26-12ca-41b2-8076-6d8374384041" containerName="dnsmasq-dns" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.428139 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="afe9fc26-12ca-41b2-8076-6d8374384041" containerName="dnsmasq-dns" Feb 18 09:18:49 crc kubenswrapper[4556]: E0218 09:18:49.428163 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02b5c258-0253-4d75-852b-30c2b26b5598" containerName="init" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.428170 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="02b5c258-0253-4d75-852b-30c2b26b5598" containerName="init" Feb 18 09:18:49 crc kubenswrapper[4556]: E0218 09:18:49.428181 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50228011-f908-465e-a754-d4b6b626dda9" containerName="horizon-log" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.428186 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="50228011-f908-465e-a754-d4b6b626dda9" containerName="horizon-log" Feb 18 09:18:49 crc kubenswrapper[4556]: E0218 09:18:49.428199 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fb56420-2137-43d2-b790-1773daf3f0c2" containerName="registry-server" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.428204 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fb56420-2137-43d2-b790-1773daf3f0c2" containerName="registry-server" Feb 18 09:18:49 crc kubenswrapper[4556]: E0218 09:18:49.428229 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18385928-f2f7-4f58-9f83-e563af72c8b4" containerName="barbican-api-log" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.428236 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="18385928-f2f7-4f58-9f83-e563af72c8b4" containerName="barbican-api-log" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.428415 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98" containerName="horizon-log" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.428427 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="50228011-f908-465e-a754-d4b6b626dda9" containerName="horizon" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.428436 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5fec43c-7b17-45d5-91c4-609b6d2ca1b9" containerName="horizon-log" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.428449 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="59767f41-27b9-4146-b723-ea8df508a92e" containerName="cinder-scheduler" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.428458 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="1fb56420-2137-43d2-b790-1773daf3f0c2" containerName="registry-server" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.428463 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="50228011-f908-465e-a754-d4b6b626dda9" containerName="horizon-log" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.428469 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="afe9fc26-12ca-41b2-8076-6d8374384041" containerName="dnsmasq-dns" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.428478 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="18385928-f2f7-4f58-9f83-e563af72c8b4" containerName="barbican-api-log" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.428484 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="b92f250f-e888-4ff5-ae26-9e31c7be3379" containerName="neutron-api" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.428491 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="59767f41-27b9-4146-b723-ea8df508a92e" containerName="probe" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.428497 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="66b0884b-6c8f-4e9d-a6ea-7891cb0c1a98" containerName="horizon" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.428507 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="b92f250f-e888-4ff5-ae26-9e31c7be3379" containerName="neutron-httpd" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.428515 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="18385928-f2f7-4f58-9f83-e563af72c8b4" containerName="barbican-api" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.428526 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="02b5c258-0253-4d75-852b-30c2b26b5598" containerName="dnsmasq-dns" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.428534 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5fec43c-7b17-45d5-91c4-609b6d2ca1b9" containerName="horizon" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.429437 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.447083 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.454207 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.569804 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca98a45d-9b95-4e00-8179-feb5a6c1ddb0-scripts\") pod \"cinder-scheduler-0\" (UID: \"ca98a45d-9b95-4e00-8179-feb5a6c1ddb0\") " pod="openstack/cinder-scheduler-0" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.569963 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ca98a45d-9b95-4e00-8179-feb5a6c1ddb0-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"ca98a45d-9b95-4e00-8179-feb5a6c1ddb0\") " pod="openstack/cinder-scheduler-0" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.570004 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbwzz\" (UniqueName: \"kubernetes.io/projected/ca98a45d-9b95-4e00-8179-feb5a6c1ddb0-kube-api-access-xbwzz\") pod \"cinder-scheduler-0\" (UID: \"ca98a45d-9b95-4e00-8179-feb5a6c1ddb0\") " pod="openstack/cinder-scheduler-0" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.570125 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca98a45d-9b95-4e00-8179-feb5a6c1ddb0-config-data\") pod \"cinder-scheduler-0\" (UID: \"ca98a45d-9b95-4e00-8179-feb5a6c1ddb0\") " pod="openstack/cinder-scheduler-0" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.570352 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ca98a45d-9b95-4e00-8179-feb5a6c1ddb0-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"ca98a45d-9b95-4e00-8179-feb5a6c1ddb0\") " pod="openstack/cinder-scheduler-0" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.570408 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca98a45d-9b95-4e00-8179-feb5a6c1ddb0-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"ca98a45d-9b95-4e00-8179-feb5a6c1ddb0\") " pod="openstack/cinder-scheduler-0" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.672381 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ca98a45d-9b95-4e00-8179-feb5a6c1ddb0-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"ca98a45d-9b95-4e00-8179-feb5a6c1ddb0\") " pod="openstack/cinder-scheduler-0" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.672426 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbwzz\" (UniqueName: \"kubernetes.io/projected/ca98a45d-9b95-4e00-8179-feb5a6c1ddb0-kube-api-access-xbwzz\") pod \"cinder-scheduler-0\" (UID: \"ca98a45d-9b95-4e00-8179-feb5a6c1ddb0\") " pod="openstack/cinder-scheduler-0" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.672478 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca98a45d-9b95-4e00-8179-feb5a6c1ddb0-config-data\") pod \"cinder-scheduler-0\" (UID: \"ca98a45d-9b95-4e00-8179-feb5a6c1ddb0\") " pod="openstack/cinder-scheduler-0" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.672516 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ca98a45d-9b95-4e00-8179-feb5a6c1ddb0-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"ca98a45d-9b95-4e00-8179-feb5a6c1ddb0\") " pod="openstack/cinder-scheduler-0" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.672548 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ca98a45d-9b95-4e00-8179-feb5a6c1ddb0-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"ca98a45d-9b95-4e00-8179-feb5a6c1ddb0\") " pod="openstack/cinder-scheduler-0" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.672679 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca98a45d-9b95-4e00-8179-feb5a6c1ddb0-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"ca98a45d-9b95-4e00-8179-feb5a6c1ddb0\") " pod="openstack/cinder-scheduler-0" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.672851 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca98a45d-9b95-4e00-8179-feb5a6c1ddb0-scripts\") pod \"cinder-scheduler-0\" (UID: \"ca98a45d-9b95-4e00-8179-feb5a6c1ddb0\") " pod="openstack/cinder-scheduler-0" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.677950 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ca98a45d-9b95-4e00-8179-feb5a6c1ddb0-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"ca98a45d-9b95-4e00-8179-feb5a6c1ddb0\") " pod="openstack/cinder-scheduler-0" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.678256 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca98a45d-9b95-4e00-8179-feb5a6c1ddb0-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"ca98a45d-9b95-4e00-8179-feb5a6c1ddb0\") " pod="openstack/cinder-scheduler-0" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.678423 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca98a45d-9b95-4e00-8179-feb5a6c1ddb0-scripts\") pod \"cinder-scheduler-0\" (UID: \"ca98a45d-9b95-4e00-8179-feb5a6c1ddb0\") " pod="openstack/cinder-scheduler-0" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.679688 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca98a45d-9b95-4e00-8179-feb5a6c1ddb0-config-data\") pod \"cinder-scheduler-0\" (UID: \"ca98a45d-9b95-4e00-8179-feb5a6c1ddb0\") " pod="openstack/cinder-scheduler-0" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.691148 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbwzz\" (UniqueName: \"kubernetes.io/projected/ca98a45d-9b95-4e00-8179-feb5a6c1ddb0-kube-api-access-xbwzz\") pod \"cinder-scheduler-0\" (UID: \"ca98a45d-9b95-4e00-8179-feb5a6c1ddb0\") " pod="openstack/cinder-scheduler-0" Feb 18 09:18:49 crc kubenswrapper[4556]: I0218 09:18:49.762228 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 18 09:18:50 crc kubenswrapper[4556]: I0218 09:18:50.200872 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 18 09:18:50 crc kubenswrapper[4556]: I0218 09:18:50.392546 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"ca98a45d-9b95-4e00-8179-feb5a6c1ddb0","Type":"ContainerStarted","Data":"5e98b475fb94a3e884997aba2302096352c1452f4929599b980db7a8dde6289d"} Feb 18 09:18:51 crc kubenswrapper[4556]: I0218 09:18:51.073633 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rmt7n"] Feb 18 09:18:51 crc kubenswrapper[4556]: I0218 09:18:51.075636 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rmt7n" Feb 18 09:18:51 crc kubenswrapper[4556]: I0218 09:18:51.084729 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rmt7n"] Feb 18 09:18:51 crc kubenswrapper[4556]: I0218 09:18:51.207029 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a557635-d67d-4e93-a82c-990c34fe7d48-utilities\") pod \"redhat-operators-rmt7n\" (UID: \"0a557635-d67d-4e93-a82c-990c34fe7d48\") " pod="openshift-marketplace/redhat-operators-rmt7n" Feb 18 09:18:51 crc kubenswrapper[4556]: I0218 09:18:51.207289 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqss7\" (UniqueName: \"kubernetes.io/projected/0a557635-d67d-4e93-a82c-990c34fe7d48-kube-api-access-rqss7\") pod \"redhat-operators-rmt7n\" (UID: \"0a557635-d67d-4e93-a82c-990c34fe7d48\") " pod="openshift-marketplace/redhat-operators-rmt7n" Feb 18 09:18:51 crc kubenswrapper[4556]: I0218 09:18:51.207349 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a557635-d67d-4e93-a82c-990c34fe7d48-catalog-content\") pod \"redhat-operators-rmt7n\" (UID: \"0a557635-d67d-4e93-a82c-990c34fe7d48\") " pod="openshift-marketplace/redhat-operators-rmt7n" Feb 18 09:18:51 crc kubenswrapper[4556]: I0218 09:18:51.290170 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59767f41-27b9-4146-b723-ea8df508a92e" path="/var/lib/kubelet/pods/59767f41-27b9-4146-b723-ea8df508a92e/volumes" Feb 18 09:18:51 crc kubenswrapper[4556]: I0218 09:18:51.309073 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a557635-d67d-4e93-a82c-990c34fe7d48-utilities\") pod \"redhat-operators-rmt7n\" (UID: \"0a557635-d67d-4e93-a82c-990c34fe7d48\") " pod="openshift-marketplace/redhat-operators-rmt7n" Feb 18 09:18:51 crc kubenswrapper[4556]: I0218 09:18:51.309116 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqss7\" (UniqueName: \"kubernetes.io/projected/0a557635-d67d-4e93-a82c-990c34fe7d48-kube-api-access-rqss7\") pod \"redhat-operators-rmt7n\" (UID: \"0a557635-d67d-4e93-a82c-990c34fe7d48\") " pod="openshift-marketplace/redhat-operators-rmt7n" Feb 18 09:18:51 crc kubenswrapper[4556]: I0218 09:18:51.309193 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a557635-d67d-4e93-a82c-990c34fe7d48-catalog-content\") pod \"redhat-operators-rmt7n\" (UID: \"0a557635-d67d-4e93-a82c-990c34fe7d48\") " pod="openshift-marketplace/redhat-operators-rmt7n" Feb 18 09:18:51 crc kubenswrapper[4556]: I0218 09:18:51.309521 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a557635-d67d-4e93-a82c-990c34fe7d48-utilities\") pod \"redhat-operators-rmt7n\" (UID: \"0a557635-d67d-4e93-a82c-990c34fe7d48\") " pod="openshift-marketplace/redhat-operators-rmt7n" Feb 18 09:18:51 crc kubenswrapper[4556]: I0218 09:18:51.309585 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a557635-d67d-4e93-a82c-990c34fe7d48-catalog-content\") pod \"redhat-operators-rmt7n\" (UID: \"0a557635-d67d-4e93-a82c-990c34fe7d48\") " pod="openshift-marketplace/redhat-operators-rmt7n" Feb 18 09:18:51 crc kubenswrapper[4556]: I0218 09:18:51.330065 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqss7\" (UniqueName: \"kubernetes.io/projected/0a557635-d67d-4e93-a82c-990c34fe7d48-kube-api-access-rqss7\") pod \"redhat-operators-rmt7n\" (UID: \"0a557635-d67d-4e93-a82c-990c34fe7d48\") " pod="openshift-marketplace/redhat-operators-rmt7n" Feb 18 09:18:51 crc kubenswrapper[4556]: I0218 09:18:51.399945 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rmt7n" Feb 18 09:18:51 crc kubenswrapper[4556]: I0218 09:18:51.404687 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"ca98a45d-9b95-4e00-8179-feb5a6c1ddb0","Type":"ContainerStarted","Data":"0fd5f13edeac3bc3a1be87688da43be04c6f7583aa3f37d257c2c31a63e6dec3"} Feb 18 09:18:51 crc kubenswrapper[4556]: I0218 09:18:51.404727 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"ca98a45d-9b95-4e00-8179-feb5a6c1ddb0","Type":"ContainerStarted","Data":"eb76e4d5a10d75fb36bfb3fa0a8f57431bc385030e331c2c1272089d31d1cb56"} Feb 18 09:18:51 crc kubenswrapper[4556]: I0218 09:18:51.425987 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=2.425966685 podStartE2EDuration="2.425966685s" podCreationTimestamp="2026-02-18 09:18:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:18:51.417489588 +0000 UTC m=+888.434450559" watchObservedRunningTime="2026-02-18 09:18:51.425966685 +0000 UTC m=+888.442927665" Feb 18 09:18:51 crc kubenswrapper[4556]: I0218 09:18:51.739167 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Feb 18 09:18:51 crc kubenswrapper[4556]: I0218 09:18:51.740307 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 18 09:18:51 crc kubenswrapper[4556]: I0218 09:18:51.744421 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Feb 18 09:18:51 crc kubenswrapper[4556]: I0218 09:18:51.744447 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Feb 18 09:18:51 crc kubenswrapper[4556]: I0218 09:18:51.744776 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-k9llr" Feb 18 09:18:51 crc kubenswrapper[4556]: I0218 09:18:51.753329 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Feb 18 09:18:51 crc kubenswrapper[4556]: I0218 09:18:51.821417 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lh22g\" (UniqueName: \"kubernetes.io/projected/49e26f9f-38d3-4b82-8ef7-0bf33efbbb25-kube-api-access-lh22g\") pod \"openstackclient\" (UID: \"49e26f9f-38d3-4b82-8ef7-0bf33efbbb25\") " pod="openstack/openstackclient" Feb 18 09:18:51 crc kubenswrapper[4556]: I0218 09:18:51.821470 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49e26f9f-38d3-4b82-8ef7-0bf33efbbb25-combined-ca-bundle\") pod \"openstackclient\" (UID: \"49e26f9f-38d3-4b82-8ef7-0bf33efbbb25\") " pod="openstack/openstackclient" Feb 18 09:18:51 crc kubenswrapper[4556]: I0218 09:18:51.821528 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/49e26f9f-38d3-4b82-8ef7-0bf33efbbb25-openstack-config\") pod \"openstackclient\" (UID: \"49e26f9f-38d3-4b82-8ef7-0bf33efbbb25\") " pod="openstack/openstackclient" Feb 18 09:18:51 crc kubenswrapper[4556]: I0218 09:18:51.821622 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/49e26f9f-38d3-4b82-8ef7-0bf33efbbb25-openstack-config-secret\") pod \"openstackclient\" (UID: \"49e26f9f-38d3-4b82-8ef7-0bf33efbbb25\") " pod="openstack/openstackclient" Feb 18 09:18:51 crc kubenswrapper[4556]: I0218 09:18:51.853463 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rmt7n"] Feb 18 09:18:51 crc kubenswrapper[4556]: I0218 09:18:51.924738 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lh22g\" (UniqueName: \"kubernetes.io/projected/49e26f9f-38d3-4b82-8ef7-0bf33efbbb25-kube-api-access-lh22g\") pod \"openstackclient\" (UID: \"49e26f9f-38d3-4b82-8ef7-0bf33efbbb25\") " pod="openstack/openstackclient" Feb 18 09:18:51 crc kubenswrapper[4556]: I0218 09:18:51.924809 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49e26f9f-38d3-4b82-8ef7-0bf33efbbb25-combined-ca-bundle\") pod \"openstackclient\" (UID: \"49e26f9f-38d3-4b82-8ef7-0bf33efbbb25\") " pod="openstack/openstackclient" Feb 18 09:18:51 crc kubenswrapper[4556]: I0218 09:18:51.924904 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/49e26f9f-38d3-4b82-8ef7-0bf33efbbb25-openstack-config\") pod \"openstackclient\" (UID: \"49e26f9f-38d3-4b82-8ef7-0bf33efbbb25\") " pod="openstack/openstackclient" Feb 18 09:18:51 crc kubenswrapper[4556]: I0218 09:18:51.925034 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/49e26f9f-38d3-4b82-8ef7-0bf33efbbb25-openstack-config-secret\") pod \"openstackclient\" (UID: \"49e26f9f-38d3-4b82-8ef7-0bf33efbbb25\") " pod="openstack/openstackclient" Feb 18 09:18:51 crc kubenswrapper[4556]: I0218 09:18:51.929088 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/49e26f9f-38d3-4b82-8ef7-0bf33efbbb25-openstack-config\") pod \"openstackclient\" (UID: \"49e26f9f-38d3-4b82-8ef7-0bf33efbbb25\") " pod="openstack/openstackclient" Feb 18 09:18:51 crc kubenswrapper[4556]: I0218 09:18:51.930687 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49e26f9f-38d3-4b82-8ef7-0bf33efbbb25-combined-ca-bundle\") pod \"openstackclient\" (UID: \"49e26f9f-38d3-4b82-8ef7-0bf33efbbb25\") " pod="openstack/openstackclient" Feb 18 09:18:51 crc kubenswrapper[4556]: I0218 09:18:51.931291 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/49e26f9f-38d3-4b82-8ef7-0bf33efbbb25-openstack-config-secret\") pod \"openstackclient\" (UID: \"49e26f9f-38d3-4b82-8ef7-0bf33efbbb25\") " pod="openstack/openstackclient" Feb 18 09:18:51 crc kubenswrapper[4556]: I0218 09:18:51.944401 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lh22g\" (UniqueName: \"kubernetes.io/projected/49e26f9f-38d3-4b82-8ef7-0bf33efbbb25-kube-api-access-lh22g\") pod \"openstackclient\" (UID: \"49e26f9f-38d3-4b82-8ef7-0bf33efbbb25\") " pod="openstack/openstackclient" Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.070992 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.145257 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-864fbf8dcf-chrjx"] Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.156449 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-864fbf8dcf-chrjx" Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.163140 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.163578 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.163932 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.228317 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-864fbf8dcf-chrjx"] Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.255099 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d678be18-d610-4ea8-b248-47843cf74ea3-combined-ca-bundle\") pod \"swift-proxy-864fbf8dcf-chrjx\" (UID: \"d678be18-d610-4ea8-b248-47843cf74ea3\") " pod="openstack/swift-proxy-864fbf8dcf-chrjx" Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.255165 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvfmp\" (UniqueName: \"kubernetes.io/projected/d678be18-d610-4ea8-b248-47843cf74ea3-kube-api-access-tvfmp\") pod \"swift-proxy-864fbf8dcf-chrjx\" (UID: \"d678be18-d610-4ea8-b248-47843cf74ea3\") " pod="openstack/swift-proxy-864fbf8dcf-chrjx" Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.255322 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d678be18-d610-4ea8-b248-47843cf74ea3-log-httpd\") pod \"swift-proxy-864fbf8dcf-chrjx\" (UID: \"d678be18-d610-4ea8-b248-47843cf74ea3\") " pod="openstack/swift-proxy-864fbf8dcf-chrjx" Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.255366 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d678be18-d610-4ea8-b248-47843cf74ea3-internal-tls-certs\") pod \"swift-proxy-864fbf8dcf-chrjx\" (UID: \"d678be18-d610-4ea8-b248-47843cf74ea3\") " pod="openstack/swift-proxy-864fbf8dcf-chrjx" Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.255401 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d678be18-d610-4ea8-b248-47843cf74ea3-config-data\") pod \"swift-proxy-864fbf8dcf-chrjx\" (UID: \"d678be18-d610-4ea8-b248-47843cf74ea3\") " pod="openstack/swift-proxy-864fbf8dcf-chrjx" Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.255512 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d678be18-d610-4ea8-b248-47843cf74ea3-public-tls-certs\") pod \"swift-proxy-864fbf8dcf-chrjx\" (UID: \"d678be18-d610-4ea8-b248-47843cf74ea3\") " pod="openstack/swift-proxy-864fbf8dcf-chrjx" Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.255560 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d678be18-d610-4ea8-b248-47843cf74ea3-run-httpd\") pod \"swift-proxy-864fbf8dcf-chrjx\" (UID: \"d678be18-d610-4ea8-b248-47843cf74ea3\") " pod="openstack/swift-proxy-864fbf8dcf-chrjx" Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.255582 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d678be18-d610-4ea8-b248-47843cf74ea3-etc-swift\") pod \"swift-proxy-864fbf8dcf-chrjx\" (UID: \"d678be18-d610-4ea8-b248-47843cf74ea3\") " pod="openstack/swift-proxy-864fbf8dcf-chrjx" Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.280636 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.290813 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.300640 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Feb 18 09:18:52 crc kubenswrapper[4556]: E0218 09:18:52.311328 4556 log.go:32] "RunPodSandbox from runtime service failed" err=< Feb 18 09:18:52 crc kubenswrapper[4556]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_49e26f9f-38d3-4b82-8ef7-0bf33efbbb25_0(c5e240311b87faca8c4175d3eebe5409b6ecebce83847326cdd0fdbce0275e75): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"c5e240311b87faca8c4175d3eebe5409b6ecebce83847326cdd0fdbce0275e75" Netns:"/var/run/netns/5ac3ab10-6e79-4f38-9c3f-3814313ae496" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=c5e240311b87faca8c4175d3eebe5409b6ecebce83847326cdd0fdbce0275e75;K8S_POD_UID=49e26f9f-38d3-4b82-8ef7-0bf33efbbb25" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/49e26f9f-38d3-4b82-8ef7-0bf33efbbb25]: expected pod UID "49e26f9f-38d3-4b82-8ef7-0bf33efbbb25" but got "97268f30-e67a-441f-93cd-0dc9771dbd42" from Kube API Feb 18 09:18:52 crc kubenswrapper[4556]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Feb 18 09:18:52 crc kubenswrapper[4556]: > Feb 18 09:18:52 crc kubenswrapper[4556]: E0218 09:18:52.311394 4556 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Feb 18 09:18:52 crc kubenswrapper[4556]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_49e26f9f-38d3-4b82-8ef7-0bf33efbbb25_0(c5e240311b87faca8c4175d3eebe5409b6ecebce83847326cdd0fdbce0275e75): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"c5e240311b87faca8c4175d3eebe5409b6ecebce83847326cdd0fdbce0275e75" Netns:"/var/run/netns/5ac3ab10-6e79-4f38-9c3f-3814313ae496" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=c5e240311b87faca8c4175d3eebe5409b6ecebce83847326cdd0fdbce0275e75;K8S_POD_UID=49e26f9f-38d3-4b82-8ef7-0bf33efbbb25" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/49e26f9f-38d3-4b82-8ef7-0bf33efbbb25]: expected pod UID "49e26f9f-38d3-4b82-8ef7-0bf33efbbb25" but got "97268f30-e67a-441f-93cd-0dc9771dbd42" from Kube API Feb 18 09:18:52 crc kubenswrapper[4556]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Feb 18 09:18:52 crc kubenswrapper[4556]: > pod="openstack/openstackclient" Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.311750 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.321504 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.361350 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d678be18-d610-4ea8-b248-47843cf74ea3-log-httpd\") pod \"swift-proxy-864fbf8dcf-chrjx\" (UID: \"d678be18-d610-4ea8-b248-47843cf74ea3\") " pod="openstack/swift-proxy-864fbf8dcf-chrjx" Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.361402 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d678be18-d610-4ea8-b248-47843cf74ea3-internal-tls-certs\") pod \"swift-proxy-864fbf8dcf-chrjx\" (UID: \"d678be18-d610-4ea8-b248-47843cf74ea3\") " pod="openstack/swift-proxy-864fbf8dcf-chrjx" Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.361434 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d678be18-d610-4ea8-b248-47843cf74ea3-config-data\") pod \"swift-proxy-864fbf8dcf-chrjx\" (UID: \"d678be18-d610-4ea8-b248-47843cf74ea3\") " pod="openstack/swift-proxy-864fbf8dcf-chrjx" Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.361507 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d678be18-d610-4ea8-b248-47843cf74ea3-public-tls-certs\") pod \"swift-proxy-864fbf8dcf-chrjx\" (UID: \"d678be18-d610-4ea8-b248-47843cf74ea3\") " pod="openstack/swift-proxy-864fbf8dcf-chrjx" Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.361542 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d678be18-d610-4ea8-b248-47843cf74ea3-run-httpd\") pod \"swift-proxy-864fbf8dcf-chrjx\" (UID: \"d678be18-d610-4ea8-b248-47843cf74ea3\") " pod="openstack/swift-proxy-864fbf8dcf-chrjx" Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.361563 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d678be18-d610-4ea8-b248-47843cf74ea3-etc-swift\") pod \"swift-proxy-864fbf8dcf-chrjx\" (UID: \"d678be18-d610-4ea8-b248-47843cf74ea3\") " pod="openstack/swift-proxy-864fbf8dcf-chrjx" Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.361597 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d678be18-d610-4ea8-b248-47843cf74ea3-combined-ca-bundle\") pod \"swift-proxy-864fbf8dcf-chrjx\" (UID: \"d678be18-d610-4ea8-b248-47843cf74ea3\") " pod="openstack/swift-proxy-864fbf8dcf-chrjx" Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.361617 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvfmp\" (UniqueName: \"kubernetes.io/projected/d678be18-d610-4ea8-b248-47843cf74ea3-kube-api-access-tvfmp\") pod \"swift-proxy-864fbf8dcf-chrjx\" (UID: \"d678be18-d610-4ea8-b248-47843cf74ea3\") " pod="openstack/swift-proxy-864fbf8dcf-chrjx" Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.362285 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d678be18-d610-4ea8-b248-47843cf74ea3-log-httpd\") pod \"swift-proxy-864fbf8dcf-chrjx\" (UID: \"d678be18-d610-4ea8-b248-47843cf74ea3\") " pod="openstack/swift-proxy-864fbf8dcf-chrjx" Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.362911 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d678be18-d610-4ea8-b248-47843cf74ea3-run-httpd\") pod \"swift-proxy-864fbf8dcf-chrjx\" (UID: \"d678be18-d610-4ea8-b248-47843cf74ea3\") " pod="openstack/swift-proxy-864fbf8dcf-chrjx" Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.371985 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d678be18-d610-4ea8-b248-47843cf74ea3-config-data\") pod \"swift-proxy-864fbf8dcf-chrjx\" (UID: \"d678be18-d610-4ea8-b248-47843cf74ea3\") " pod="openstack/swift-proxy-864fbf8dcf-chrjx" Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.373046 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d678be18-d610-4ea8-b248-47843cf74ea3-public-tls-certs\") pod \"swift-proxy-864fbf8dcf-chrjx\" (UID: \"d678be18-d610-4ea8-b248-47843cf74ea3\") " pod="openstack/swift-proxy-864fbf8dcf-chrjx" Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.376673 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d678be18-d610-4ea8-b248-47843cf74ea3-combined-ca-bundle\") pod \"swift-proxy-864fbf8dcf-chrjx\" (UID: \"d678be18-d610-4ea8-b248-47843cf74ea3\") " pod="openstack/swift-proxy-864fbf8dcf-chrjx" Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.376728 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d678be18-d610-4ea8-b248-47843cf74ea3-internal-tls-certs\") pod \"swift-proxy-864fbf8dcf-chrjx\" (UID: \"d678be18-d610-4ea8-b248-47843cf74ea3\") " pod="openstack/swift-proxy-864fbf8dcf-chrjx" Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.379031 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d678be18-d610-4ea8-b248-47843cf74ea3-etc-swift\") pod \"swift-proxy-864fbf8dcf-chrjx\" (UID: \"d678be18-d610-4ea8-b248-47843cf74ea3\") " pod="openstack/swift-proxy-864fbf8dcf-chrjx" Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.383149 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvfmp\" (UniqueName: \"kubernetes.io/projected/d678be18-d610-4ea8-b248-47843cf74ea3-kube-api-access-tvfmp\") pod \"swift-proxy-864fbf8dcf-chrjx\" (UID: \"d678be18-d610-4ea8-b248-47843cf74ea3\") " pod="openstack/swift-proxy-864fbf8dcf-chrjx" Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.416324 4556 generic.go:334] "Generic (PLEG): container finished" podID="0a557635-d67d-4e93-a82c-990c34fe7d48" containerID="920ec97ab27b16c8945b61d64a2df962837b1fd7a6881c99c07a5be5234c0d07" exitCode=0 Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.416825 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rmt7n" event={"ID":"0a557635-d67d-4e93-a82c-990c34fe7d48","Type":"ContainerDied","Data":"920ec97ab27b16c8945b61d64a2df962837b1fd7a6881c99c07a5be5234c0d07"} Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.416912 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rmt7n" event={"ID":"0a557635-d67d-4e93-a82c-990c34fe7d48","Type":"ContainerStarted","Data":"7a953807970cdb89372bde18f3bdab42a137614c5d4ce0ea1c1d2f7f0b718a42"} Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.416979 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.422562 4556 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="49e26f9f-38d3-4b82-8ef7-0bf33efbbb25" podUID="97268f30-e67a-441f-93cd-0dc9771dbd42" Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.427433 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.463257 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/97268f30-e67a-441f-93cd-0dc9771dbd42-openstack-config-secret\") pod \"openstackclient\" (UID: \"97268f30-e67a-441f-93cd-0dc9771dbd42\") " pod="openstack/openstackclient" Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.463375 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97268f30-e67a-441f-93cd-0dc9771dbd42-combined-ca-bundle\") pod \"openstackclient\" (UID: \"97268f30-e67a-441f-93cd-0dc9771dbd42\") " pod="openstack/openstackclient" Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.463488 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/97268f30-e67a-441f-93cd-0dc9771dbd42-openstack-config\") pod \"openstackclient\" (UID: \"97268f30-e67a-441f-93cd-0dc9771dbd42\") " pod="openstack/openstackclient" Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.463582 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2tnl\" (UniqueName: \"kubernetes.io/projected/97268f30-e67a-441f-93cd-0dc9771dbd42-kube-api-access-n2tnl\") pod \"openstackclient\" (UID: \"97268f30-e67a-441f-93cd-0dc9771dbd42\") " pod="openstack/openstackclient" Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.564561 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/49e26f9f-38d3-4b82-8ef7-0bf33efbbb25-openstack-config\") pod \"49e26f9f-38d3-4b82-8ef7-0bf33efbbb25\" (UID: \"49e26f9f-38d3-4b82-8ef7-0bf33efbbb25\") " Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.564703 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/49e26f9f-38d3-4b82-8ef7-0bf33efbbb25-openstack-config-secret\") pod \"49e26f9f-38d3-4b82-8ef7-0bf33efbbb25\" (UID: \"49e26f9f-38d3-4b82-8ef7-0bf33efbbb25\") " Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.564801 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49e26f9f-38d3-4b82-8ef7-0bf33efbbb25-combined-ca-bundle\") pod \"49e26f9f-38d3-4b82-8ef7-0bf33efbbb25\" (UID: \"49e26f9f-38d3-4b82-8ef7-0bf33efbbb25\") " Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.564826 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lh22g\" (UniqueName: \"kubernetes.io/projected/49e26f9f-38d3-4b82-8ef7-0bf33efbbb25-kube-api-access-lh22g\") pod \"49e26f9f-38d3-4b82-8ef7-0bf33efbbb25\" (UID: \"49e26f9f-38d3-4b82-8ef7-0bf33efbbb25\") " Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.565129 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/97268f30-e67a-441f-93cd-0dc9771dbd42-openstack-config\") pod \"openstackclient\" (UID: \"97268f30-e67a-441f-93cd-0dc9771dbd42\") " pod="openstack/openstackclient" Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.565237 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2tnl\" (UniqueName: \"kubernetes.io/projected/97268f30-e67a-441f-93cd-0dc9771dbd42-kube-api-access-n2tnl\") pod \"openstackclient\" (UID: \"97268f30-e67a-441f-93cd-0dc9771dbd42\") " pod="openstack/openstackclient" Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.565320 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/97268f30-e67a-441f-93cd-0dc9771dbd42-openstack-config-secret\") pod \"openstackclient\" (UID: \"97268f30-e67a-441f-93cd-0dc9771dbd42\") " pod="openstack/openstackclient" Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.565428 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97268f30-e67a-441f-93cd-0dc9771dbd42-combined-ca-bundle\") pod \"openstackclient\" (UID: \"97268f30-e67a-441f-93cd-0dc9771dbd42\") " pod="openstack/openstackclient" Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.566418 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49e26f9f-38d3-4b82-8ef7-0bf33efbbb25-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "49e26f9f-38d3-4b82-8ef7-0bf33efbbb25" (UID: "49e26f9f-38d3-4b82-8ef7-0bf33efbbb25"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.567515 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/97268f30-e67a-441f-93cd-0dc9771dbd42-openstack-config\") pod \"openstackclient\" (UID: \"97268f30-e67a-441f-93cd-0dc9771dbd42\") " pod="openstack/openstackclient" Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.568255 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-864fbf8dcf-chrjx" Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.571813 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97268f30-e67a-441f-93cd-0dc9771dbd42-combined-ca-bundle\") pod \"openstackclient\" (UID: \"97268f30-e67a-441f-93cd-0dc9771dbd42\") " pod="openstack/openstackclient" Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.573269 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49e26f9f-38d3-4b82-8ef7-0bf33efbbb25-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "49e26f9f-38d3-4b82-8ef7-0bf33efbbb25" (UID: "49e26f9f-38d3-4b82-8ef7-0bf33efbbb25"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.573385 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49e26f9f-38d3-4b82-8ef7-0bf33efbbb25-kube-api-access-lh22g" (OuterVolumeSpecName: "kube-api-access-lh22g") pod "49e26f9f-38d3-4b82-8ef7-0bf33efbbb25" (UID: "49e26f9f-38d3-4b82-8ef7-0bf33efbbb25"). InnerVolumeSpecName "kube-api-access-lh22g". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.582899 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2tnl\" (UniqueName: \"kubernetes.io/projected/97268f30-e67a-441f-93cd-0dc9771dbd42-kube-api-access-n2tnl\") pod \"openstackclient\" (UID: \"97268f30-e67a-441f-93cd-0dc9771dbd42\") " pod="openstack/openstackclient" Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.586408 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49e26f9f-38d3-4b82-8ef7-0bf33efbbb25-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "49e26f9f-38d3-4b82-8ef7-0bf33efbbb25" (UID: "49e26f9f-38d3-4b82-8ef7-0bf33efbbb25"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.590562 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/97268f30-e67a-441f-93cd-0dc9771dbd42-openstack-config-secret\") pod \"openstackclient\" (UID: \"97268f30-e67a-441f-93cd-0dc9771dbd42\") " pod="openstack/openstackclient" Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.649077 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.667838 4556 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/49e26f9f-38d3-4b82-8ef7-0bf33efbbb25-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.667879 4556 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49e26f9f-38d3-4b82-8ef7-0bf33efbbb25-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.667890 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lh22g\" (UniqueName: \"kubernetes.io/projected/49e26f9f-38d3-4b82-8ef7-0bf33efbbb25-kube-api-access-lh22g\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:52 crc kubenswrapper[4556]: I0218 09:18:52.667898 4556 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/49e26f9f-38d3-4b82-8ef7-0bf33efbbb25-openstack-config\") on node \"crc\" DevicePath \"\"" Feb 18 09:18:53 crc kubenswrapper[4556]: I0218 09:18:53.106043 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-864fbf8dcf-chrjx"] Feb 18 09:18:53 crc kubenswrapper[4556]: I0218 09:18:53.236489 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Feb 18 09:18:53 crc kubenswrapper[4556]: I0218 09:18:53.290671 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49e26f9f-38d3-4b82-8ef7-0bf33efbbb25" path="/var/lib/kubelet/pods/49e26f9f-38d3-4b82-8ef7-0bf33efbbb25/volumes" Feb 18 09:18:53 crc kubenswrapper[4556]: I0218 09:18:53.429021 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rmt7n" event={"ID":"0a557635-d67d-4e93-a82c-990c34fe7d48","Type":"ContainerStarted","Data":"aa6c9b508b068a81e21c5d33efe4beaa37c6cc99d038679316b57fc58dddfbe3"} Feb 18 09:18:53 crc kubenswrapper[4556]: I0218 09:18:53.430554 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"97268f30-e67a-441f-93cd-0dc9771dbd42","Type":"ContainerStarted","Data":"2508cd3c9f0c8de39e3fab0c035e441d1e20a94c0eed283769d2100b6fd4f96d"} Feb 18 09:18:53 crc kubenswrapper[4556]: I0218 09:18:53.433141 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 18 09:18:53 crc kubenswrapper[4556]: I0218 09:18:53.433241 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-864fbf8dcf-chrjx" event={"ID":"d678be18-d610-4ea8-b248-47843cf74ea3","Type":"ContainerStarted","Data":"5d2fbdd7bda6443f99c723784fdaaa383e68c3b9d4c95589fd442ae5eabe6818"} Feb 18 09:18:53 crc kubenswrapper[4556]: I0218 09:18:53.451730 4556 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="49e26f9f-38d3-4b82-8ef7-0bf33efbbb25" podUID="97268f30-e67a-441f-93cd-0dc9771dbd42" Feb 18 09:18:53 crc kubenswrapper[4556]: I0218 09:18:53.646227 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-9855c4778-9n9f2" Feb 18 09:18:53 crc kubenswrapper[4556]: I0218 09:18:53.700618 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-9855c4778-9n9f2" Feb 18 09:18:54 crc kubenswrapper[4556]: I0218 09:18:54.070754 4556 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-85bb8c66f4-fhblh" podUID="3354b930-6b00-4aa7-a4bd-97ec410cc863" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Feb 18 09:18:54 crc kubenswrapper[4556]: I0218 09:18:54.443665 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-864fbf8dcf-chrjx" event={"ID":"d678be18-d610-4ea8-b248-47843cf74ea3","Type":"ContainerStarted","Data":"fa24a7c545815af9a4716318baea11c6d042c9c5349e84865ea690d4673606a3"} Feb 18 09:18:54 crc kubenswrapper[4556]: I0218 09:18:54.443772 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-864fbf8dcf-chrjx" event={"ID":"d678be18-d610-4ea8-b248-47843cf74ea3","Type":"ContainerStarted","Data":"2f7016bf65b5d32d6d26acbf5286fcabadd506550f63ba0c31cea1cb44210b78"} Feb 18 09:18:54 crc kubenswrapper[4556]: I0218 09:18:54.466334 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-864fbf8dcf-chrjx" podStartSLOduration=2.466317031 podStartE2EDuration="2.466317031s" podCreationTimestamp="2026-02-18 09:18:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:18:54.461454887 +0000 UTC m=+891.478415867" watchObservedRunningTime="2026-02-18 09:18:54.466317031 +0000 UTC m=+891.483278012" Feb 18 09:18:54 crc kubenswrapper[4556]: I0218 09:18:54.762324 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Feb 18 09:18:55 crc kubenswrapper[4556]: I0218 09:18:55.451401 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-864fbf8dcf-chrjx" Feb 18 09:18:55 crc kubenswrapper[4556]: I0218 09:18:55.451732 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-864fbf8dcf-chrjx" Feb 18 09:18:56 crc kubenswrapper[4556]: I0218 09:18:56.280212 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-nghmb"] Feb 18 09:18:56 crc kubenswrapper[4556]: I0218 09:18:56.282351 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nghmb" Feb 18 09:18:56 crc kubenswrapper[4556]: I0218 09:18:56.291532 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nghmb"] Feb 18 09:18:56 crc kubenswrapper[4556]: I0218 09:18:56.458869 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a0954e9-38c5-435c-9ced-d8e910e284f9-catalog-content\") pod \"redhat-marketplace-nghmb\" (UID: \"3a0954e9-38c5-435c-9ced-d8e910e284f9\") " pod="openshift-marketplace/redhat-marketplace-nghmb" Feb 18 09:18:56 crc kubenswrapper[4556]: I0218 09:18:56.458942 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpd8w\" (UniqueName: \"kubernetes.io/projected/3a0954e9-38c5-435c-9ced-d8e910e284f9-kube-api-access-qpd8w\") pod \"redhat-marketplace-nghmb\" (UID: \"3a0954e9-38c5-435c-9ced-d8e910e284f9\") " pod="openshift-marketplace/redhat-marketplace-nghmb" Feb 18 09:18:56 crc kubenswrapper[4556]: I0218 09:18:56.459003 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a0954e9-38c5-435c-9ced-d8e910e284f9-utilities\") pod \"redhat-marketplace-nghmb\" (UID: \"3a0954e9-38c5-435c-9ced-d8e910e284f9\") " pod="openshift-marketplace/redhat-marketplace-nghmb" Feb 18 09:18:56 crc kubenswrapper[4556]: I0218 09:18:56.465388 4556 generic.go:334] "Generic (PLEG): container finished" podID="0a557635-d67d-4e93-a82c-990c34fe7d48" containerID="aa6c9b508b068a81e21c5d33efe4beaa37c6cc99d038679316b57fc58dddfbe3" exitCode=0 Feb 18 09:18:56 crc kubenswrapper[4556]: I0218 09:18:56.465501 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rmt7n" event={"ID":"0a557635-d67d-4e93-a82c-990c34fe7d48","Type":"ContainerDied","Data":"aa6c9b508b068a81e21c5d33efe4beaa37c6cc99d038679316b57fc58dddfbe3"} Feb 18 09:18:56 crc kubenswrapper[4556]: I0218 09:18:56.561762 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a0954e9-38c5-435c-9ced-d8e910e284f9-catalog-content\") pod \"redhat-marketplace-nghmb\" (UID: \"3a0954e9-38c5-435c-9ced-d8e910e284f9\") " pod="openshift-marketplace/redhat-marketplace-nghmb" Feb 18 09:18:56 crc kubenswrapper[4556]: I0218 09:18:56.561848 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpd8w\" (UniqueName: \"kubernetes.io/projected/3a0954e9-38c5-435c-9ced-d8e910e284f9-kube-api-access-qpd8w\") pod \"redhat-marketplace-nghmb\" (UID: \"3a0954e9-38c5-435c-9ced-d8e910e284f9\") " pod="openshift-marketplace/redhat-marketplace-nghmb" Feb 18 09:18:56 crc kubenswrapper[4556]: I0218 09:18:56.561948 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a0954e9-38c5-435c-9ced-d8e910e284f9-utilities\") pod \"redhat-marketplace-nghmb\" (UID: \"3a0954e9-38c5-435c-9ced-d8e910e284f9\") " pod="openshift-marketplace/redhat-marketplace-nghmb" Feb 18 09:18:56 crc kubenswrapper[4556]: I0218 09:18:56.562360 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a0954e9-38c5-435c-9ced-d8e910e284f9-utilities\") pod \"redhat-marketplace-nghmb\" (UID: \"3a0954e9-38c5-435c-9ced-d8e910e284f9\") " pod="openshift-marketplace/redhat-marketplace-nghmb" Feb 18 09:18:56 crc kubenswrapper[4556]: I0218 09:18:56.562435 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a0954e9-38c5-435c-9ced-d8e910e284f9-catalog-content\") pod \"redhat-marketplace-nghmb\" (UID: \"3a0954e9-38c5-435c-9ced-d8e910e284f9\") " pod="openshift-marketplace/redhat-marketplace-nghmb" Feb 18 09:18:56 crc kubenswrapper[4556]: I0218 09:18:56.595261 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpd8w\" (UniqueName: \"kubernetes.io/projected/3a0954e9-38c5-435c-9ced-d8e910e284f9-kube-api-access-qpd8w\") pod \"redhat-marketplace-nghmb\" (UID: \"3a0954e9-38c5-435c-9ced-d8e910e284f9\") " pod="openshift-marketplace/redhat-marketplace-nghmb" Feb 18 09:18:56 crc kubenswrapper[4556]: I0218 09:18:56.608101 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nghmb" Feb 18 09:18:57 crc kubenswrapper[4556]: I0218 09:18:57.129210 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nghmb"] Feb 18 09:18:57 crc kubenswrapper[4556]: I0218 09:18:57.478560 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rmt7n" event={"ID":"0a557635-d67d-4e93-a82c-990c34fe7d48","Type":"ContainerStarted","Data":"e9d99eae194d31abfd1573901538558b29b3ebf6244cb332929af9b0130f99f8"} Feb 18 09:18:57 crc kubenswrapper[4556]: I0218 09:18:57.485941 4556 generic.go:334] "Generic (PLEG): container finished" podID="3a0954e9-38c5-435c-9ced-d8e910e284f9" containerID="75e22c711b14ccf5ce6a0acd5950f61dc69f87d5df160be30e872749562bc6f8" exitCode=0 Feb 18 09:18:57 crc kubenswrapper[4556]: I0218 09:18:57.486025 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nghmb" event={"ID":"3a0954e9-38c5-435c-9ced-d8e910e284f9","Type":"ContainerDied","Data":"75e22c711b14ccf5ce6a0acd5950f61dc69f87d5df160be30e872749562bc6f8"} Feb 18 09:18:57 crc kubenswrapper[4556]: I0218 09:18:57.486111 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nghmb" event={"ID":"3a0954e9-38c5-435c-9ced-d8e910e284f9","Type":"ContainerStarted","Data":"7ee7f916a52e0517e42b13a26438a3541d4ef44c10d559bfff948a0f9aee1ced"} Feb 18 09:18:57 crc kubenswrapper[4556]: I0218 09:18:57.498598 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rmt7n" podStartSLOduration=1.7927868679999999 podStartE2EDuration="6.498579838s" podCreationTimestamp="2026-02-18 09:18:51 +0000 UTC" firstStartedPulling="2026-02-18 09:18:52.418701149 +0000 UTC m=+889.435662130" lastFinishedPulling="2026-02-18 09:18:57.12449412 +0000 UTC m=+894.141455100" observedRunningTime="2026-02-18 09:18:57.496328942 +0000 UTC m=+894.513289921" watchObservedRunningTime="2026-02-18 09:18:57.498579838 +0000 UTC m=+894.515540818" Feb 18 09:18:58 crc kubenswrapper[4556]: I0218 09:18:58.502093 4556 generic.go:334] "Generic (PLEG): container finished" podID="3a0954e9-38c5-435c-9ced-d8e910e284f9" containerID="fad04e69ea4a5850f128f2634e47f2422613db7a840386f5fd27819d8bcb0ff6" exitCode=0 Feb 18 09:18:58 crc kubenswrapper[4556]: I0218 09:18:58.502483 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nghmb" event={"ID":"3a0954e9-38c5-435c-9ced-d8e910e284f9","Type":"ContainerDied","Data":"fad04e69ea4a5850f128f2634e47f2422613db7a840386f5fd27819d8bcb0ff6"} Feb 18 09:18:59 crc kubenswrapper[4556]: I0218 09:18:59.523998 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nghmb" event={"ID":"3a0954e9-38c5-435c-9ced-d8e910e284f9","Type":"ContainerStarted","Data":"307bdbb1384330f314b33f4e5a051ba4a57c6383a5b0f9f4720db0713087c297"} Feb 18 09:19:00 crc kubenswrapper[4556]: I0218 09:19:00.000388 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Feb 18 09:19:00 crc kubenswrapper[4556]: I0218 09:19:00.021261 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-nghmb" podStartSLOduration=2.524276813 podStartE2EDuration="4.021244155s" podCreationTimestamp="2026-02-18 09:18:56 +0000 UTC" firstStartedPulling="2026-02-18 09:18:57.489085393 +0000 UTC m=+894.506046372" lastFinishedPulling="2026-02-18 09:18:58.986052733 +0000 UTC m=+896.003013714" observedRunningTime="2026-02-18 09:18:59.542843515 +0000 UTC m=+896.559804496" watchObservedRunningTime="2026-02-18 09:19:00.021244155 +0000 UTC m=+897.038205135" Feb 18 09:19:01 crc kubenswrapper[4556]: I0218 09:19:01.401689 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rmt7n" Feb 18 09:19:01 crc kubenswrapper[4556]: I0218 09:19:01.402065 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rmt7n" Feb 18 09:19:01 crc kubenswrapper[4556]: I0218 09:19:01.727534 4556 patch_prober.go:28] interesting pod/machine-config-daemon-f76hs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 09:19:01 crc kubenswrapper[4556]: I0218 09:19:01.727592 4556 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 09:19:02 crc kubenswrapper[4556]: I0218 09:19:02.273531 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-54664bdbc-cxnvs" Feb 18 09:19:02 crc kubenswrapper[4556]: I0218 09:19:02.325898 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-d8cc75c64-fwzfq"] Feb 18 09:19:02 crc kubenswrapper[4556]: I0218 09:19:02.326245 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-d8cc75c64-fwzfq" podUID="2bb44e10-7532-4173-a140-140b161601a4" containerName="neutron-api" containerID="cri-o://0d354e35ff2cfc3bc438333dde43e23a4e77b08fcb22a13609433bf275ed1538" gracePeriod=30 Feb 18 09:19:02 crc kubenswrapper[4556]: I0218 09:19:02.326450 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-d8cc75c64-fwzfq" podUID="2bb44e10-7532-4173-a140-140b161601a4" containerName="neutron-httpd" containerID="cri-o://4975d3e5e950a678a1bad9398685c9357404c16df42f27b08a96ae18291c4230" gracePeriod=30 Feb 18 09:19:02 crc kubenswrapper[4556]: I0218 09:19:02.443968 4556 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rmt7n" podUID="0a557635-d67d-4e93-a82c-990c34fe7d48" containerName="registry-server" probeResult="failure" output=< Feb 18 09:19:02 crc kubenswrapper[4556]: timeout: failed to connect service ":50051" within 1s Feb 18 09:19:02 crc kubenswrapper[4556]: > Feb 18 09:19:02 crc kubenswrapper[4556]: I0218 09:19:02.555263 4556 generic.go:334] "Generic (PLEG): container finished" podID="2bb44e10-7532-4173-a140-140b161601a4" containerID="4975d3e5e950a678a1bad9398685c9357404c16df42f27b08a96ae18291c4230" exitCode=0 Feb 18 09:19:02 crc kubenswrapper[4556]: I0218 09:19:02.555330 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d8cc75c64-fwzfq" event={"ID":"2bb44e10-7532-4173-a140-140b161601a4","Type":"ContainerDied","Data":"4975d3e5e950a678a1bad9398685c9357404c16df42f27b08a96ae18291c4230"} Feb 18 09:19:02 crc kubenswrapper[4556]: I0218 09:19:02.573660 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-864fbf8dcf-chrjx" Feb 18 09:19:02 crc kubenswrapper[4556]: I0218 09:19:02.574607 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-864fbf8dcf-chrjx" Feb 18 09:19:04 crc kubenswrapper[4556]: I0218 09:19:04.070750 4556 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-85bb8c66f4-fhblh" podUID="3354b930-6b00-4aa7-a4bd-97ec410cc863" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Feb 18 09:19:04 crc kubenswrapper[4556]: I0218 09:19:04.071354 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-85bb8c66f4-fhblh" Feb 18 09:19:04 crc kubenswrapper[4556]: I0218 09:19:04.598881 4556 generic.go:334] "Generic (PLEG): container finished" podID="2bb44e10-7532-4173-a140-140b161601a4" containerID="0d354e35ff2cfc3bc438333dde43e23a4e77b08fcb22a13609433bf275ed1538" exitCode=0 Feb 18 09:19:04 crc kubenswrapper[4556]: I0218 09:19:04.598939 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d8cc75c64-fwzfq" event={"ID":"2bb44e10-7532-4173-a140-140b161601a4","Type":"ContainerDied","Data":"0d354e35ff2cfc3bc438333dde43e23a4e77b08fcb22a13609433bf275ed1538"} Feb 18 09:19:05 crc kubenswrapper[4556]: I0218 09:19:05.612104 4556 generic.go:334] "Generic (PLEG): container finished" podID="e0739f72-891b-46e1-bce5-e17c54a407ab" containerID="c7b35b462aab32a2bd6aa0f4634c9496147ec2c072365494e87a9b8128d8030b" exitCode=137 Feb 18 09:19:05 crc kubenswrapper[4556]: I0218 09:19:05.612164 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e0739f72-891b-46e1-bce5-e17c54a407ab","Type":"ContainerDied","Data":"c7b35b462aab32a2bd6aa0f4634c9496147ec2c072365494e87a9b8128d8030b"} Feb 18 09:19:06 crc kubenswrapper[4556]: I0218 09:19:06.608771 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-nghmb" Feb 18 09:19:06 crc kubenswrapper[4556]: I0218 09:19:06.609177 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-nghmb" Feb 18 09:19:06 crc kubenswrapper[4556]: I0218 09:19:06.658593 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-nghmb" Feb 18 09:19:06 crc kubenswrapper[4556]: I0218 09:19:06.700172 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-nghmb" Feb 18 09:19:06 crc kubenswrapper[4556]: I0218 09:19:06.898505 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nghmb"] Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.259759 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.360882 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d8cc75c64-fwzfq" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.457419 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0739f72-891b-46e1-bce5-e17c54a407ab-config-data\") pod \"e0739f72-891b-46e1-bce5-e17c54a407ab\" (UID: \"e0739f72-891b-46e1-bce5-e17c54a407ab\") " Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.457510 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0739f72-891b-46e1-bce5-e17c54a407ab-scripts\") pod \"e0739f72-891b-46e1-bce5-e17c54a407ab\" (UID: \"e0739f72-891b-46e1-bce5-e17c54a407ab\") " Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.457556 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e0739f72-891b-46e1-bce5-e17c54a407ab-run-httpd\") pod \"e0739f72-891b-46e1-bce5-e17c54a407ab\" (UID: \"e0739f72-891b-46e1-bce5-e17c54a407ab\") " Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.457632 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e0739f72-891b-46e1-bce5-e17c54a407ab-log-httpd\") pod \"e0739f72-891b-46e1-bce5-e17c54a407ab\" (UID: \"e0739f72-891b-46e1-bce5-e17c54a407ab\") " Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.457656 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c985q\" (UniqueName: \"kubernetes.io/projected/e0739f72-891b-46e1-bce5-e17c54a407ab-kube-api-access-c985q\") pod \"e0739f72-891b-46e1-bce5-e17c54a407ab\" (UID: \"e0739f72-891b-46e1-bce5-e17c54a407ab\") " Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.457842 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0739f72-891b-46e1-bce5-e17c54a407ab-combined-ca-bundle\") pod \"e0739f72-891b-46e1-bce5-e17c54a407ab\" (UID: \"e0739f72-891b-46e1-bce5-e17c54a407ab\") " Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.457924 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e0739f72-891b-46e1-bce5-e17c54a407ab-sg-core-conf-yaml\") pod \"e0739f72-891b-46e1-bce5-e17c54a407ab\" (UID: \"e0739f72-891b-46e1-bce5-e17c54a407ab\") " Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.458801 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0739f72-891b-46e1-bce5-e17c54a407ab-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "e0739f72-891b-46e1-bce5-e17c54a407ab" (UID: "e0739f72-891b-46e1-bce5-e17c54a407ab"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.459656 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0739f72-891b-46e1-bce5-e17c54a407ab-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "e0739f72-891b-46e1-bce5-e17c54a407ab" (UID: "e0739f72-891b-46e1-bce5-e17c54a407ab"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.463517 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0739f72-891b-46e1-bce5-e17c54a407ab-kube-api-access-c985q" (OuterVolumeSpecName: "kube-api-access-c985q") pod "e0739f72-891b-46e1-bce5-e17c54a407ab" (UID: "e0739f72-891b-46e1-bce5-e17c54a407ab"). InnerVolumeSpecName "kube-api-access-c985q". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.464770 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0739f72-891b-46e1-bce5-e17c54a407ab-scripts" (OuterVolumeSpecName: "scripts") pod "e0739f72-891b-46e1-bce5-e17c54a407ab" (UID: "e0739f72-891b-46e1-bce5-e17c54a407ab"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.503298 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0739f72-891b-46e1-bce5-e17c54a407ab-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "e0739f72-891b-46e1-bce5-e17c54a407ab" (UID: "e0739f72-891b-46e1-bce5-e17c54a407ab"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.510669 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.557980 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0739f72-891b-46e1-bce5-e17c54a407ab-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e0739f72-891b-46e1-bce5-e17c54a407ab" (UID: "e0739f72-891b-46e1-bce5-e17c54a407ab"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.564238 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2bb44e10-7532-4173-a140-140b161601a4-config\") pod \"2bb44e10-7532-4173-a140-140b161601a4\" (UID: \"2bb44e10-7532-4173-a140-140b161601a4\") " Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.564279 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6c8xm\" (UniqueName: \"kubernetes.io/projected/2bb44e10-7532-4173-a140-140b161601a4-kube-api-access-6c8xm\") pod \"2bb44e10-7532-4173-a140-140b161601a4\" (UID: \"2bb44e10-7532-4173-a140-140b161601a4\") " Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.564312 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/2bb44e10-7532-4173-a140-140b161601a4-httpd-config\") pod \"2bb44e10-7532-4173-a140-140b161601a4\" (UID: \"2bb44e10-7532-4173-a140-140b161601a4\") " Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.564384 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bb44e10-7532-4173-a140-140b161601a4-combined-ca-bundle\") pod \"2bb44e10-7532-4173-a140-140b161601a4\" (UID: \"2bb44e10-7532-4173-a140-140b161601a4\") " Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.564817 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2bb44e10-7532-4173-a140-140b161601a4-ovndb-tls-certs\") pod \"2bb44e10-7532-4173-a140-140b161601a4\" (UID: \"2bb44e10-7532-4173-a140-140b161601a4\") " Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.565584 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e15b6829-420b-493f-a31c-3bde1c53035f-scripts\") pod \"e15b6829-420b-493f-a31c-3bde1c53035f\" (UID: \"e15b6829-420b-493f-a31c-3bde1c53035f\") " Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.566224 4556 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0739f72-891b-46e1-bce5-e17c54a407ab-scripts\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.566261 4556 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e0739f72-891b-46e1-bce5-e17c54a407ab-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.566270 4556 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e0739f72-891b-46e1-bce5-e17c54a407ab-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.566278 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c985q\" (UniqueName: \"kubernetes.io/projected/e0739f72-891b-46e1-bce5-e17c54a407ab-kube-api-access-c985q\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.566287 4556 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0739f72-891b-46e1-bce5-e17c54a407ab-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.566295 4556 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e0739f72-891b-46e1-bce5-e17c54a407ab-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.569318 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e15b6829-420b-493f-a31c-3bde1c53035f-scripts" (OuterVolumeSpecName: "scripts") pod "e15b6829-420b-493f-a31c-3bde1c53035f" (UID: "e15b6829-420b-493f-a31c-3bde1c53035f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.569948 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bb44e10-7532-4173-a140-140b161601a4-kube-api-access-6c8xm" (OuterVolumeSpecName: "kube-api-access-6c8xm") pod "2bb44e10-7532-4173-a140-140b161601a4" (UID: "2bb44e10-7532-4173-a140-140b161601a4"). InnerVolumeSpecName "kube-api-access-6c8xm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.572246 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bb44e10-7532-4173-a140-140b161601a4-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "2bb44e10-7532-4173-a140-140b161601a4" (UID: "2bb44e10-7532-4173-a140-140b161601a4"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.591317 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0739f72-891b-46e1-bce5-e17c54a407ab-config-data" (OuterVolumeSpecName: "config-data") pod "e0739f72-891b-46e1-bce5-e17c54a407ab" (UID: "e0739f72-891b-46e1-bce5-e17c54a407ab"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.615413 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bb44e10-7532-4173-a140-140b161601a4-config" (OuterVolumeSpecName: "config") pod "2bb44e10-7532-4173-a140-140b161601a4" (UID: "2bb44e10-7532-4173-a140-140b161601a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.618973 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bb44e10-7532-4173-a140-140b161601a4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2bb44e10-7532-4173-a140-140b161601a4" (UID: "2bb44e10-7532-4173-a140-140b161601a4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.634203 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bb44e10-7532-4173-a140-140b161601a4-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "2bb44e10-7532-4173-a140-140b161601a4" (UID: "2bb44e10-7532-4173-a140-140b161601a4"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.636389 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d8cc75c64-fwzfq" event={"ID":"2bb44e10-7532-4173-a140-140b161601a4","Type":"ContainerDied","Data":"f097d1897907d95895e9c90beb30d488dd020fb1b4fe76bc9601482a0968b723"} Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.636440 4556 scope.go:117] "RemoveContainer" containerID="4975d3e5e950a678a1bad9398685c9357404c16df42f27b08a96ae18291c4230" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.636646 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d8cc75c64-fwzfq" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.640388 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"97268f30-e67a-441f-93cd-0dc9771dbd42","Type":"ContainerStarted","Data":"5b1c4779b9d41d04539fc5bb906d851de888a772de0e43dfffa47e3080d0550b"} Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.656253 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e0739f72-891b-46e1-bce5-e17c54a407ab","Type":"ContainerDied","Data":"f429b65a29cc2a73cc0988982cefba8e5a956982006946c9fcfc145106a29642"} Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.656562 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.659244 4556 scope.go:117] "RemoveContainer" containerID="0d354e35ff2cfc3bc438333dde43e23a4e77b08fcb22a13609433bf275ed1538" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.664551 4556 generic.go:334] "Generic (PLEG): container finished" podID="e15b6829-420b-493f-a31c-3bde1c53035f" containerID="9d80e4b26b777401843d9382d2ebfd2f637f602a3747547b652ad119072af823" exitCode=137 Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.665025 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.665185 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.026805145 podStartE2EDuration="15.665170289s" podCreationTimestamp="2026-02-18 09:18:52 +0000 UTC" firstStartedPulling="2026-02-18 09:18:53.253396276 +0000 UTC m=+890.270357256" lastFinishedPulling="2026-02-18 09:19:06.89176142 +0000 UTC m=+903.908722400" observedRunningTime="2026-02-18 09:19:07.661008376 +0000 UTC m=+904.677969356" watchObservedRunningTime="2026-02-18 09:19:07.665170289 +0000 UTC m=+904.682131269" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.665215 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e15b6829-420b-493f-a31c-3bde1c53035f","Type":"ContainerDied","Data":"9d80e4b26b777401843d9382d2ebfd2f637f602a3747547b652ad119072af823"} Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.665236 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e15b6829-420b-493f-a31c-3bde1c53035f","Type":"ContainerDied","Data":"b6353f56ef8b6ec4206c8a95be662cb4da864e7e58416f7d69100aa872d0c181"} Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.667259 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e15b6829-420b-493f-a31c-3bde1c53035f-config-data-custom\") pod \"e15b6829-420b-493f-a31c-3bde1c53035f\" (UID: \"e15b6829-420b-493f-a31c-3bde1c53035f\") " Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.667758 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qrswm\" (UniqueName: \"kubernetes.io/projected/e15b6829-420b-493f-a31c-3bde1c53035f-kube-api-access-qrswm\") pod \"e15b6829-420b-493f-a31c-3bde1c53035f\" (UID: \"e15b6829-420b-493f-a31c-3bde1c53035f\") " Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.667847 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e15b6829-420b-493f-a31c-3bde1c53035f-etc-machine-id\") pod \"e15b6829-420b-493f-a31c-3bde1c53035f\" (UID: \"e15b6829-420b-493f-a31c-3bde1c53035f\") " Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.667878 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e15b6829-420b-493f-a31c-3bde1c53035f-combined-ca-bundle\") pod \"e15b6829-420b-493f-a31c-3bde1c53035f\" (UID: \"e15b6829-420b-493f-a31c-3bde1c53035f\") " Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.667932 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e15b6829-420b-493f-a31c-3bde1c53035f-config-data\") pod \"e15b6829-420b-493f-a31c-3bde1c53035f\" (UID: \"e15b6829-420b-493f-a31c-3bde1c53035f\") " Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.667964 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e15b6829-420b-493f-a31c-3bde1c53035f-logs\") pod \"e15b6829-420b-493f-a31c-3bde1c53035f\" (UID: \"e15b6829-420b-493f-a31c-3bde1c53035f\") " Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.668337 4556 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2bb44e10-7532-4173-a140-140b161601a4-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.668353 4556 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e15b6829-420b-493f-a31c-3bde1c53035f-scripts\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.668362 4556 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0739f72-891b-46e1-bce5-e17c54a407ab-config-data\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.668371 4556 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/2bb44e10-7532-4173-a140-140b161601a4-config\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.668379 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6c8xm\" (UniqueName: \"kubernetes.io/projected/2bb44e10-7532-4173-a140-140b161601a4-kube-api-access-6c8xm\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.668388 4556 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/2bb44e10-7532-4173-a140-140b161601a4-httpd-config\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.668396 4556 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bb44e10-7532-4173-a140-140b161601a4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.669145 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e15b6829-420b-493f-a31c-3bde1c53035f-logs" (OuterVolumeSpecName: "logs") pod "e15b6829-420b-493f-a31c-3bde1c53035f" (UID: "e15b6829-420b-493f-a31c-3bde1c53035f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.669216 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e15b6829-420b-493f-a31c-3bde1c53035f-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "e15b6829-420b-493f-a31c-3bde1c53035f" (UID: "e15b6829-420b-493f-a31c-3bde1c53035f"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.672509 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e15b6829-420b-493f-a31c-3bde1c53035f-kube-api-access-qrswm" (OuterVolumeSpecName: "kube-api-access-qrswm") pod "e15b6829-420b-493f-a31c-3bde1c53035f" (UID: "e15b6829-420b-493f-a31c-3bde1c53035f"). InnerVolumeSpecName "kube-api-access-qrswm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.679668 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e15b6829-420b-493f-a31c-3bde1c53035f-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "e15b6829-420b-493f-a31c-3bde1c53035f" (UID: "e15b6829-420b-493f-a31c-3bde1c53035f"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.683631 4556 scope.go:117] "RemoveContainer" containerID="c7b35b462aab32a2bd6aa0f4634c9496147ec2c072365494e87a9b8128d8030b" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.690361 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-d8cc75c64-fwzfq"] Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.700951 4556 scope.go:117] "RemoveContainer" containerID="f7367e0b2c468791b656052267d5c9e3f713321ffc299c0bae2758bf91f088f7" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.702736 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-d8cc75c64-fwzfq"] Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.707729 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.711620 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e15b6829-420b-493f-a31c-3bde1c53035f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e15b6829-420b-493f-a31c-3bde1c53035f" (UID: "e15b6829-420b-493f-a31c-3bde1c53035f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.712953 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.730600 4556 scope.go:117] "RemoveContainer" containerID="f644379761b7c18c34aaa29cf6ae6f5433a7bb623d8ff1f7d0fd70f9539f4a9f" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.730779 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e15b6829-420b-493f-a31c-3bde1c53035f-config-data" (OuterVolumeSpecName: "config-data") pod "e15b6829-420b-493f-a31c-3bde1c53035f" (UID: "e15b6829-420b-493f-a31c-3bde1c53035f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.734290 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 18 09:19:07 crc kubenswrapper[4556]: E0218 09:19:07.734711 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0739f72-891b-46e1-bce5-e17c54a407ab" containerName="ceilometer-notification-agent" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.734727 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0739f72-891b-46e1-bce5-e17c54a407ab" containerName="ceilometer-notification-agent" Feb 18 09:19:07 crc kubenswrapper[4556]: E0218 09:19:07.734745 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e15b6829-420b-493f-a31c-3bde1c53035f" containerName="cinder-api-log" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.734751 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="e15b6829-420b-493f-a31c-3bde1c53035f" containerName="cinder-api-log" Feb 18 09:19:07 crc kubenswrapper[4556]: E0218 09:19:07.734763 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0739f72-891b-46e1-bce5-e17c54a407ab" containerName="ceilometer-central-agent" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.734770 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0739f72-891b-46e1-bce5-e17c54a407ab" containerName="ceilometer-central-agent" Feb 18 09:19:07 crc kubenswrapper[4556]: E0218 09:19:07.734779 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0739f72-891b-46e1-bce5-e17c54a407ab" containerName="sg-core" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.734785 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0739f72-891b-46e1-bce5-e17c54a407ab" containerName="sg-core" Feb 18 09:19:07 crc kubenswrapper[4556]: E0218 09:19:07.734793 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bb44e10-7532-4173-a140-140b161601a4" containerName="neutron-api" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.734799 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bb44e10-7532-4173-a140-140b161601a4" containerName="neutron-api" Feb 18 09:19:07 crc kubenswrapper[4556]: E0218 09:19:07.734811 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e15b6829-420b-493f-a31c-3bde1c53035f" containerName="cinder-api" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.734816 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="e15b6829-420b-493f-a31c-3bde1c53035f" containerName="cinder-api" Feb 18 09:19:07 crc kubenswrapper[4556]: E0218 09:19:07.734839 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0739f72-891b-46e1-bce5-e17c54a407ab" containerName="proxy-httpd" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.734846 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0739f72-891b-46e1-bce5-e17c54a407ab" containerName="proxy-httpd" Feb 18 09:19:07 crc kubenswrapper[4556]: E0218 09:19:07.734857 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bb44e10-7532-4173-a140-140b161601a4" containerName="neutron-httpd" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.734863 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bb44e10-7532-4173-a140-140b161601a4" containerName="neutron-httpd" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.735049 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0739f72-891b-46e1-bce5-e17c54a407ab" containerName="sg-core" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.735067 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="e15b6829-420b-493f-a31c-3bde1c53035f" containerName="cinder-api" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.735080 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0739f72-891b-46e1-bce5-e17c54a407ab" containerName="ceilometer-notification-agent" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.735092 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bb44e10-7532-4173-a140-140b161601a4" containerName="neutron-api" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.735100 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bb44e10-7532-4173-a140-140b161601a4" containerName="neutron-httpd" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.735110 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="e15b6829-420b-493f-a31c-3bde1c53035f" containerName="cinder-api-log" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.735119 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0739f72-891b-46e1-bce5-e17c54a407ab" containerName="ceilometer-central-agent" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.735128 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0739f72-891b-46e1-bce5-e17c54a407ab" containerName="proxy-httpd" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.736766 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.739470 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.739700 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.744764 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.765436 4556 scope.go:117] "RemoveContainer" containerID="5efdef38fecec015f3f0924c79231a8c7d9c34252e406c92aefa44ef006d639b" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.769748 4556 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e15b6829-420b-493f-a31c-3bde1c53035f-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.769782 4556 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e15b6829-420b-493f-a31c-3bde1c53035f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.769794 4556 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e15b6829-420b-493f-a31c-3bde1c53035f-config-data\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.769807 4556 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e15b6829-420b-493f-a31c-3bde1c53035f-logs\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.769816 4556 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e15b6829-420b-493f-a31c-3bde1c53035f-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.769825 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qrswm\" (UniqueName: \"kubernetes.io/projected/e15b6829-420b-493f-a31c-3bde1c53035f-kube-api-access-qrswm\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.783973 4556 scope.go:117] "RemoveContainer" containerID="9d80e4b26b777401843d9382d2ebfd2f637f602a3747547b652ad119072af823" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.812813 4556 scope.go:117] "RemoveContainer" containerID="d67852ff68b4610cf54e2b45a423c4b0b67eec6e82b39b44a3cf5c7922d54615" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.828429 4556 scope.go:117] "RemoveContainer" containerID="9d80e4b26b777401843d9382d2ebfd2f637f602a3747547b652ad119072af823" Feb 18 09:19:07 crc kubenswrapper[4556]: E0218 09:19:07.829415 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d80e4b26b777401843d9382d2ebfd2f637f602a3747547b652ad119072af823\": container with ID starting with 9d80e4b26b777401843d9382d2ebfd2f637f602a3747547b652ad119072af823 not found: ID does not exist" containerID="9d80e4b26b777401843d9382d2ebfd2f637f602a3747547b652ad119072af823" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.829446 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d80e4b26b777401843d9382d2ebfd2f637f602a3747547b652ad119072af823"} err="failed to get container status \"9d80e4b26b777401843d9382d2ebfd2f637f602a3747547b652ad119072af823\": rpc error: code = NotFound desc = could not find container \"9d80e4b26b777401843d9382d2ebfd2f637f602a3747547b652ad119072af823\": container with ID starting with 9d80e4b26b777401843d9382d2ebfd2f637f602a3747547b652ad119072af823 not found: ID does not exist" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.829470 4556 scope.go:117] "RemoveContainer" containerID="d67852ff68b4610cf54e2b45a423c4b0b67eec6e82b39b44a3cf5c7922d54615" Feb 18 09:19:07 crc kubenswrapper[4556]: E0218 09:19:07.829957 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d67852ff68b4610cf54e2b45a423c4b0b67eec6e82b39b44a3cf5c7922d54615\": container with ID starting with d67852ff68b4610cf54e2b45a423c4b0b67eec6e82b39b44a3cf5c7922d54615 not found: ID does not exist" containerID="d67852ff68b4610cf54e2b45a423c4b0b67eec6e82b39b44a3cf5c7922d54615" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.829981 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d67852ff68b4610cf54e2b45a423c4b0b67eec6e82b39b44a3cf5c7922d54615"} err="failed to get container status \"d67852ff68b4610cf54e2b45a423c4b0b67eec6e82b39b44a3cf5c7922d54615\": rpc error: code = NotFound desc = could not find container \"d67852ff68b4610cf54e2b45a423c4b0b67eec6e82b39b44a3cf5c7922d54615\": container with ID starting with d67852ff68b4610cf54e2b45a423c4b0b67eec6e82b39b44a3cf5c7922d54615 not found: ID does not exist" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.873240 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40b701d5-4495-47c3-bb6e-6ba466958355-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"40b701d5-4495-47c3-bb6e-6ba466958355\") " pod="openstack/ceilometer-0" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.873480 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40b701d5-4495-47c3-bb6e-6ba466958355-config-data\") pod \"ceilometer-0\" (UID: \"40b701d5-4495-47c3-bb6e-6ba466958355\") " pod="openstack/ceilometer-0" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.873537 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/40b701d5-4495-47c3-bb6e-6ba466958355-run-httpd\") pod \"ceilometer-0\" (UID: \"40b701d5-4495-47c3-bb6e-6ba466958355\") " pod="openstack/ceilometer-0" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.873714 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8blb\" (UniqueName: \"kubernetes.io/projected/40b701d5-4495-47c3-bb6e-6ba466958355-kube-api-access-p8blb\") pod \"ceilometer-0\" (UID: \"40b701d5-4495-47c3-bb6e-6ba466958355\") " pod="openstack/ceilometer-0" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.873746 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/40b701d5-4495-47c3-bb6e-6ba466958355-scripts\") pod \"ceilometer-0\" (UID: \"40b701d5-4495-47c3-bb6e-6ba466958355\") " pod="openstack/ceilometer-0" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.873854 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/40b701d5-4495-47c3-bb6e-6ba466958355-log-httpd\") pod \"ceilometer-0\" (UID: \"40b701d5-4495-47c3-bb6e-6ba466958355\") " pod="openstack/ceilometer-0" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.873952 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/40b701d5-4495-47c3-bb6e-6ba466958355-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"40b701d5-4495-47c3-bb6e-6ba466958355\") " pod="openstack/ceilometer-0" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.975080 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40b701d5-4495-47c3-bb6e-6ba466958355-config-data\") pod \"ceilometer-0\" (UID: \"40b701d5-4495-47c3-bb6e-6ba466958355\") " pod="openstack/ceilometer-0" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.975130 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/40b701d5-4495-47c3-bb6e-6ba466958355-run-httpd\") pod \"ceilometer-0\" (UID: \"40b701d5-4495-47c3-bb6e-6ba466958355\") " pod="openstack/ceilometer-0" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.975191 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8blb\" (UniqueName: \"kubernetes.io/projected/40b701d5-4495-47c3-bb6e-6ba466958355-kube-api-access-p8blb\") pod \"ceilometer-0\" (UID: \"40b701d5-4495-47c3-bb6e-6ba466958355\") " pod="openstack/ceilometer-0" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.975214 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/40b701d5-4495-47c3-bb6e-6ba466958355-scripts\") pod \"ceilometer-0\" (UID: \"40b701d5-4495-47c3-bb6e-6ba466958355\") " pod="openstack/ceilometer-0" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.975251 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/40b701d5-4495-47c3-bb6e-6ba466958355-log-httpd\") pod \"ceilometer-0\" (UID: \"40b701d5-4495-47c3-bb6e-6ba466958355\") " pod="openstack/ceilometer-0" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.975282 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/40b701d5-4495-47c3-bb6e-6ba466958355-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"40b701d5-4495-47c3-bb6e-6ba466958355\") " pod="openstack/ceilometer-0" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.975304 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40b701d5-4495-47c3-bb6e-6ba466958355-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"40b701d5-4495-47c3-bb6e-6ba466958355\") " pod="openstack/ceilometer-0" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.977780 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/40b701d5-4495-47c3-bb6e-6ba466958355-run-httpd\") pod \"ceilometer-0\" (UID: \"40b701d5-4495-47c3-bb6e-6ba466958355\") " pod="openstack/ceilometer-0" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.977808 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/40b701d5-4495-47c3-bb6e-6ba466958355-log-httpd\") pod \"ceilometer-0\" (UID: \"40b701d5-4495-47c3-bb6e-6ba466958355\") " pod="openstack/ceilometer-0" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.979797 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40b701d5-4495-47c3-bb6e-6ba466958355-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"40b701d5-4495-47c3-bb6e-6ba466958355\") " pod="openstack/ceilometer-0" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.979851 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/40b701d5-4495-47c3-bb6e-6ba466958355-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"40b701d5-4495-47c3-bb6e-6ba466958355\") " pod="openstack/ceilometer-0" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.981144 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/40b701d5-4495-47c3-bb6e-6ba466958355-scripts\") pod \"ceilometer-0\" (UID: \"40b701d5-4495-47c3-bb6e-6ba466958355\") " pod="openstack/ceilometer-0" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.983209 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40b701d5-4495-47c3-bb6e-6ba466958355-config-data\") pod \"ceilometer-0\" (UID: \"40b701d5-4495-47c3-bb6e-6ba466958355\") " pod="openstack/ceilometer-0" Feb 18 09:19:07 crc kubenswrapper[4556]: I0218 09:19:07.999725 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8blb\" (UniqueName: \"kubernetes.io/projected/40b701d5-4495-47c3-bb6e-6ba466958355-kube-api-access-p8blb\") pod \"ceilometer-0\" (UID: \"40b701d5-4495-47c3-bb6e-6ba466958355\") " pod="openstack/ceilometer-0" Feb 18 09:19:08 crc kubenswrapper[4556]: I0218 09:19:08.009265 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Feb 18 09:19:08 crc kubenswrapper[4556]: I0218 09:19:08.013688 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Feb 18 09:19:08 crc kubenswrapper[4556]: I0218 09:19:08.023318 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Feb 18 09:19:08 crc kubenswrapper[4556]: I0218 09:19:08.026080 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 18 09:19:08 crc kubenswrapper[4556]: I0218 09:19:08.029003 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Feb 18 09:19:08 crc kubenswrapper[4556]: I0218 09:19:08.029013 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Feb 18 09:19:08 crc kubenswrapper[4556]: I0218 09:19:08.029138 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Feb 18 09:19:08 crc kubenswrapper[4556]: I0218 09:19:08.036763 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 18 09:19:08 crc kubenswrapper[4556]: I0218 09:19:08.060580 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 18 09:19:08 crc kubenswrapper[4556]: I0218 09:19:08.078715 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d\") " pod="openstack/cinder-api-0" Feb 18 09:19:08 crc kubenswrapper[4556]: I0218 09:19:08.078824 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqqwg\" (UniqueName: \"kubernetes.io/projected/e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d-kube-api-access-cqqwg\") pod \"cinder-api-0\" (UID: \"e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d\") " pod="openstack/cinder-api-0" Feb 18 09:19:08 crc kubenswrapper[4556]: I0218 09:19:08.078942 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d\") " pod="openstack/cinder-api-0" Feb 18 09:19:08 crc kubenswrapper[4556]: I0218 09:19:08.078961 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d-public-tls-certs\") pod \"cinder-api-0\" (UID: \"e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d\") " pod="openstack/cinder-api-0" Feb 18 09:19:08 crc kubenswrapper[4556]: I0218 09:19:08.078976 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d-config-data-custom\") pod \"cinder-api-0\" (UID: \"e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d\") " pod="openstack/cinder-api-0" Feb 18 09:19:08 crc kubenswrapper[4556]: I0218 09:19:08.079025 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d-logs\") pod \"cinder-api-0\" (UID: \"e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d\") " pod="openstack/cinder-api-0" Feb 18 09:19:08 crc kubenswrapper[4556]: I0218 09:19:08.079099 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d-scripts\") pod \"cinder-api-0\" (UID: \"e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d\") " pod="openstack/cinder-api-0" Feb 18 09:19:08 crc kubenswrapper[4556]: I0218 09:19:08.079162 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d-config-data\") pod \"cinder-api-0\" (UID: \"e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d\") " pod="openstack/cinder-api-0" Feb 18 09:19:08 crc kubenswrapper[4556]: I0218 09:19:08.079356 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d\") " pod="openstack/cinder-api-0" Feb 18 09:19:08 crc kubenswrapper[4556]: I0218 09:19:08.180373 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d-config-data\") pod \"cinder-api-0\" (UID: \"e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d\") " pod="openstack/cinder-api-0" Feb 18 09:19:08 crc kubenswrapper[4556]: I0218 09:19:08.180778 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d\") " pod="openstack/cinder-api-0" Feb 18 09:19:08 crc kubenswrapper[4556]: I0218 09:19:08.180844 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d\") " pod="openstack/cinder-api-0" Feb 18 09:19:08 crc kubenswrapper[4556]: I0218 09:19:08.180875 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqqwg\" (UniqueName: \"kubernetes.io/projected/e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d-kube-api-access-cqqwg\") pod \"cinder-api-0\" (UID: \"e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d\") " pod="openstack/cinder-api-0" Feb 18 09:19:08 crc kubenswrapper[4556]: I0218 09:19:08.181460 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d\") " pod="openstack/cinder-api-0" Feb 18 09:19:08 crc kubenswrapper[4556]: I0218 09:19:08.181506 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d-public-tls-certs\") pod \"cinder-api-0\" (UID: \"e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d\") " pod="openstack/cinder-api-0" Feb 18 09:19:08 crc kubenswrapper[4556]: I0218 09:19:08.181524 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d-config-data-custom\") pod \"cinder-api-0\" (UID: \"e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d\") " pod="openstack/cinder-api-0" Feb 18 09:19:08 crc kubenswrapper[4556]: I0218 09:19:08.181548 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d-logs\") pod \"cinder-api-0\" (UID: \"e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d\") " pod="openstack/cinder-api-0" Feb 18 09:19:08 crc kubenswrapper[4556]: I0218 09:19:08.181061 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d\") " pod="openstack/cinder-api-0" Feb 18 09:19:08 crc kubenswrapper[4556]: I0218 09:19:08.182110 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d-logs\") pod \"cinder-api-0\" (UID: \"e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d\") " pod="openstack/cinder-api-0" Feb 18 09:19:08 crc kubenswrapper[4556]: I0218 09:19:08.182124 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d-scripts\") pod \"cinder-api-0\" (UID: \"e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d\") " pod="openstack/cinder-api-0" Feb 18 09:19:08 crc kubenswrapper[4556]: I0218 09:19:08.185730 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d-scripts\") pod \"cinder-api-0\" (UID: \"e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d\") " pod="openstack/cinder-api-0" Feb 18 09:19:08 crc kubenswrapper[4556]: I0218 09:19:08.187128 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d\") " pod="openstack/cinder-api-0" Feb 18 09:19:08 crc kubenswrapper[4556]: I0218 09:19:08.187562 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d-public-tls-certs\") pod \"cinder-api-0\" (UID: \"e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d\") " pod="openstack/cinder-api-0" Feb 18 09:19:08 crc kubenswrapper[4556]: I0218 09:19:08.187682 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d\") " pod="openstack/cinder-api-0" Feb 18 09:19:08 crc kubenswrapper[4556]: I0218 09:19:08.188584 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d-config-data-custom\") pod \"cinder-api-0\" (UID: \"e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d\") " pod="openstack/cinder-api-0" Feb 18 09:19:08 crc kubenswrapper[4556]: I0218 09:19:08.190695 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d-config-data\") pod \"cinder-api-0\" (UID: \"e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d\") " pod="openstack/cinder-api-0" Feb 18 09:19:08 crc kubenswrapper[4556]: I0218 09:19:08.197979 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqqwg\" (UniqueName: \"kubernetes.io/projected/e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d-kube-api-access-cqqwg\") pod \"cinder-api-0\" (UID: \"e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d\") " pod="openstack/cinder-api-0" Feb 18 09:19:08 crc kubenswrapper[4556]: I0218 09:19:08.372838 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 18 09:19:08 crc kubenswrapper[4556]: I0218 09:19:08.532182 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 18 09:19:08 crc kubenswrapper[4556]: I0218 09:19:08.701514 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"40b701d5-4495-47c3-bb6e-6ba466958355","Type":"ContainerStarted","Data":"e11c71227e2adf3a67dc47bc79ad4548bdf54b4ca48ccc8a77137ac6f5f17844"} Feb 18 09:19:08 crc kubenswrapper[4556]: I0218 09:19:08.702326 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-nghmb" podUID="3a0954e9-38c5-435c-9ced-d8e910e284f9" containerName="registry-server" containerID="cri-o://307bdbb1384330f314b33f4e5a051ba4a57c6383a5b0f9f4720db0713087c297" gracePeriod=2 Feb 18 09:19:08 crc kubenswrapper[4556]: I0218 09:19:08.866955 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.140585 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nghmb" Feb 18 09:19:09 crc kubenswrapper[4556]: E0218 09:19:09.177548 4556 manager.go:1116] Failed to create existing container: /kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2bb44e10_7532_4173_a140_140b161601a4.slice/crio-f097d1897907d95895e9c90beb30d488dd020fb1b4fe76bc9601482a0968b723: Error finding container f097d1897907d95895e9c90beb30d488dd020fb1b4fe76bc9601482a0968b723: Status 404 returned error can't find the container with id f097d1897907d95895e9c90beb30d488dd020fb1b4fe76bc9601482a0968b723 Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.214205 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qpd8w\" (UniqueName: \"kubernetes.io/projected/3a0954e9-38c5-435c-9ced-d8e910e284f9-kube-api-access-qpd8w\") pod \"3a0954e9-38c5-435c-9ced-d8e910e284f9\" (UID: \"3a0954e9-38c5-435c-9ced-d8e910e284f9\") " Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.214244 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a0954e9-38c5-435c-9ced-d8e910e284f9-catalog-content\") pod \"3a0954e9-38c5-435c-9ced-d8e910e284f9\" (UID: \"3a0954e9-38c5-435c-9ced-d8e910e284f9\") " Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.214330 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a0954e9-38c5-435c-9ced-d8e910e284f9-utilities\") pod \"3a0954e9-38c5-435c-9ced-d8e910e284f9\" (UID: \"3a0954e9-38c5-435c-9ced-d8e910e284f9\") " Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.217565 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a0954e9-38c5-435c-9ced-d8e910e284f9-utilities" (OuterVolumeSpecName: "utilities") pod "3a0954e9-38c5-435c-9ced-d8e910e284f9" (UID: "3a0954e9-38c5-435c-9ced-d8e910e284f9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.221563 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a0954e9-38c5-435c-9ced-d8e910e284f9-kube-api-access-qpd8w" (OuterVolumeSpecName: "kube-api-access-qpd8w") pod "3a0954e9-38c5-435c-9ced-d8e910e284f9" (UID: "3a0954e9-38c5-435c-9ced-d8e910e284f9"). InnerVolumeSpecName "kube-api-access-qpd8w". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.234284 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a0954e9-38c5-435c-9ced-d8e910e284f9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3a0954e9-38c5-435c-9ced-d8e910e284f9" (UID: "3a0954e9-38c5-435c-9ced-d8e910e284f9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.294370 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2bb44e10-7532-4173-a140-140b161601a4" path="/var/lib/kubelet/pods/2bb44e10-7532-4173-a140-140b161601a4/volumes" Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.295044 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0739f72-891b-46e1-bce5-e17c54a407ab" path="/var/lib/kubelet/pods/e0739f72-891b-46e1-bce5-e17c54a407ab/volumes" Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.295806 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e15b6829-420b-493f-a31c-3bde1c53035f" path="/var/lib/kubelet/pods/e15b6829-420b-493f-a31c-3bde1c53035f/volumes" Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.317513 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qpd8w\" (UniqueName: \"kubernetes.io/projected/3a0954e9-38c5-435c-9ced-d8e910e284f9-kube-api-access-qpd8w\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.317539 4556 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a0954e9-38c5-435c-9ced-d8e910e284f9-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.317550 4556 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a0954e9-38c5-435c-9ced-d8e910e284f9-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:09 crc kubenswrapper[4556]: E0218 09:19:09.390073 4556 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3354b930_6b00_4aa7_a4bd_97ec410cc863.slice/crio-c1df47a378b25fad3361d8a29971e5ef0d47d8783acc451f619234849683c0f1.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3a0954e9_38c5_435c_9ced_d8e910e284f9.slice/crio-307bdbb1384330f314b33f4e5a051ba4a57c6383a5b0f9f4720db0713087c297.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3354b930_6b00_4aa7_a4bd_97ec410cc863.slice/crio-conmon-c1df47a378b25fad3361d8a29971e5ef0d47d8783acc451f619234849683c0f1.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3a0954e9_38c5_435c_9ced_d8e910e284f9.slice/crio-conmon-307bdbb1384330f314b33f4e5a051ba4a57c6383a5b0f9f4720db0713087c297.scope\": RecentStats: unable to find data in memory cache]" Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.452464 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-85bb8c66f4-fhblh" Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.519979 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3354b930-6b00-4aa7-a4bd-97ec410cc863-scripts\") pod \"3354b930-6b00-4aa7-a4bd-97ec410cc863\" (UID: \"3354b930-6b00-4aa7-a4bd-97ec410cc863\") " Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.520111 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3354b930-6b00-4aa7-a4bd-97ec410cc863-combined-ca-bundle\") pod \"3354b930-6b00-4aa7-a4bd-97ec410cc863\" (UID: \"3354b930-6b00-4aa7-a4bd-97ec410cc863\") " Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.520242 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3354b930-6b00-4aa7-a4bd-97ec410cc863-logs\") pod \"3354b930-6b00-4aa7-a4bd-97ec410cc863\" (UID: \"3354b930-6b00-4aa7-a4bd-97ec410cc863\") " Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.520431 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3354b930-6b00-4aa7-a4bd-97ec410cc863-horizon-secret-key\") pod \"3354b930-6b00-4aa7-a4bd-97ec410cc863\" (UID: \"3354b930-6b00-4aa7-a4bd-97ec410cc863\") " Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.520520 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3354b930-6b00-4aa7-a4bd-97ec410cc863-config-data\") pod \"3354b930-6b00-4aa7-a4bd-97ec410cc863\" (UID: \"3354b930-6b00-4aa7-a4bd-97ec410cc863\") " Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.520621 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/3354b930-6b00-4aa7-a4bd-97ec410cc863-horizon-tls-certs\") pod \"3354b930-6b00-4aa7-a4bd-97ec410cc863\" (UID: \"3354b930-6b00-4aa7-a4bd-97ec410cc863\") " Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.520711 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ffnjj\" (UniqueName: \"kubernetes.io/projected/3354b930-6b00-4aa7-a4bd-97ec410cc863-kube-api-access-ffnjj\") pod \"3354b930-6b00-4aa7-a4bd-97ec410cc863\" (UID: \"3354b930-6b00-4aa7-a4bd-97ec410cc863\") " Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.522023 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3354b930-6b00-4aa7-a4bd-97ec410cc863-logs" (OuterVolumeSpecName: "logs") pod "3354b930-6b00-4aa7-a4bd-97ec410cc863" (UID: "3354b930-6b00-4aa7-a4bd-97ec410cc863"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.526354 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3354b930-6b00-4aa7-a4bd-97ec410cc863-kube-api-access-ffnjj" (OuterVolumeSpecName: "kube-api-access-ffnjj") pod "3354b930-6b00-4aa7-a4bd-97ec410cc863" (UID: "3354b930-6b00-4aa7-a4bd-97ec410cc863"). InnerVolumeSpecName "kube-api-access-ffnjj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.527273 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3354b930-6b00-4aa7-a4bd-97ec410cc863-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "3354b930-6b00-4aa7-a4bd-97ec410cc863" (UID: "3354b930-6b00-4aa7-a4bd-97ec410cc863"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.554818 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3354b930-6b00-4aa7-a4bd-97ec410cc863-config-data" (OuterVolumeSpecName: "config-data") pod "3354b930-6b00-4aa7-a4bd-97ec410cc863" (UID: "3354b930-6b00-4aa7-a4bd-97ec410cc863"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.557335 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3354b930-6b00-4aa7-a4bd-97ec410cc863-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3354b930-6b00-4aa7-a4bd-97ec410cc863" (UID: "3354b930-6b00-4aa7-a4bd-97ec410cc863"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.558915 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3354b930-6b00-4aa7-a4bd-97ec410cc863-scripts" (OuterVolumeSpecName: "scripts") pod "3354b930-6b00-4aa7-a4bd-97ec410cc863" (UID: "3354b930-6b00-4aa7-a4bd-97ec410cc863"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.587261 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3354b930-6b00-4aa7-a4bd-97ec410cc863-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "3354b930-6b00-4aa7-a4bd-97ec410cc863" (UID: "3354b930-6b00-4aa7-a4bd-97ec410cc863"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.623050 4556 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3354b930-6b00-4aa7-a4bd-97ec410cc863-scripts\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.623090 4556 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3354b930-6b00-4aa7-a4bd-97ec410cc863-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.623100 4556 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3354b930-6b00-4aa7-a4bd-97ec410cc863-logs\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.623112 4556 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3354b930-6b00-4aa7-a4bd-97ec410cc863-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.623122 4556 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3354b930-6b00-4aa7-a4bd-97ec410cc863-config-data\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.623131 4556 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/3354b930-6b00-4aa7-a4bd-97ec410cc863-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.623176 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ffnjj\" (UniqueName: \"kubernetes.io/projected/3354b930-6b00-4aa7-a4bd-97ec410cc863-kube-api-access-ffnjj\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.714058 4556 generic.go:334] "Generic (PLEG): container finished" podID="3a0954e9-38c5-435c-9ced-d8e910e284f9" containerID="307bdbb1384330f314b33f4e5a051ba4a57c6383a5b0f9f4720db0713087c297" exitCode=0 Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.714103 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nghmb" event={"ID":"3a0954e9-38c5-435c-9ced-d8e910e284f9","Type":"ContainerDied","Data":"307bdbb1384330f314b33f4e5a051ba4a57c6383a5b0f9f4720db0713087c297"} Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.714192 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nghmb" event={"ID":"3a0954e9-38c5-435c-9ced-d8e910e284f9","Type":"ContainerDied","Data":"7ee7f916a52e0517e42b13a26438a3541d4ef44c10d559bfff948a0f9aee1ced"} Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.714217 4556 scope.go:117] "RemoveContainer" containerID="307bdbb1384330f314b33f4e5a051ba4a57c6383a5b0f9f4720db0713087c297" Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.714126 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nghmb" Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.721821 4556 generic.go:334] "Generic (PLEG): container finished" podID="3354b930-6b00-4aa7-a4bd-97ec410cc863" containerID="c1df47a378b25fad3361d8a29971e5ef0d47d8783acc451f619234849683c0f1" exitCode=137 Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.721885 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-85bb8c66f4-fhblh" Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.721889 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-85bb8c66f4-fhblh" event={"ID":"3354b930-6b00-4aa7-a4bd-97ec410cc863","Type":"ContainerDied","Data":"c1df47a378b25fad3361d8a29971e5ef0d47d8783acc451f619234849683c0f1"} Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.722456 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-85bb8c66f4-fhblh" event={"ID":"3354b930-6b00-4aa7-a4bd-97ec410cc863","Type":"ContainerDied","Data":"0ab01d8a9a09a2f53fe5451dc6aba62a035bf8862bbcdd87335a79bca55b4e46"} Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.726876 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d","Type":"ContainerStarted","Data":"3d2364c0036c52bdbf61757ac78903747390f7f43904e9c49d4048a814f3e90b"} Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.726914 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d","Type":"ContainerStarted","Data":"fb73b93ec8e8796bf8945508467c0bc8dd7784e5ad82a78704a4608e25387996"} Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.732551 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"40b701d5-4495-47c3-bb6e-6ba466958355","Type":"ContainerStarted","Data":"b0838fd427fe9647da3a2c765ef31d6f5e3b1d99df1a0a099104c13db47316a4"} Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.736219 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nghmb"] Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.745966 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-nghmb"] Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.746475 4556 scope.go:117] "RemoveContainer" containerID="fad04e69ea4a5850f128f2634e47f2422613db7a840386f5fd27819d8bcb0ff6" Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.767689 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-85bb8c66f4-fhblh"] Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.774782 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-85bb8c66f4-fhblh"] Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.786770 4556 scope.go:117] "RemoveContainer" containerID="75e22c711b14ccf5ce6a0acd5950f61dc69f87d5df160be30e872749562bc6f8" Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.811875 4556 scope.go:117] "RemoveContainer" containerID="307bdbb1384330f314b33f4e5a051ba4a57c6383a5b0f9f4720db0713087c297" Feb 18 09:19:09 crc kubenswrapper[4556]: E0218 09:19:09.813493 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"307bdbb1384330f314b33f4e5a051ba4a57c6383a5b0f9f4720db0713087c297\": container with ID starting with 307bdbb1384330f314b33f4e5a051ba4a57c6383a5b0f9f4720db0713087c297 not found: ID does not exist" containerID="307bdbb1384330f314b33f4e5a051ba4a57c6383a5b0f9f4720db0713087c297" Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.813540 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"307bdbb1384330f314b33f4e5a051ba4a57c6383a5b0f9f4720db0713087c297"} err="failed to get container status \"307bdbb1384330f314b33f4e5a051ba4a57c6383a5b0f9f4720db0713087c297\": rpc error: code = NotFound desc = could not find container \"307bdbb1384330f314b33f4e5a051ba4a57c6383a5b0f9f4720db0713087c297\": container with ID starting with 307bdbb1384330f314b33f4e5a051ba4a57c6383a5b0f9f4720db0713087c297 not found: ID does not exist" Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.813559 4556 scope.go:117] "RemoveContainer" containerID="fad04e69ea4a5850f128f2634e47f2422613db7a840386f5fd27819d8bcb0ff6" Feb 18 09:19:09 crc kubenswrapper[4556]: E0218 09:19:09.814118 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fad04e69ea4a5850f128f2634e47f2422613db7a840386f5fd27819d8bcb0ff6\": container with ID starting with fad04e69ea4a5850f128f2634e47f2422613db7a840386f5fd27819d8bcb0ff6 not found: ID does not exist" containerID="fad04e69ea4a5850f128f2634e47f2422613db7a840386f5fd27819d8bcb0ff6" Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.814172 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fad04e69ea4a5850f128f2634e47f2422613db7a840386f5fd27819d8bcb0ff6"} err="failed to get container status \"fad04e69ea4a5850f128f2634e47f2422613db7a840386f5fd27819d8bcb0ff6\": rpc error: code = NotFound desc = could not find container \"fad04e69ea4a5850f128f2634e47f2422613db7a840386f5fd27819d8bcb0ff6\": container with ID starting with fad04e69ea4a5850f128f2634e47f2422613db7a840386f5fd27819d8bcb0ff6 not found: ID does not exist" Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.814215 4556 scope.go:117] "RemoveContainer" containerID="75e22c711b14ccf5ce6a0acd5950f61dc69f87d5df160be30e872749562bc6f8" Feb 18 09:19:09 crc kubenswrapper[4556]: E0218 09:19:09.814500 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75e22c711b14ccf5ce6a0acd5950f61dc69f87d5df160be30e872749562bc6f8\": container with ID starting with 75e22c711b14ccf5ce6a0acd5950f61dc69f87d5df160be30e872749562bc6f8 not found: ID does not exist" containerID="75e22c711b14ccf5ce6a0acd5950f61dc69f87d5df160be30e872749562bc6f8" Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.814521 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75e22c711b14ccf5ce6a0acd5950f61dc69f87d5df160be30e872749562bc6f8"} err="failed to get container status \"75e22c711b14ccf5ce6a0acd5950f61dc69f87d5df160be30e872749562bc6f8\": rpc error: code = NotFound desc = could not find container \"75e22c711b14ccf5ce6a0acd5950f61dc69f87d5df160be30e872749562bc6f8\": container with ID starting with 75e22c711b14ccf5ce6a0acd5950f61dc69f87d5df160be30e872749562bc6f8 not found: ID does not exist" Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.814536 4556 scope.go:117] "RemoveContainer" containerID="4423fda867205940e352d08e4c633d2caee4bc72b78e6280dd5fd52606caa4dd" Feb 18 09:19:09 crc kubenswrapper[4556]: I0218 09:19:09.985464 4556 scope.go:117] "RemoveContainer" containerID="c1df47a378b25fad3361d8a29971e5ef0d47d8783acc451f619234849683c0f1" Feb 18 09:19:10 crc kubenswrapper[4556]: I0218 09:19:10.002710 4556 scope.go:117] "RemoveContainer" containerID="4423fda867205940e352d08e4c633d2caee4bc72b78e6280dd5fd52606caa4dd" Feb 18 09:19:10 crc kubenswrapper[4556]: E0218 09:19:10.003239 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4423fda867205940e352d08e4c633d2caee4bc72b78e6280dd5fd52606caa4dd\": container with ID starting with 4423fda867205940e352d08e4c633d2caee4bc72b78e6280dd5fd52606caa4dd not found: ID does not exist" containerID="4423fda867205940e352d08e4c633d2caee4bc72b78e6280dd5fd52606caa4dd" Feb 18 09:19:10 crc kubenswrapper[4556]: I0218 09:19:10.003272 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4423fda867205940e352d08e4c633d2caee4bc72b78e6280dd5fd52606caa4dd"} err="failed to get container status \"4423fda867205940e352d08e4c633d2caee4bc72b78e6280dd5fd52606caa4dd\": rpc error: code = NotFound desc = could not find container \"4423fda867205940e352d08e4c633d2caee4bc72b78e6280dd5fd52606caa4dd\": container with ID starting with 4423fda867205940e352d08e4c633d2caee4bc72b78e6280dd5fd52606caa4dd not found: ID does not exist" Feb 18 09:19:10 crc kubenswrapper[4556]: I0218 09:19:10.003307 4556 scope.go:117] "RemoveContainer" containerID="c1df47a378b25fad3361d8a29971e5ef0d47d8783acc451f619234849683c0f1" Feb 18 09:19:10 crc kubenswrapper[4556]: E0218 09:19:10.003594 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1df47a378b25fad3361d8a29971e5ef0d47d8783acc451f619234849683c0f1\": container with ID starting with c1df47a378b25fad3361d8a29971e5ef0d47d8783acc451f619234849683c0f1 not found: ID does not exist" containerID="c1df47a378b25fad3361d8a29971e5ef0d47d8783acc451f619234849683c0f1" Feb 18 09:19:10 crc kubenswrapper[4556]: I0218 09:19:10.003618 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1df47a378b25fad3361d8a29971e5ef0d47d8783acc451f619234849683c0f1"} err="failed to get container status \"c1df47a378b25fad3361d8a29971e5ef0d47d8783acc451f619234849683c0f1\": rpc error: code = NotFound desc = could not find container \"c1df47a378b25fad3361d8a29971e5ef0d47d8783acc451f619234849683c0f1\": container with ID starting with c1df47a378b25fad3361d8a29971e5ef0d47d8783acc451f619234849683c0f1 not found: ID does not exist" Feb 18 09:19:10 crc kubenswrapper[4556]: I0218 09:19:10.744611 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"40b701d5-4495-47c3-bb6e-6ba466958355","Type":"ContainerStarted","Data":"471e8d3bb0e3ae1e40387a1ab945b5b86c109e0aae7d0c0de060669dafb0b2e1"} Feb 18 09:19:10 crc kubenswrapper[4556]: I0218 09:19:10.751594 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d","Type":"ContainerStarted","Data":"71fb8f4e38efd86e0d7495d1e332dc658112e8a77c889d80ba70b883f478cdfe"} Feb 18 09:19:10 crc kubenswrapper[4556]: I0218 09:19:10.752570 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Feb 18 09:19:11 crc kubenswrapper[4556]: I0218 09:19:11.291795 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3354b930-6b00-4aa7-a4bd-97ec410cc863" path="/var/lib/kubelet/pods/3354b930-6b00-4aa7-a4bd-97ec410cc863/volumes" Feb 18 09:19:11 crc kubenswrapper[4556]: I0218 09:19:11.292693 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a0954e9-38c5-435c-9ced-d8e910e284f9" path="/var/lib/kubelet/pods/3a0954e9-38c5-435c-9ced-d8e910e284f9/volumes" Feb 18 09:19:11 crc kubenswrapper[4556]: I0218 09:19:11.522270 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.522255171 podStartE2EDuration="4.522255171s" podCreationTimestamp="2026-02-18 09:19:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:19:10.77340852 +0000 UTC m=+907.790369500" watchObservedRunningTime="2026-02-18 09:19:11.522255171 +0000 UTC m=+908.539216151" Feb 18 09:19:11 crc kubenswrapper[4556]: I0218 09:19:11.525485 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-f7xdx"] Feb 18 09:19:11 crc kubenswrapper[4556]: E0218 09:19:11.525804 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3354b930-6b00-4aa7-a4bd-97ec410cc863" containerName="horizon-log" Feb 18 09:19:11 crc kubenswrapper[4556]: I0218 09:19:11.525829 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="3354b930-6b00-4aa7-a4bd-97ec410cc863" containerName="horizon-log" Feb 18 09:19:11 crc kubenswrapper[4556]: E0218 09:19:11.525844 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a0954e9-38c5-435c-9ced-d8e910e284f9" containerName="extract-content" Feb 18 09:19:11 crc kubenswrapper[4556]: I0218 09:19:11.525851 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a0954e9-38c5-435c-9ced-d8e910e284f9" containerName="extract-content" Feb 18 09:19:11 crc kubenswrapper[4556]: E0218 09:19:11.525868 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a0954e9-38c5-435c-9ced-d8e910e284f9" containerName="extract-utilities" Feb 18 09:19:11 crc kubenswrapper[4556]: I0218 09:19:11.525873 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a0954e9-38c5-435c-9ced-d8e910e284f9" containerName="extract-utilities" Feb 18 09:19:11 crc kubenswrapper[4556]: E0218 09:19:11.525883 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a0954e9-38c5-435c-9ced-d8e910e284f9" containerName="registry-server" Feb 18 09:19:11 crc kubenswrapper[4556]: I0218 09:19:11.525888 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a0954e9-38c5-435c-9ced-d8e910e284f9" containerName="registry-server" Feb 18 09:19:11 crc kubenswrapper[4556]: E0218 09:19:11.525913 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3354b930-6b00-4aa7-a4bd-97ec410cc863" containerName="horizon" Feb 18 09:19:11 crc kubenswrapper[4556]: I0218 09:19:11.525918 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="3354b930-6b00-4aa7-a4bd-97ec410cc863" containerName="horizon" Feb 18 09:19:11 crc kubenswrapper[4556]: I0218 09:19:11.526103 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a0954e9-38c5-435c-9ced-d8e910e284f9" containerName="registry-server" Feb 18 09:19:11 crc kubenswrapper[4556]: I0218 09:19:11.526123 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="3354b930-6b00-4aa7-a4bd-97ec410cc863" containerName="horizon-log" Feb 18 09:19:11 crc kubenswrapper[4556]: I0218 09:19:11.526131 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="3354b930-6b00-4aa7-a4bd-97ec410cc863" containerName="horizon" Feb 18 09:19:11 crc kubenswrapper[4556]: I0218 09:19:11.528090 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-f7xdx" Feb 18 09:19:11 crc kubenswrapper[4556]: I0218 09:19:11.539209 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-f7xdx"] Feb 18 09:19:11 crc kubenswrapper[4556]: I0218 09:19:11.565273 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/96a4b91d-0008-4af7-8522-6842a8c627a8-operator-scripts\") pod \"nova-api-db-create-f7xdx\" (UID: \"96a4b91d-0008-4af7-8522-6842a8c627a8\") " pod="openstack/nova-api-db-create-f7xdx" Feb 18 09:19:11 crc kubenswrapper[4556]: I0218 09:19:11.565653 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxkpr\" (UniqueName: \"kubernetes.io/projected/96a4b91d-0008-4af7-8522-6842a8c627a8-kube-api-access-xxkpr\") pod \"nova-api-db-create-f7xdx\" (UID: \"96a4b91d-0008-4af7-8522-6842a8c627a8\") " pod="openstack/nova-api-db-create-f7xdx" Feb 18 09:19:11 crc kubenswrapper[4556]: I0218 09:19:11.667677 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxkpr\" (UniqueName: \"kubernetes.io/projected/96a4b91d-0008-4af7-8522-6842a8c627a8-kube-api-access-xxkpr\") pod \"nova-api-db-create-f7xdx\" (UID: \"96a4b91d-0008-4af7-8522-6842a8c627a8\") " pod="openstack/nova-api-db-create-f7xdx" Feb 18 09:19:11 crc kubenswrapper[4556]: I0218 09:19:11.667868 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/96a4b91d-0008-4af7-8522-6842a8c627a8-operator-scripts\") pod \"nova-api-db-create-f7xdx\" (UID: \"96a4b91d-0008-4af7-8522-6842a8c627a8\") " pod="openstack/nova-api-db-create-f7xdx" Feb 18 09:19:11 crc kubenswrapper[4556]: I0218 09:19:11.668617 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/96a4b91d-0008-4af7-8522-6842a8c627a8-operator-scripts\") pod \"nova-api-db-create-f7xdx\" (UID: \"96a4b91d-0008-4af7-8522-6842a8c627a8\") " pod="openstack/nova-api-db-create-f7xdx" Feb 18 09:19:11 crc kubenswrapper[4556]: I0218 09:19:11.700816 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxkpr\" (UniqueName: \"kubernetes.io/projected/96a4b91d-0008-4af7-8522-6842a8c627a8-kube-api-access-xxkpr\") pod \"nova-api-db-create-f7xdx\" (UID: \"96a4b91d-0008-4af7-8522-6842a8c627a8\") " pod="openstack/nova-api-db-create-f7xdx" Feb 18 09:19:11 crc kubenswrapper[4556]: I0218 09:19:11.750681 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-nq759"] Feb 18 09:19:11 crc kubenswrapper[4556]: I0218 09:19:11.751932 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-nq759" Feb 18 09:19:11 crc kubenswrapper[4556]: I0218 09:19:11.761705 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"40b701d5-4495-47c3-bb6e-6ba466958355","Type":"ContainerStarted","Data":"f984aeec06953af9d589b06b93e201997457368d82ecc83ab082489d056e155f"} Feb 18 09:19:11 crc kubenswrapper[4556]: I0218 09:19:11.765290 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-nq759"] Feb 18 09:19:11 crc kubenswrapper[4556]: I0218 09:19:11.769580 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/65bf8803-f4f0-4366-8f1f-d894f2f407e2-operator-scripts\") pod \"nova-cell0-db-create-nq759\" (UID: \"65bf8803-f4f0-4366-8f1f-d894f2f407e2\") " pod="openstack/nova-cell0-db-create-nq759" Feb 18 09:19:11 crc kubenswrapper[4556]: I0218 09:19:11.769702 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mx8c2\" (UniqueName: \"kubernetes.io/projected/65bf8803-f4f0-4366-8f1f-d894f2f407e2-kube-api-access-mx8c2\") pod \"nova-cell0-db-create-nq759\" (UID: \"65bf8803-f4f0-4366-8f1f-d894f2f407e2\") " pod="openstack/nova-cell0-db-create-nq759" Feb 18 09:19:11 crc kubenswrapper[4556]: I0218 09:19:11.770884 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-5bef-account-create-update-g2mzr"] Feb 18 09:19:11 crc kubenswrapper[4556]: I0218 09:19:11.771913 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-5bef-account-create-update-g2mzr" Feb 18 09:19:11 crc kubenswrapper[4556]: I0218 09:19:11.775480 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Feb 18 09:19:11 crc kubenswrapper[4556]: I0218 09:19:11.810364 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-5bef-account-create-update-g2mzr"] Feb 18 09:19:11 crc kubenswrapper[4556]: I0218 09:19:11.843399 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-f7xdx" Feb 18 09:19:11 crc kubenswrapper[4556]: I0218 09:19:11.866279 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-ln2s2"] Feb 18 09:19:11 crc kubenswrapper[4556]: I0218 09:19:11.867400 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-ln2s2" Feb 18 09:19:11 crc kubenswrapper[4556]: I0218 09:19:11.873951 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mx8c2\" (UniqueName: \"kubernetes.io/projected/65bf8803-f4f0-4366-8f1f-d894f2f407e2-kube-api-access-mx8c2\") pod \"nova-cell0-db-create-nq759\" (UID: \"65bf8803-f4f0-4366-8f1f-d894f2f407e2\") " pod="openstack/nova-cell0-db-create-nq759" Feb 18 09:19:11 crc kubenswrapper[4556]: I0218 09:19:11.874056 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7da9e29c-e9f2-49bf-a3b3-6e1b764c2b78-operator-scripts\") pod \"nova-api-5bef-account-create-update-g2mzr\" (UID: \"7da9e29c-e9f2-49bf-a3b3-6e1b764c2b78\") " pod="openstack/nova-api-5bef-account-create-update-g2mzr" Feb 18 09:19:11 crc kubenswrapper[4556]: I0218 09:19:11.874190 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/65bf8803-f4f0-4366-8f1f-d894f2f407e2-operator-scripts\") pod \"nova-cell0-db-create-nq759\" (UID: \"65bf8803-f4f0-4366-8f1f-d894f2f407e2\") " pod="openstack/nova-cell0-db-create-nq759" Feb 18 09:19:11 crc kubenswrapper[4556]: I0218 09:19:11.874285 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6jcn\" (UniqueName: \"kubernetes.io/projected/7da9e29c-e9f2-49bf-a3b3-6e1b764c2b78-kube-api-access-k6jcn\") pod \"nova-api-5bef-account-create-update-g2mzr\" (UID: \"7da9e29c-e9f2-49bf-a3b3-6e1b764c2b78\") " pod="openstack/nova-api-5bef-account-create-update-g2mzr" Feb 18 09:19:11 crc kubenswrapper[4556]: I0218 09:19:11.876259 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/65bf8803-f4f0-4366-8f1f-d894f2f407e2-operator-scripts\") pod \"nova-cell0-db-create-nq759\" (UID: \"65bf8803-f4f0-4366-8f1f-d894f2f407e2\") " pod="openstack/nova-cell0-db-create-nq759" Feb 18 09:19:11 crc kubenswrapper[4556]: I0218 09:19:11.914209 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-ln2s2"] Feb 18 09:19:11 crc kubenswrapper[4556]: I0218 09:19:11.932885 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mx8c2\" (UniqueName: \"kubernetes.io/projected/65bf8803-f4f0-4366-8f1f-d894f2f407e2-kube-api-access-mx8c2\") pod \"nova-cell0-db-create-nq759\" (UID: \"65bf8803-f4f0-4366-8f1f-d894f2f407e2\") " pod="openstack/nova-cell0-db-create-nq759" Feb 18 09:19:11 crc kubenswrapper[4556]: I0218 09:19:11.974485 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-6c31-account-create-update-2wxx2"] Feb 18 09:19:11 crc kubenswrapper[4556]: I0218 09:19:11.976355 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7da9e29c-e9f2-49bf-a3b3-6e1b764c2b78-operator-scripts\") pod \"nova-api-5bef-account-create-update-g2mzr\" (UID: \"7da9e29c-e9f2-49bf-a3b3-6e1b764c2b78\") " pod="openstack/nova-api-5bef-account-create-update-g2mzr" Feb 18 09:19:11 crc kubenswrapper[4556]: I0218 09:19:11.976427 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9d482227-d966-4319-bcfd-10944085417b-operator-scripts\") pod \"nova-cell1-db-create-ln2s2\" (UID: \"9d482227-d966-4319-bcfd-10944085417b\") " pod="openstack/nova-cell1-db-create-ln2s2" Feb 18 09:19:11 crc kubenswrapper[4556]: I0218 09:19:11.976481 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6jcn\" (UniqueName: \"kubernetes.io/projected/7da9e29c-e9f2-49bf-a3b3-6e1b764c2b78-kube-api-access-k6jcn\") pod \"nova-api-5bef-account-create-update-g2mzr\" (UID: \"7da9e29c-e9f2-49bf-a3b3-6e1b764c2b78\") " pod="openstack/nova-api-5bef-account-create-update-g2mzr" Feb 18 09:19:11 crc kubenswrapper[4556]: I0218 09:19:11.976499 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9jd5\" (UniqueName: \"kubernetes.io/projected/9d482227-d966-4319-bcfd-10944085417b-kube-api-access-x9jd5\") pod \"nova-cell1-db-create-ln2s2\" (UID: \"9d482227-d966-4319-bcfd-10944085417b\") " pod="openstack/nova-cell1-db-create-ln2s2" Feb 18 09:19:11 crc kubenswrapper[4556]: I0218 09:19:11.976432 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-6c31-account-create-update-2wxx2" Feb 18 09:19:11 crc kubenswrapper[4556]: I0218 09:19:11.977147 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7da9e29c-e9f2-49bf-a3b3-6e1b764c2b78-operator-scripts\") pod \"nova-api-5bef-account-create-update-g2mzr\" (UID: \"7da9e29c-e9f2-49bf-a3b3-6e1b764c2b78\") " pod="openstack/nova-api-5bef-account-create-update-g2mzr" Feb 18 09:19:11 crc kubenswrapper[4556]: I0218 09:19:11.990100 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Feb 18 09:19:12 crc kubenswrapper[4556]: I0218 09:19:12.013815 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6jcn\" (UniqueName: \"kubernetes.io/projected/7da9e29c-e9f2-49bf-a3b3-6e1b764c2b78-kube-api-access-k6jcn\") pod \"nova-api-5bef-account-create-update-g2mzr\" (UID: \"7da9e29c-e9f2-49bf-a3b3-6e1b764c2b78\") " pod="openstack/nova-api-5bef-account-create-update-g2mzr" Feb 18 09:19:12 crc kubenswrapper[4556]: I0218 09:19:12.035203 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-6c31-account-create-update-2wxx2"] Feb 18 09:19:12 crc kubenswrapper[4556]: I0218 09:19:12.067801 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-nq759" Feb 18 09:19:12 crc kubenswrapper[4556]: I0218 09:19:12.083602 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3bab749-15ef-4b1d-a9ab-f51acb445a96-operator-scripts\") pod \"nova-cell0-6c31-account-create-update-2wxx2\" (UID: \"b3bab749-15ef-4b1d-a9ab-f51acb445a96\") " pod="openstack/nova-cell0-6c31-account-create-update-2wxx2" Feb 18 09:19:12 crc kubenswrapper[4556]: I0218 09:19:12.083666 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qf8rk\" (UniqueName: \"kubernetes.io/projected/b3bab749-15ef-4b1d-a9ab-f51acb445a96-kube-api-access-qf8rk\") pod \"nova-cell0-6c31-account-create-update-2wxx2\" (UID: \"b3bab749-15ef-4b1d-a9ab-f51acb445a96\") " pod="openstack/nova-cell0-6c31-account-create-update-2wxx2" Feb 18 09:19:12 crc kubenswrapper[4556]: I0218 09:19:12.083706 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9d482227-d966-4319-bcfd-10944085417b-operator-scripts\") pod \"nova-cell1-db-create-ln2s2\" (UID: \"9d482227-d966-4319-bcfd-10944085417b\") " pod="openstack/nova-cell1-db-create-ln2s2" Feb 18 09:19:12 crc kubenswrapper[4556]: I0218 09:19:12.083784 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9jd5\" (UniqueName: \"kubernetes.io/projected/9d482227-d966-4319-bcfd-10944085417b-kube-api-access-x9jd5\") pod \"nova-cell1-db-create-ln2s2\" (UID: \"9d482227-d966-4319-bcfd-10944085417b\") " pod="openstack/nova-cell1-db-create-ln2s2" Feb 18 09:19:12 crc kubenswrapper[4556]: I0218 09:19:12.086059 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9d482227-d966-4319-bcfd-10944085417b-operator-scripts\") pod \"nova-cell1-db-create-ln2s2\" (UID: \"9d482227-d966-4319-bcfd-10944085417b\") " pod="openstack/nova-cell1-db-create-ln2s2" Feb 18 09:19:12 crc kubenswrapper[4556]: I0218 09:19:12.094662 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-5bef-account-create-update-g2mzr" Feb 18 09:19:12 crc kubenswrapper[4556]: I0218 09:19:12.110989 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9jd5\" (UniqueName: \"kubernetes.io/projected/9d482227-d966-4319-bcfd-10944085417b-kube-api-access-x9jd5\") pod \"nova-cell1-db-create-ln2s2\" (UID: \"9d482227-d966-4319-bcfd-10944085417b\") " pod="openstack/nova-cell1-db-create-ln2s2" Feb 18 09:19:12 crc kubenswrapper[4556]: I0218 09:19:12.168963 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-2c5c-account-create-update-7z6nx"] Feb 18 09:19:12 crc kubenswrapper[4556]: I0218 09:19:12.170444 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-2c5c-account-create-update-7z6nx" Feb 18 09:19:12 crc kubenswrapper[4556]: I0218 09:19:12.172844 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Feb 18 09:19:12 crc kubenswrapper[4556]: I0218 09:19:12.179443 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-2c5c-account-create-update-7z6nx"] Feb 18 09:19:12 crc kubenswrapper[4556]: I0218 09:19:12.189193 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3bab749-15ef-4b1d-a9ab-f51acb445a96-operator-scripts\") pod \"nova-cell0-6c31-account-create-update-2wxx2\" (UID: \"b3bab749-15ef-4b1d-a9ab-f51acb445a96\") " pod="openstack/nova-cell0-6c31-account-create-update-2wxx2" Feb 18 09:19:12 crc kubenswrapper[4556]: I0218 09:19:12.189502 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qf8rk\" (UniqueName: \"kubernetes.io/projected/b3bab749-15ef-4b1d-a9ab-f51acb445a96-kube-api-access-qf8rk\") pod \"nova-cell0-6c31-account-create-update-2wxx2\" (UID: \"b3bab749-15ef-4b1d-a9ab-f51acb445a96\") " pod="openstack/nova-cell0-6c31-account-create-update-2wxx2" Feb 18 09:19:12 crc kubenswrapper[4556]: I0218 09:19:12.191179 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3bab749-15ef-4b1d-a9ab-f51acb445a96-operator-scripts\") pod \"nova-cell0-6c31-account-create-update-2wxx2\" (UID: \"b3bab749-15ef-4b1d-a9ab-f51acb445a96\") " pod="openstack/nova-cell0-6c31-account-create-update-2wxx2" Feb 18 09:19:12 crc kubenswrapper[4556]: I0218 09:19:12.192559 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-ln2s2" Feb 18 09:19:12 crc kubenswrapper[4556]: I0218 09:19:12.210906 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qf8rk\" (UniqueName: \"kubernetes.io/projected/b3bab749-15ef-4b1d-a9ab-f51acb445a96-kube-api-access-qf8rk\") pod \"nova-cell0-6c31-account-create-update-2wxx2\" (UID: \"b3bab749-15ef-4b1d-a9ab-f51acb445a96\") " pod="openstack/nova-cell0-6c31-account-create-update-2wxx2" Feb 18 09:19:12 crc kubenswrapper[4556]: I0218 09:19:12.292754 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5zs4\" (UniqueName: \"kubernetes.io/projected/d89c1ba7-214f-45af-8a78-4f52474b16f7-kube-api-access-t5zs4\") pod \"nova-cell1-2c5c-account-create-update-7z6nx\" (UID: \"d89c1ba7-214f-45af-8a78-4f52474b16f7\") " pod="openstack/nova-cell1-2c5c-account-create-update-7z6nx" Feb 18 09:19:12 crc kubenswrapper[4556]: I0218 09:19:12.292823 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d89c1ba7-214f-45af-8a78-4f52474b16f7-operator-scripts\") pod \"nova-cell1-2c5c-account-create-update-7z6nx\" (UID: \"d89c1ba7-214f-45af-8a78-4f52474b16f7\") " pod="openstack/nova-cell1-2c5c-account-create-update-7z6nx" Feb 18 09:19:12 crc kubenswrapper[4556]: I0218 09:19:12.309297 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-6c31-account-create-update-2wxx2" Feb 18 09:19:12 crc kubenswrapper[4556]: I0218 09:19:12.394601 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d89c1ba7-214f-45af-8a78-4f52474b16f7-operator-scripts\") pod \"nova-cell1-2c5c-account-create-update-7z6nx\" (UID: \"d89c1ba7-214f-45af-8a78-4f52474b16f7\") " pod="openstack/nova-cell1-2c5c-account-create-update-7z6nx" Feb 18 09:19:12 crc kubenswrapper[4556]: I0218 09:19:12.395072 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5zs4\" (UniqueName: \"kubernetes.io/projected/d89c1ba7-214f-45af-8a78-4f52474b16f7-kube-api-access-t5zs4\") pod \"nova-cell1-2c5c-account-create-update-7z6nx\" (UID: \"d89c1ba7-214f-45af-8a78-4f52474b16f7\") " pod="openstack/nova-cell1-2c5c-account-create-update-7z6nx" Feb 18 09:19:12 crc kubenswrapper[4556]: I0218 09:19:12.400141 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d89c1ba7-214f-45af-8a78-4f52474b16f7-operator-scripts\") pod \"nova-cell1-2c5c-account-create-update-7z6nx\" (UID: \"d89c1ba7-214f-45af-8a78-4f52474b16f7\") " pod="openstack/nova-cell1-2c5c-account-create-update-7z6nx" Feb 18 09:19:12 crc kubenswrapper[4556]: I0218 09:19:12.412100 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5zs4\" (UniqueName: \"kubernetes.io/projected/d89c1ba7-214f-45af-8a78-4f52474b16f7-kube-api-access-t5zs4\") pod \"nova-cell1-2c5c-account-create-update-7z6nx\" (UID: \"d89c1ba7-214f-45af-8a78-4f52474b16f7\") " pod="openstack/nova-cell1-2c5c-account-create-update-7z6nx" Feb 18 09:19:12 crc kubenswrapper[4556]: I0218 09:19:12.436354 4556 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rmt7n" podUID="0a557635-d67d-4e93-a82c-990c34fe7d48" containerName="registry-server" probeResult="failure" output=< Feb 18 09:19:12 crc kubenswrapper[4556]: timeout: failed to connect service ":50051" within 1s Feb 18 09:19:12 crc kubenswrapper[4556]: > Feb 18 09:19:12 crc kubenswrapper[4556]: I0218 09:19:12.503852 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-f7xdx"] Feb 18 09:19:12 crc kubenswrapper[4556]: I0218 09:19:12.556021 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-2c5c-account-create-update-7z6nx" Feb 18 09:19:12 crc kubenswrapper[4556]: I0218 09:19:12.660132 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-nq759"] Feb 18 09:19:12 crc kubenswrapper[4556]: W0218 09:19:12.685281 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod65bf8803_f4f0_4366_8f1f_d894f2f407e2.slice/crio-798affd86559ea714ab4872cb270f39b9b32b2ccec55f6c813766c29ae1b8ceb WatchSource:0}: Error finding container 798affd86559ea714ab4872cb270f39b9b32b2ccec55f6c813766c29ae1b8ceb: Status 404 returned error can't find the container with id 798affd86559ea714ab4872cb270f39b9b32b2ccec55f6c813766c29ae1b8ceb Feb 18 09:19:12 crc kubenswrapper[4556]: I0218 09:19:12.767122 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-ln2s2"] Feb 18 09:19:12 crc kubenswrapper[4556]: W0218 09:19:12.776781 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9d482227_d966_4319_bcfd_10944085417b.slice/crio-e0708635a045706bf17a0d78a812d1df1d813b5a4c0ca4897ef559b2aa23945c WatchSource:0}: Error finding container e0708635a045706bf17a0d78a812d1df1d813b5a4c0ca4897ef559b2aa23945c: Status 404 returned error can't find the container with id e0708635a045706bf17a0d78a812d1df1d813b5a4c0ca4897ef559b2aa23945c Feb 18 09:19:12 crc kubenswrapper[4556]: I0218 09:19:12.804083 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"40b701d5-4495-47c3-bb6e-6ba466958355","Type":"ContainerStarted","Data":"1dda491fdb344030848ca5df7d5c32e96139ad7022a869bb7c2f0fc02c1bb6ed"} Feb 18 09:19:12 crc kubenswrapper[4556]: I0218 09:19:12.804592 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 18 09:19:12 crc kubenswrapper[4556]: I0218 09:19:12.811194 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-f7xdx" event={"ID":"96a4b91d-0008-4af7-8522-6842a8c627a8","Type":"ContainerStarted","Data":"e759ce739b673f51aade7482b8cc97d043d1849906a5e5a2a1550260f264e5e4"} Feb 18 09:19:12 crc kubenswrapper[4556]: I0218 09:19:12.819316 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-nq759" event={"ID":"65bf8803-f4f0-4366-8f1f-d894f2f407e2","Type":"ContainerStarted","Data":"798affd86559ea714ab4872cb270f39b9b32b2ccec55f6c813766c29ae1b8ceb"} Feb 18 09:19:12 crc kubenswrapper[4556]: I0218 09:19:12.821576 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-5bef-account-create-update-g2mzr"] Feb 18 09:19:12 crc kubenswrapper[4556]: I0218 09:19:12.880614 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.987380193 podStartE2EDuration="5.880595192s" podCreationTimestamp="2026-02-18 09:19:07 +0000 UTC" firstStartedPulling="2026-02-18 09:19:08.540483782 +0000 UTC m=+905.557444762" lastFinishedPulling="2026-02-18 09:19:12.433698781 +0000 UTC m=+909.450659761" observedRunningTime="2026-02-18 09:19:12.841722449 +0000 UTC m=+909.858683430" watchObservedRunningTime="2026-02-18 09:19:12.880595192 +0000 UTC m=+909.897556172" Feb 18 09:19:12 crc kubenswrapper[4556]: I0218 09:19:12.964127 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-6c31-account-create-update-2wxx2"] Feb 18 09:19:13 crc kubenswrapper[4556]: I0218 09:19:13.147011 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-2c5c-account-create-update-7z6nx"] Feb 18 09:19:13 crc kubenswrapper[4556]: W0218 09:19:13.159088 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd89c1ba7_214f_45af_8a78_4f52474b16f7.slice/crio-9d6714c1d3f29b78c31d4bf6ef407ca521ace7e68760f03a8f124ce859d9d447 WatchSource:0}: Error finding container 9d6714c1d3f29b78c31d4bf6ef407ca521ace7e68760f03a8f124ce859d9d447: Status 404 returned error can't find the container with id 9d6714c1d3f29b78c31d4bf6ef407ca521ace7e68760f03a8f124ce859d9d447 Feb 18 09:19:13 crc kubenswrapper[4556]: I0218 09:19:13.596400 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 18 09:19:13 crc kubenswrapper[4556]: I0218 09:19:13.596881 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9" containerName="glance-log" containerID="cri-o://c31bb188cd36a64068b7d79dd9886e17c0884ce62c635deb9aaf356133dad0ae" gracePeriod=30 Feb 18 09:19:13 crc kubenswrapper[4556]: I0218 09:19:13.597005 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9" containerName="glance-httpd" containerID="cri-o://4f89dee65414ab2d433b8d26869688906b5bc9fad55290401fe98c3fb76403fd" gracePeriod=30 Feb 18 09:19:13 crc kubenswrapper[4556]: I0218 09:19:13.827621 4556 generic.go:334] "Generic (PLEG): container finished" podID="b3bab749-15ef-4b1d-a9ab-f51acb445a96" containerID="c94cc4c9927e86110534aaf1c83ff7cd2ee6e8e5117aa4499d7c535f0ad637de" exitCode=0 Feb 18 09:19:13 crc kubenswrapper[4556]: I0218 09:19:13.827681 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-6c31-account-create-update-2wxx2" event={"ID":"b3bab749-15ef-4b1d-a9ab-f51acb445a96","Type":"ContainerDied","Data":"c94cc4c9927e86110534aaf1c83ff7cd2ee6e8e5117aa4499d7c535f0ad637de"} Feb 18 09:19:13 crc kubenswrapper[4556]: I0218 09:19:13.827710 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-6c31-account-create-update-2wxx2" event={"ID":"b3bab749-15ef-4b1d-a9ab-f51acb445a96","Type":"ContainerStarted","Data":"b1bfc7a1ba809de5c18f82922667053ca1e2a28941c9905eb811695f8f553d40"} Feb 18 09:19:13 crc kubenswrapper[4556]: I0218 09:19:13.833272 4556 generic.go:334] "Generic (PLEG): container finished" podID="65bf8803-f4f0-4366-8f1f-d894f2f407e2" containerID="d53f239a31c608a32853374e8528b40c92c47ff4879c4c4ffac72086bcf2cbc9" exitCode=0 Feb 18 09:19:13 crc kubenswrapper[4556]: I0218 09:19:13.833390 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-nq759" event={"ID":"65bf8803-f4f0-4366-8f1f-d894f2f407e2","Type":"ContainerDied","Data":"d53f239a31c608a32853374e8528b40c92c47ff4879c4c4ffac72086bcf2cbc9"} Feb 18 09:19:13 crc kubenswrapper[4556]: I0218 09:19:13.837842 4556 generic.go:334] "Generic (PLEG): container finished" podID="d89c1ba7-214f-45af-8a78-4f52474b16f7" containerID="21545e68d19b79db777bb501d6ab6fae34c8d8b1e4b0d8e7049d0803a0b1d066" exitCode=0 Feb 18 09:19:13 crc kubenswrapper[4556]: I0218 09:19:13.837944 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-2c5c-account-create-update-7z6nx" event={"ID":"d89c1ba7-214f-45af-8a78-4f52474b16f7","Type":"ContainerDied","Data":"21545e68d19b79db777bb501d6ab6fae34c8d8b1e4b0d8e7049d0803a0b1d066"} Feb 18 09:19:13 crc kubenswrapper[4556]: I0218 09:19:13.837978 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-2c5c-account-create-update-7z6nx" event={"ID":"d89c1ba7-214f-45af-8a78-4f52474b16f7","Type":"ContainerStarted","Data":"9d6714c1d3f29b78c31d4bf6ef407ca521ace7e68760f03a8f124ce859d9d447"} Feb 18 09:19:13 crc kubenswrapper[4556]: I0218 09:19:13.843517 4556 generic.go:334] "Generic (PLEG): container finished" podID="7da9e29c-e9f2-49bf-a3b3-6e1b764c2b78" containerID="f8c3715c17081b2315cdcf8d51278635a41a1190d1e5b321467b1cf3d6a471b6" exitCode=0 Feb 18 09:19:13 crc kubenswrapper[4556]: I0218 09:19:13.843752 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-5bef-account-create-update-g2mzr" event={"ID":"7da9e29c-e9f2-49bf-a3b3-6e1b764c2b78","Type":"ContainerDied","Data":"f8c3715c17081b2315cdcf8d51278635a41a1190d1e5b321467b1cf3d6a471b6"} Feb 18 09:19:13 crc kubenswrapper[4556]: I0218 09:19:13.843799 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-5bef-account-create-update-g2mzr" event={"ID":"7da9e29c-e9f2-49bf-a3b3-6e1b764c2b78","Type":"ContainerStarted","Data":"a8ef86948e0f7b35b6ac6759df067a677ca7898da3b68d89c20f489cf3949f07"} Feb 18 09:19:13 crc kubenswrapper[4556]: I0218 09:19:13.845131 4556 generic.go:334] "Generic (PLEG): container finished" podID="9d482227-d966-4319-bcfd-10944085417b" containerID="11a37ae750e25db51acdd71f8efb9605029d7f4853cb1d8aed87ad1132925aac" exitCode=0 Feb 18 09:19:13 crc kubenswrapper[4556]: I0218 09:19:13.845189 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-ln2s2" event={"ID":"9d482227-d966-4319-bcfd-10944085417b","Type":"ContainerDied","Data":"11a37ae750e25db51acdd71f8efb9605029d7f4853cb1d8aed87ad1132925aac"} Feb 18 09:19:13 crc kubenswrapper[4556]: I0218 09:19:13.845206 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-ln2s2" event={"ID":"9d482227-d966-4319-bcfd-10944085417b","Type":"ContainerStarted","Data":"e0708635a045706bf17a0d78a812d1df1d813b5a4c0ca4897ef559b2aa23945c"} Feb 18 09:19:13 crc kubenswrapper[4556]: I0218 09:19:13.850137 4556 generic.go:334] "Generic (PLEG): container finished" podID="2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9" containerID="c31bb188cd36a64068b7d79dd9886e17c0884ce62c635deb9aaf356133dad0ae" exitCode=143 Feb 18 09:19:13 crc kubenswrapper[4556]: I0218 09:19:13.850206 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9","Type":"ContainerDied","Data":"c31bb188cd36a64068b7d79dd9886e17c0884ce62c635deb9aaf356133dad0ae"} Feb 18 09:19:13 crc kubenswrapper[4556]: I0218 09:19:13.852233 4556 generic.go:334] "Generic (PLEG): container finished" podID="96a4b91d-0008-4af7-8522-6842a8c627a8" containerID="69ace43bee98401345011fa1708b2b02040a330a77aef476145808875a5f4267" exitCode=0 Feb 18 09:19:13 crc kubenswrapper[4556]: I0218 09:19:13.852748 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-f7xdx" event={"ID":"96a4b91d-0008-4af7-8522-6842a8c627a8","Type":"ContainerDied","Data":"69ace43bee98401345011fa1708b2b02040a330a77aef476145808875a5f4267"} Feb 18 09:19:14 crc kubenswrapper[4556]: I0218 09:19:14.433574 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 18 09:19:14 crc kubenswrapper[4556]: I0218 09:19:14.434123 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="35011d99-3341-4081-9733-a6f42f69a6d2" containerName="glance-log" containerID="cri-o://4f6a9471fabb777523494b87d377f21524f2dc8cb583c2a3116675d5d8900f75" gracePeriod=30 Feb 18 09:19:14 crc kubenswrapper[4556]: I0218 09:19:14.434264 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="35011d99-3341-4081-9733-a6f42f69a6d2" containerName="glance-httpd" containerID="cri-o://e18dd7ef339961839ac3bcbecba5e030be1878f9424536354a1a994260d68983" gracePeriod=30 Feb 18 09:19:14 crc kubenswrapper[4556]: I0218 09:19:14.863867 4556 generic.go:334] "Generic (PLEG): container finished" podID="35011d99-3341-4081-9733-a6f42f69a6d2" containerID="4f6a9471fabb777523494b87d377f21524f2dc8cb583c2a3116675d5d8900f75" exitCode=143 Feb 18 09:19:14 crc kubenswrapper[4556]: I0218 09:19:14.864773 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"35011d99-3341-4081-9733-a6f42f69a6d2","Type":"ContainerDied","Data":"4f6a9471fabb777523494b87d377f21524f2dc8cb583c2a3116675d5d8900f75"} Feb 18 09:19:15 crc kubenswrapper[4556]: I0218 09:19:15.326732 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-5bef-account-create-update-g2mzr" Feb 18 09:19:15 crc kubenswrapper[4556]: I0218 09:19:15.474207 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k6jcn\" (UniqueName: \"kubernetes.io/projected/7da9e29c-e9f2-49bf-a3b3-6e1b764c2b78-kube-api-access-k6jcn\") pod \"7da9e29c-e9f2-49bf-a3b3-6e1b764c2b78\" (UID: \"7da9e29c-e9f2-49bf-a3b3-6e1b764c2b78\") " Feb 18 09:19:15 crc kubenswrapper[4556]: I0218 09:19:15.474317 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7da9e29c-e9f2-49bf-a3b3-6e1b764c2b78-operator-scripts\") pod \"7da9e29c-e9f2-49bf-a3b3-6e1b764c2b78\" (UID: \"7da9e29c-e9f2-49bf-a3b3-6e1b764c2b78\") " Feb 18 09:19:15 crc kubenswrapper[4556]: I0218 09:19:15.474872 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7da9e29c-e9f2-49bf-a3b3-6e1b764c2b78-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7da9e29c-e9f2-49bf-a3b3-6e1b764c2b78" (UID: "7da9e29c-e9f2-49bf-a3b3-6e1b764c2b78"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:19:15 crc kubenswrapper[4556]: I0218 09:19:15.475844 4556 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7da9e29c-e9f2-49bf-a3b3-6e1b764c2b78-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:15 crc kubenswrapper[4556]: I0218 09:19:15.476232 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-ln2s2" Feb 18 09:19:15 crc kubenswrapper[4556]: I0218 09:19:15.480141 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7da9e29c-e9f2-49bf-a3b3-6e1b764c2b78-kube-api-access-k6jcn" (OuterVolumeSpecName: "kube-api-access-k6jcn") pod "7da9e29c-e9f2-49bf-a3b3-6e1b764c2b78" (UID: "7da9e29c-e9f2-49bf-a3b3-6e1b764c2b78"). InnerVolumeSpecName "kube-api-access-k6jcn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:19:15 crc kubenswrapper[4556]: I0218 09:19:15.482472 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-6c31-account-create-update-2wxx2" Feb 18 09:19:15 crc kubenswrapper[4556]: I0218 09:19:15.489105 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-f7xdx" Feb 18 09:19:15 crc kubenswrapper[4556]: I0218 09:19:15.494444 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-nq759" Feb 18 09:19:15 crc kubenswrapper[4556]: I0218 09:19:15.508089 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-2c5c-account-create-update-7z6nx" Feb 18 09:19:15 crc kubenswrapper[4556]: I0218 09:19:15.576813 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x9jd5\" (UniqueName: \"kubernetes.io/projected/9d482227-d966-4319-bcfd-10944085417b-kube-api-access-x9jd5\") pod \"9d482227-d966-4319-bcfd-10944085417b\" (UID: \"9d482227-d966-4319-bcfd-10944085417b\") " Feb 18 09:19:15 crc kubenswrapper[4556]: I0218 09:19:15.576992 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9d482227-d966-4319-bcfd-10944085417b-operator-scripts\") pod \"9d482227-d966-4319-bcfd-10944085417b\" (UID: \"9d482227-d966-4319-bcfd-10944085417b\") " Feb 18 09:19:15 crc kubenswrapper[4556]: I0218 09:19:15.577589 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d482227-d966-4319-bcfd-10944085417b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9d482227-d966-4319-bcfd-10944085417b" (UID: "9d482227-d966-4319-bcfd-10944085417b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:19:15 crc kubenswrapper[4556]: I0218 09:19:15.580237 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d482227-d966-4319-bcfd-10944085417b-kube-api-access-x9jd5" (OuterVolumeSpecName: "kube-api-access-x9jd5") pod "9d482227-d966-4319-bcfd-10944085417b" (UID: "9d482227-d966-4319-bcfd-10944085417b"). InnerVolumeSpecName "kube-api-access-x9jd5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:19:15 crc kubenswrapper[4556]: I0218 09:19:15.580600 4556 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9d482227-d966-4319-bcfd-10944085417b-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:15 crc kubenswrapper[4556]: I0218 09:19:15.580621 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k6jcn\" (UniqueName: \"kubernetes.io/projected/7da9e29c-e9f2-49bf-a3b3-6e1b764c2b78-kube-api-access-k6jcn\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:15 crc kubenswrapper[4556]: I0218 09:19:15.580634 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x9jd5\" (UniqueName: \"kubernetes.io/projected/9d482227-d966-4319-bcfd-10944085417b-kube-api-access-x9jd5\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:15 crc kubenswrapper[4556]: I0218 09:19:15.682145 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d89c1ba7-214f-45af-8a78-4f52474b16f7-operator-scripts\") pod \"d89c1ba7-214f-45af-8a78-4f52474b16f7\" (UID: \"d89c1ba7-214f-45af-8a78-4f52474b16f7\") " Feb 18 09:19:15 crc kubenswrapper[4556]: I0218 09:19:15.682237 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mx8c2\" (UniqueName: \"kubernetes.io/projected/65bf8803-f4f0-4366-8f1f-d894f2f407e2-kube-api-access-mx8c2\") pod \"65bf8803-f4f0-4366-8f1f-d894f2f407e2\" (UID: \"65bf8803-f4f0-4366-8f1f-d894f2f407e2\") " Feb 18 09:19:15 crc kubenswrapper[4556]: I0218 09:19:15.682269 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t5zs4\" (UniqueName: \"kubernetes.io/projected/d89c1ba7-214f-45af-8a78-4f52474b16f7-kube-api-access-t5zs4\") pod \"d89c1ba7-214f-45af-8a78-4f52474b16f7\" (UID: \"d89c1ba7-214f-45af-8a78-4f52474b16f7\") " Feb 18 09:19:15 crc kubenswrapper[4556]: I0218 09:19:15.682398 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/96a4b91d-0008-4af7-8522-6842a8c627a8-operator-scripts\") pod \"96a4b91d-0008-4af7-8522-6842a8c627a8\" (UID: \"96a4b91d-0008-4af7-8522-6842a8c627a8\") " Feb 18 09:19:15 crc kubenswrapper[4556]: I0218 09:19:15.682443 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d89c1ba7-214f-45af-8a78-4f52474b16f7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d89c1ba7-214f-45af-8a78-4f52474b16f7" (UID: "d89c1ba7-214f-45af-8a78-4f52474b16f7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:19:15 crc kubenswrapper[4556]: I0218 09:19:15.682452 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3bab749-15ef-4b1d-a9ab-f51acb445a96-operator-scripts\") pod \"b3bab749-15ef-4b1d-a9ab-f51acb445a96\" (UID: \"b3bab749-15ef-4b1d-a9ab-f51acb445a96\") " Feb 18 09:19:15 crc kubenswrapper[4556]: I0218 09:19:15.682517 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/65bf8803-f4f0-4366-8f1f-d894f2f407e2-operator-scripts\") pod \"65bf8803-f4f0-4366-8f1f-d894f2f407e2\" (UID: \"65bf8803-f4f0-4366-8f1f-d894f2f407e2\") " Feb 18 09:19:15 crc kubenswrapper[4556]: I0218 09:19:15.682543 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xxkpr\" (UniqueName: \"kubernetes.io/projected/96a4b91d-0008-4af7-8522-6842a8c627a8-kube-api-access-xxkpr\") pod \"96a4b91d-0008-4af7-8522-6842a8c627a8\" (UID: \"96a4b91d-0008-4af7-8522-6842a8c627a8\") " Feb 18 09:19:15 crc kubenswrapper[4556]: I0218 09:19:15.682599 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qf8rk\" (UniqueName: \"kubernetes.io/projected/b3bab749-15ef-4b1d-a9ab-f51acb445a96-kube-api-access-qf8rk\") pod \"b3bab749-15ef-4b1d-a9ab-f51acb445a96\" (UID: \"b3bab749-15ef-4b1d-a9ab-f51acb445a96\") " Feb 18 09:19:15 crc kubenswrapper[4556]: I0218 09:19:15.682721 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3bab749-15ef-4b1d-a9ab-f51acb445a96-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b3bab749-15ef-4b1d-a9ab-f51acb445a96" (UID: "b3bab749-15ef-4b1d-a9ab-f51acb445a96"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:19:15 crc kubenswrapper[4556]: I0218 09:19:15.683229 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/65bf8803-f4f0-4366-8f1f-d894f2f407e2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "65bf8803-f4f0-4366-8f1f-d894f2f407e2" (UID: "65bf8803-f4f0-4366-8f1f-d894f2f407e2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:19:15 crc kubenswrapper[4556]: I0218 09:19:15.683387 4556 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3bab749-15ef-4b1d-a9ab-f51acb445a96-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:15 crc kubenswrapper[4556]: I0218 09:19:15.683528 4556 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d89c1ba7-214f-45af-8a78-4f52474b16f7-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:15 crc kubenswrapper[4556]: I0218 09:19:15.684044 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96a4b91d-0008-4af7-8522-6842a8c627a8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "96a4b91d-0008-4af7-8522-6842a8c627a8" (UID: "96a4b91d-0008-4af7-8522-6842a8c627a8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:19:15 crc kubenswrapper[4556]: I0218 09:19:15.688300 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3bab749-15ef-4b1d-a9ab-f51acb445a96-kube-api-access-qf8rk" (OuterVolumeSpecName: "kube-api-access-qf8rk") pod "b3bab749-15ef-4b1d-a9ab-f51acb445a96" (UID: "b3bab749-15ef-4b1d-a9ab-f51acb445a96"). InnerVolumeSpecName "kube-api-access-qf8rk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:19:15 crc kubenswrapper[4556]: I0218 09:19:15.688408 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96a4b91d-0008-4af7-8522-6842a8c627a8-kube-api-access-xxkpr" (OuterVolumeSpecName: "kube-api-access-xxkpr") pod "96a4b91d-0008-4af7-8522-6842a8c627a8" (UID: "96a4b91d-0008-4af7-8522-6842a8c627a8"). InnerVolumeSpecName "kube-api-access-xxkpr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:19:15 crc kubenswrapper[4556]: I0218 09:19:15.702307 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65bf8803-f4f0-4366-8f1f-d894f2f407e2-kube-api-access-mx8c2" (OuterVolumeSpecName: "kube-api-access-mx8c2") pod "65bf8803-f4f0-4366-8f1f-d894f2f407e2" (UID: "65bf8803-f4f0-4366-8f1f-d894f2f407e2"). InnerVolumeSpecName "kube-api-access-mx8c2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:19:15 crc kubenswrapper[4556]: I0218 09:19:15.702346 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d89c1ba7-214f-45af-8a78-4f52474b16f7-kube-api-access-t5zs4" (OuterVolumeSpecName: "kube-api-access-t5zs4") pod "d89c1ba7-214f-45af-8a78-4f52474b16f7" (UID: "d89c1ba7-214f-45af-8a78-4f52474b16f7"). InnerVolumeSpecName "kube-api-access-t5zs4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:19:15 crc kubenswrapper[4556]: I0218 09:19:15.784774 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mx8c2\" (UniqueName: \"kubernetes.io/projected/65bf8803-f4f0-4366-8f1f-d894f2f407e2-kube-api-access-mx8c2\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:15 crc kubenswrapper[4556]: I0218 09:19:15.784802 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t5zs4\" (UniqueName: \"kubernetes.io/projected/d89c1ba7-214f-45af-8a78-4f52474b16f7-kube-api-access-t5zs4\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:15 crc kubenswrapper[4556]: I0218 09:19:15.784812 4556 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/96a4b91d-0008-4af7-8522-6842a8c627a8-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:15 crc kubenswrapper[4556]: I0218 09:19:15.784820 4556 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/65bf8803-f4f0-4366-8f1f-d894f2f407e2-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:15 crc kubenswrapper[4556]: I0218 09:19:15.784831 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xxkpr\" (UniqueName: \"kubernetes.io/projected/96a4b91d-0008-4af7-8522-6842a8c627a8-kube-api-access-xxkpr\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:15 crc kubenswrapper[4556]: I0218 09:19:15.784839 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qf8rk\" (UniqueName: \"kubernetes.io/projected/b3bab749-15ef-4b1d-a9ab-f51acb445a96-kube-api-access-qf8rk\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:15 crc kubenswrapper[4556]: I0218 09:19:15.879638 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-ln2s2" event={"ID":"9d482227-d966-4319-bcfd-10944085417b","Type":"ContainerDied","Data":"e0708635a045706bf17a0d78a812d1df1d813b5a4c0ca4897ef559b2aa23945c"} Feb 18 09:19:15 crc kubenswrapper[4556]: I0218 09:19:15.879687 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e0708635a045706bf17a0d78a812d1df1d813b5a4c0ca4897ef559b2aa23945c" Feb 18 09:19:15 crc kubenswrapper[4556]: I0218 09:19:15.879685 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-ln2s2" Feb 18 09:19:15 crc kubenswrapper[4556]: I0218 09:19:15.881724 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-f7xdx" event={"ID":"96a4b91d-0008-4af7-8522-6842a8c627a8","Type":"ContainerDied","Data":"e759ce739b673f51aade7482b8cc97d043d1849906a5e5a2a1550260f264e5e4"} Feb 18 09:19:15 crc kubenswrapper[4556]: I0218 09:19:15.881770 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e759ce739b673f51aade7482b8cc97d043d1849906a5e5a2a1550260f264e5e4" Feb 18 09:19:15 crc kubenswrapper[4556]: I0218 09:19:15.881835 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-f7xdx" Feb 18 09:19:15 crc kubenswrapper[4556]: I0218 09:19:15.906581 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-6c31-account-create-update-2wxx2" event={"ID":"b3bab749-15ef-4b1d-a9ab-f51acb445a96","Type":"ContainerDied","Data":"b1bfc7a1ba809de5c18f82922667053ca1e2a28941c9905eb811695f8f553d40"} Feb 18 09:19:15 crc kubenswrapper[4556]: I0218 09:19:15.906628 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b1bfc7a1ba809de5c18f82922667053ca1e2a28941c9905eb811695f8f553d40" Feb 18 09:19:15 crc kubenswrapper[4556]: I0218 09:19:15.906628 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-6c31-account-create-update-2wxx2" Feb 18 09:19:15 crc kubenswrapper[4556]: I0218 09:19:15.910293 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-nq759" event={"ID":"65bf8803-f4f0-4366-8f1f-d894f2f407e2","Type":"ContainerDied","Data":"798affd86559ea714ab4872cb270f39b9b32b2ccec55f6c813766c29ae1b8ceb"} Feb 18 09:19:15 crc kubenswrapper[4556]: I0218 09:19:15.910309 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-nq759" Feb 18 09:19:15 crc kubenswrapper[4556]: I0218 09:19:15.910325 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="798affd86559ea714ab4872cb270f39b9b32b2ccec55f6c813766c29ae1b8ceb" Feb 18 09:19:15 crc kubenswrapper[4556]: I0218 09:19:15.912723 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-2c5c-account-create-update-7z6nx" event={"ID":"d89c1ba7-214f-45af-8a78-4f52474b16f7","Type":"ContainerDied","Data":"9d6714c1d3f29b78c31d4bf6ef407ca521ace7e68760f03a8f124ce859d9d447"} Feb 18 09:19:15 crc kubenswrapper[4556]: I0218 09:19:15.912761 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9d6714c1d3f29b78c31d4bf6ef407ca521ace7e68760f03a8f124ce859d9d447" Feb 18 09:19:15 crc kubenswrapper[4556]: I0218 09:19:15.912810 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-2c5c-account-create-update-7z6nx" Feb 18 09:19:15 crc kubenswrapper[4556]: I0218 09:19:15.921454 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-5bef-account-create-update-g2mzr" event={"ID":"7da9e29c-e9f2-49bf-a3b3-6e1b764c2b78","Type":"ContainerDied","Data":"a8ef86948e0f7b35b6ac6759df067a677ca7898da3b68d89c20f489cf3949f07"} Feb 18 09:19:15 crc kubenswrapper[4556]: I0218 09:19:15.921480 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a8ef86948e0f7b35b6ac6759df067a677ca7898da3b68d89c20f489cf3949f07" Feb 18 09:19:15 crc kubenswrapper[4556]: I0218 09:19:15.921525 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-5bef-account-create-update-g2mzr" Feb 18 09:19:16 crc kubenswrapper[4556]: I0218 09:19:16.368737 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 18 09:19:16 crc kubenswrapper[4556]: I0218 09:19:16.369214 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="40b701d5-4495-47c3-bb6e-6ba466958355" containerName="proxy-httpd" containerID="cri-o://1dda491fdb344030848ca5df7d5c32e96139ad7022a869bb7c2f0fc02c1bb6ed" gracePeriod=30 Feb 18 09:19:16 crc kubenswrapper[4556]: I0218 09:19:16.369234 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="40b701d5-4495-47c3-bb6e-6ba466958355" containerName="ceilometer-notification-agent" containerID="cri-o://471e8d3bb0e3ae1e40387a1ab945b5b86c109e0aae7d0c0de060669dafb0b2e1" gracePeriod=30 Feb 18 09:19:16 crc kubenswrapper[4556]: I0218 09:19:16.369399 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="40b701d5-4495-47c3-bb6e-6ba466958355" containerName="sg-core" containerID="cri-o://f984aeec06953af9d589b06b93e201997457368d82ecc83ab082489d056e155f" gracePeriod=30 Feb 18 09:19:16 crc kubenswrapper[4556]: I0218 09:19:16.369565 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="40b701d5-4495-47c3-bb6e-6ba466958355" containerName="ceilometer-central-agent" containerID="cri-o://b0838fd427fe9647da3a2c765ef31d6f5e3b1d99df1a0a099104c13db47316a4" gracePeriod=30 Feb 18 09:19:16 crc kubenswrapper[4556]: I0218 09:19:16.732122 4556 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.153:9292/healthcheck\": read tcp 10.217.0.2:53346->10.217.0.153:9292: read: connection reset by peer" Feb 18 09:19:16 crc kubenswrapper[4556]: I0218 09:19:16.732139 4556 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.153:9292/healthcheck\": read tcp 10.217.0.2:53348->10.217.0.153:9292: read: connection reset by peer" Feb 18 09:19:16 crc kubenswrapper[4556]: I0218 09:19:16.931563 4556 generic.go:334] "Generic (PLEG): container finished" podID="40b701d5-4495-47c3-bb6e-6ba466958355" containerID="1dda491fdb344030848ca5df7d5c32e96139ad7022a869bb7c2f0fc02c1bb6ed" exitCode=0 Feb 18 09:19:16 crc kubenswrapper[4556]: I0218 09:19:16.931596 4556 generic.go:334] "Generic (PLEG): container finished" podID="40b701d5-4495-47c3-bb6e-6ba466958355" containerID="f984aeec06953af9d589b06b93e201997457368d82ecc83ab082489d056e155f" exitCode=2 Feb 18 09:19:16 crc kubenswrapper[4556]: I0218 09:19:16.931603 4556 generic.go:334] "Generic (PLEG): container finished" podID="40b701d5-4495-47c3-bb6e-6ba466958355" containerID="471e8d3bb0e3ae1e40387a1ab945b5b86c109e0aae7d0c0de060669dafb0b2e1" exitCode=0 Feb 18 09:19:16 crc kubenswrapper[4556]: I0218 09:19:16.931636 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"40b701d5-4495-47c3-bb6e-6ba466958355","Type":"ContainerDied","Data":"1dda491fdb344030848ca5df7d5c32e96139ad7022a869bb7c2f0fc02c1bb6ed"} Feb 18 09:19:16 crc kubenswrapper[4556]: I0218 09:19:16.931664 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"40b701d5-4495-47c3-bb6e-6ba466958355","Type":"ContainerDied","Data":"f984aeec06953af9d589b06b93e201997457368d82ecc83ab082489d056e155f"} Feb 18 09:19:16 crc kubenswrapper[4556]: I0218 09:19:16.931673 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"40b701d5-4495-47c3-bb6e-6ba466958355","Type":"ContainerDied","Data":"471e8d3bb0e3ae1e40387a1ab945b5b86c109e0aae7d0c0de060669dafb0b2e1"} Feb 18 09:19:16 crc kubenswrapper[4556]: I0218 09:19:16.934169 4556 generic.go:334] "Generic (PLEG): container finished" podID="2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9" containerID="4f89dee65414ab2d433b8d26869688906b5bc9fad55290401fe98c3fb76403fd" exitCode=0 Feb 18 09:19:16 crc kubenswrapper[4556]: I0218 09:19:16.934192 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9","Type":"ContainerDied","Data":"4f89dee65414ab2d433b8d26869688906b5bc9fad55290401fe98c3fb76403fd"} Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.179170 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-wxltl"] Feb 18 09:19:17 crc kubenswrapper[4556]: E0218 09:19:17.179743 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d482227-d966-4319-bcfd-10944085417b" containerName="mariadb-database-create" Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.179759 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d482227-d966-4319-bcfd-10944085417b" containerName="mariadb-database-create" Feb 18 09:19:17 crc kubenswrapper[4556]: E0218 09:19:17.179773 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3bab749-15ef-4b1d-a9ab-f51acb445a96" containerName="mariadb-account-create-update" Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.179779 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3bab749-15ef-4b1d-a9ab-f51acb445a96" containerName="mariadb-account-create-update" Feb 18 09:19:17 crc kubenswrapper[4556]: E0218 09:19:17.179791 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7da9e29c-e9f2-49bf-a3b3-6e1b764c2b78" containerName="mariadb-account-create-update" Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.179797 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="7da9e29c-e9f2-49bf-a3b3-6e1b764c2b78" containerName="mariadb-account-create-update" Feb 18 09:19:17 crc kubenswrapper[4556]: E0218 09:19:17.179810 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d89c1ba7-214f-45af-8a78-4f52474b16f7" containerName="mariadb-account-create-update" Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.179815 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="d89c1ba7-214f-45af-8a78-4f52474b16f7" containerName="mariadb-account-create-update" Feb 18 09:19:17 crc kubenswrapper[4556]: E0218 09:19:17.179828 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65bf8803-f4f0-4366-8f1f-d894f2f407e2" containerName="mariadb-database-create" Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.179833 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="65bf8803-f4f0-4366-8f1f-d894f2f407e2" containerName="mariadb-database-create" Feb 18 09:19:17 crc kubenswrapper[4556]: E0218 09:19:17.179844 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96a4b91d-0008-4af7-8522-6842a8c627a8" containerName="mariadb-database-create" Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.179857 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="96a4b91d-0008-4af7-8522-6842a8c627a8" containerName="mariadb-database-create" Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.180035 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="7da9e29c-e9f2-49bf-a3b3-6e1b764c2b78" containerName="mariadb-account-create-update" Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.180047 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3bab749-15ef-4b1d-a9ab-f51acb445a96" containerName="mariadb-account-create-update" Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.180058 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d482227-d966-4319-bcfd-10944085417b" containerName="mariadb-database-create" Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.180067 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="d89c1ba7-214f-45af-8a78-4f52474b16f7" containerName="mariadb-account-create-update" Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.180077 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="96a4b91d-0008-4af7-8522-6842a8c627a8" containerName="mariadb-database-create" Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.180083 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="65bf8803-f4f0-4366-8f1f-d894f2f407e2" containerName="mariadb-database-create" Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.182643 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-wxltl" Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.189687 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-wwj98" Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.193974 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.194137 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.220203 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-wxltl"] Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.222272 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4a792c8-97e0-46f2-8a40-747a7cae58f5-scripts\") pod \"nova-cell0-conductor-db-sync-wxltl\" (UID: \"f4a792c8-97e0-46f2-8a40-747a7cae58f5\") " pod="openstack/nova-cell0-conductor-db-sync-wxltl" Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.222322 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkkw5\" (UniqueName: \"kubernetes.io/projected/f4a792c8-97e0-46f2-8a40-747a7cae58f5-kube-api-access-kkkw5\") pod \"nova-cell0-conductor-db-sync-wxltl\" (UID: \"f4a792c8-97e0-46f2-8a40-747a7cae58f5\") " pod="openstack/nova-cell0-conductor-db-sync-wxltl" Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.222375 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4a792c8-97e0-46f2-8a40-747a7cae58f5-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-wxltl\" (UID: \"f4a792c8-97e0-46f2-8a40-747a7cae58f5\") " pod="openstack/nova-cell0-conductor-db-sync-wxltl" Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.222536 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4a792c8-97e0-46f2-8a40-747a7cae58f5-config-data\") pod \"nova-cell0-conductor-db-sync-wxltl\" (UID: \"f4a792c8-97e0-46f2-8a40-747a7cae58f5\") " pod="openstack/nova-cell0-conductor-db-sync-wxltl" Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.282357 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.323535 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9\" (UID: \"2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9\") " Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.323638 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9-combined-ca-bundle\") pod \"2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9\" (UID: \"2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9\") " Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.323674 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9-public-tls-certs\") pod \"2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9\" (UID: \"2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9\") " Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.323745 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9-logs\") pod \"2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9\" (UID: \"2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9\") " Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.323775 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9-httpd-run\") pod \"2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9\" (UID: \"2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9\") " Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.323804 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9-scripts\") pod \"2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9\" (UID: \"2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9\") " Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.323886 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9-config-data\") pod \"2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9\" (UID: \"2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9\") " Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.323928 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gkcrq\" (UniqueName: \"kubernetes.io/projected/2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9-kube-api-access-gkcrq\") pod \"2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9\" (UID: \"2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9\") " Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.324162 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4a792c8-97e0-46f2-8a40-747a7cae58f5-config-data\") pod \"nova-cell0-conductor-db-sync-wxltl\" (UID: \"f4a792c8-97e0-46f2-8a40-747a7cae58f5\") " pod="openstack/nova-cell0-conductor-db-sync-wxltl" Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.329643 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9" (UID: "2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.331913 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4a792c8-97e0-46f2-8a40-747a7cae58f5-scripts\") pod \"nova-cell0-conductor-db-sync-wxltl\" (UID: \"f4a792c8-97e0-46f2-8a40-747a7cae58f5\") " pod="openstack/nova-cell0-conductor-db-sync-wxltl" Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.331977 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkkw5\" (UniqueName: \"kubernetes.io/projected/f4a792c8-97e0-46f2-8a40-747a7cae58f5-kube-api-access-kkkw5\") pod \"nova-cell0-conductor-db-sync-wxltl\" (UID: \"f4a792c8-97e0-46f2-8a40-747a7cae58f5\") " pod="openstack/nova-cell0-conductor-db-sync-wxltl" Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.332043 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4a792c8-97e0-46f2-8a40-747a7cae58f5-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-wxltl\" (UID: \"f4a792c8-97e0-46f2-8a40-747a7cae58f5\") " pod="openstack/nova-cell0-conductor-db-sync-wxltl" Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.332371 4556 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.333479 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9-logs" (OuterVolumeSpecName: "logs") pod "2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9" (UID: "2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.342496 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9-scripts" (OuterVolumeSpecName: "scripts") pod "2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9" (UID: "2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.343185 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9" (UID: "2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.346501 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9-kube-api-access-gkcrq" (OuterVolumeSpecName: "kube-api-access-gkcrq") pod "2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9" (UID: "2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9"). InnerVolumeSpecName "kube-api-access-gkcrq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.367766 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4a792c8-97e0-46f2-8a40-747a7cae58f5-scripts\") pod \"nova-cell0-conductor-db-sync-wxltl\" (UID: \"f4a792c8-97e0-46f2-8a40-747a7cae58f5\") " pod="openstack/nova-cell0-conductor-db-sync-wxltl" Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.371667 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4a792c8-97e0-46f2-8a40-747a7cae58f5-config-data\") pod \"nova-cell0-conductor-db-sync-wxltl\" (UID: \"f4a792c8-97e0-46f2-8a40-747a7cae58f5\") " pod="openstack/nova-cell0-conductor-db-sync-wxltl" Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.377177 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4a792c8-97e0-46f2-8a40-747a7cae58f5-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-wxltl\" (UID: \"f4a792c8-97e0-46f2-8a40-747a7cae58f5\") " pod="openstack/nova-cell0-conductor-db-sync-wxltl" Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.398233 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkkw5\" (UniqueName: \"kubernetes.io/projected/f4a792c8-97e0-46f2-8a40-747a7cae58f5-kube-api-access-kkkw5\") pod \"nova-cell0-conductor-db-sync-wxltl\" (UID: \"f4a792c8-97e0-46f2-8a40-747a7cae58f5\") " pod="openstack/nova-cell0-conductor-db-sync-wxltl" Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.434540 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gkcrq\" (UniqueName: \"kubernetes.io/projected/2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9-kube-api-access-gkcrq\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.434585 4556 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.434597 4556 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9-logs\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.434606 4556 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9-scripts\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.463269 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9" (UID: "2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.472502 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9-config-data" (OuterVolumeSpecName: "config-data") pod "2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9" (UID: "2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.493305 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9" (UID: "2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.496397 4556 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.536808 4556 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.536838 4556 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.536847 4556 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9-config-data\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.536856 4556 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.578701 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-wxltl" Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.946738 4556 generic.go:334] "Generic (PLEG): container finished" podID="35011d99-3341-4081-9733-a6f42f69a6d2" containerID="e18dd7ef339961839ac3bcbecba5e030be1878f9424536354a1a994260d68983" exitCode=0 Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.947383 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"35011d99-3341-4081-9733-a6f42f69a6d2","Type":"ContainerDied","Data":"e18dd7ef339961839ac3bcbecba5e030be1878f9424536354a1a994260d68983"} Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.952864 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9","Type":"ContainerDied","Data":"10752ccd372681e784cd9b6d7a7a940de3084c11a711f90554b2dd4b3a778e5a"} Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.952911 4556 scope.go:117] "RemoveContainer" containerID="4f89dee65414ab2d433b8d26869688906b5bc9fad55290401fe98c3fb76403fd" Feb 18 09:19:17 crc kubenswrapper[4556]: I0218 09:19:17.952964 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.019941 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.029482 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.030134 4556 scope.go:117] "RemoveContainer" containerID="c31bb188cd36a64068b7d79dd9886e17c0884ce62c635deb9aaf356133dad0ae" Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.039661 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 18 09:19:18 crc kubenswrapper[4556]: E0218 09:19:18.040005 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9" containerName="glance-httpd" Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.040024 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9" containerName="glance-httpd" Feb 18 09:19:18 crc kubenswrapper[4556]: E0218 09:19:18.040043 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9" containerName="glance-log" Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.040049 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9" containerName="glance-log" Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.040254 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9" containerName="glance-httpd" Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.040284 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9" containerName="glance-log" Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.041066 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.044534 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.046536 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.063200 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-wxltl"] Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.074228 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.107793 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.147517 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/23a483a0-fda5-4bf6-bfea-06a84d8d6533-scripts\") pod \"glance-default-external-api-0\" (UID: \"23a483a0-fda5-4bf6-bfea-06a84d8d6533\") " pod="openstack/glance-default-external-api-0" Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.147585 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23a483a0-fda5-4bf6-bfea-06a84d8d6533-config-data\") pod \"glance-default-external-api-0\" (UID: \"23a483a0-fda5-4bf6-bfea-06a84d8d6533\") " pod="openstack/glance-default-external-api-0" Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.154260 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"23a483a0-fda5-4bf6-bfea-06a84d8d6533\") " pod="openstack/glance-default-external-api-0" Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.154706 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4vpb\" (UniqueName: \"kubernetes.io/projected/23a483a0-fda5-4bf6-bfea-06a84d8d6533-kube-api-access-s4vpb\") pod \"glance-default-external-api-0\" (UID: \"23a483a0-fda5-4bf6-bfea-06a84d8d6533\") " pod="openstack/glance-default-external-api-0" Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.154814 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/23a483a0-fda5-4bf6-bfea-06a84d8d6533-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"23a483a0-fda5-4bf6-bfea-06a84d8d6533\") " pod="openstack/glance-default-external-api-0" Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.154865 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23a483a0-fda5-4bf6-bfea-06a84d8d6533-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"23a483a0-fda5-4bf6-bfea-06a84d8d6533\") " pod="openstack/glance-default-external-api-0" Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.154945 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23a483a0-fda5-4bf6-bfea-06a84d8d6533-logs\") pod \"glance-default-external-api-0\" (UID: \"23a483a0-fda5-4bf6-bfea-06a84d8d6533\") " pod="openstack/glance-default-external-api-0" Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.154979 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/23a483a0-fda5-4bf6-bfea-06a84d8d6533-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"23a483a0-fda5-4bf6-bfea-06a84d8d6533\") " pod="openstack/glance-default-external-api-0" Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.256430 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/35011d99-3341-4081-9733-a6f42f69a6d2-logs\") pod \"35011d99-3341-4081-9733-a6f42f69a6d2\" (UID: \"35011d99-3341-4081-9733-a6f42f69a6d2\") " Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.256492 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/35011d99-3341-4081-9733-a6f42f69a6d2-httpd-run\") pod \"35011d99-3341-4081-9733-a6f42f69a6d2\" (UID: \"35011d99-3341-4081-9733-a6f42f69a6d2\") " Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.256602 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35011d99-3341-4081-9733-a6f42f69a6d2-scripts\") pod \"35011d99-3341-4081-9733-a6f42f69a6d2\" (UID: \"35011d99-3341-4081-9733-a6f42f69a6d2\") " Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.256626 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"35011d99-3341-4081-9733-a6f42f69a6d2\" (UID: \"35011d99-3341-4081-9733-a6f42f69a6d2\") " Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.256664 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35011d99-3341-4081-9733-a6f42f69a6d2-combined-ca-bundle\") pod \"35011d99-3341-4081-9733-a6f42f69a6d2\" (UID: \"35011d99-3341-4081-9733-a6f42f69a6d2\") " Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.256689 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h98ck\" (UniqueName: \"kubernetes.io/projected/35011d99-3341-4081-9733-a6f42f69a6d2-kube-api-access-h98ck\") pod \"35011d99-3341-4081-9733-a6f42f69a6d2\" (UID: \"35011d99-3341-4081-9733-a6f42f69a6d2\") " Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.256763 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35011d99-3341-4081-9733-a6f42f69a6d2-config-data\") pod \"35011d99-3341-4081-9733-a6f42f69a6d2\" (UID: \"35011d99-3341-4081-9733-a6f42f69a6d2\") " Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.256873 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/35011d99-3341-4081-9733-a6f42f69a6d2-internal-tls-certs\") pod \"35011d99-3341-4081-9733-a6f42f69a6d2\" (UID: \"35011d99-3341-4081-9733-a6f42f69a6d2\") " Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.256930 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35011d99-3341-4081-9733-a6f42f69a6d2-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "35011d99-3341-4081-9733-a6f42f69a6d2" (UID: "35011d99-3341-4081-9733-a6f42f69a6d2"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.257000 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35011d99-3341-4081-9733-a6f42f69a6d2-logs" (OuterVolumeSpecName: "logs") pod "35011d99-3341-4081-9733-a6f42f69a6d2" (UID: "35011d99-3341-4081-9733-a6f42f69a6d2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.257111 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23a483a0-fda5-4bf6-bfea-06a84d8d6533-logs\") pod \"glance-default-external-api-0\" (UID: \"23a483a0-fda5-4bf6-bfea-06a84d8d6533\") " pod="openstack/glance-default-external-api-0" Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.257166 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/23a483a0-fda5-4bf6-bfea-06a84d8d6533-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"23a483a0-fda5-4bf6-bfea-06a84d8d6533\") " pod="openstack/glance-default-external-api-0" Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.257221 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/23a483a0-fda5-4bf6-bfea-06a84d8d6533-scripts\") pod \"glance-default-external-api-0\" (UID: \"23a483a0-fda5-4bf6-bfea-06a84d8d6533\") " pod="openstack/glance-default-external-api-0" Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.257251 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23a483a0-fda5-4bf6-bfea-06a84d8d6533-config-data\") pod \"glance-default-external-api-0\" (UID: \"23a483a0-fda5-4bf6-bfea-06a84d8d6533\") " pod="openstack/glance-default-external-api-0" Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.257280 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"23a483a0-fda5-4bf6-bfea-06a84d8d6533\") " pod="openstack/glance-default-external-api-0" Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.257371 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4vpb\" (UniqueName: \"kubernetes.io/projected/23a483a0-fda5-4bf6-bfea-06a84d8d6533-kube-api-access-s4vpb\") pod \"glance-default-external-api-0\" (UID: \"23a483a0-fda5-4bf6-bfea-06a84d8d6533\") " pod="openstack/glance-default-external-api-0" Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.257430 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/23a483a0-fda5-4bf6-bfea-06a84d8d6533-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"23a483a0-fda5-4bf6-bfea-06a84d8d6533\") " pod="openstack/glance-default-external-api-0" Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.257463 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23a483a0-fda5-4bf6-bfea-06a84d8d6533-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"23a483a0-fda5-4bf6-bfea-06a84d8d6533\") " pod="openstack/glance-default-external-api-0" Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.257510 4556 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/35011d99-3341-4081-9733-a6f42f69a6d2-logs\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.257519 4556 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/35011d99-3341-4081-9733-a6f42f69a6d2-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.257834 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23a483a0-fda5-4bf6-bfea-06a84d8d6533-logs\") pod \"glance-default-external-api-0\" (UID: \"23a483a0-fda5-4bf6-bfea-06a84d8d6533\") " pod="openstack/glance-default-external-api-0" Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.263313 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35011d99-3341-4081-9733-a6f42f69a6d2-scripts" (OuterVolumeSpecName: "scripts") pod "35011d99-3341-4081-9733-a6f42f69a6d2" (UID: "35011d99-3341-4081-9733-a6f42f69a6d2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.263595 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/23a483a0-fda5-4bf6-bfea-06a84d8d6533-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"23a483a0-fda5-4bf6-bfea-06a84d8d6533\") " pod="openstack/glance-default-external-api-0" Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.264029 4556 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"23a483a0-fda5-4bf6-bfea-06a84d8d6533\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-external-api-0" Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.272791 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/23a483a0-fda5-4bf6-bfea-06a84d8d6533-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"23a483a0-fda5-4bf6-bfea-06a84d8d6533\") " pod="openstack/glance-default-external-api-0" Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.274293 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35011d99-3341-4081-9733-a6f42f69a6d2-kube-api-access-h98ck" (OuterVolumeSpecName: "kube-api-access-h98ck") pod "35011d99-3341-4081-9733-a6f42f69a6d2" (UID: "35011d99-3341-4081-9733-a6f42f69a6d2"). InnerVolumeSpecName "kube-api-access-h98ck". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.276596 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23a483a0-fda5-4bf6-bfea-06a84d8d6533-config-data\") pod \"glance-default-external-api-0\" (UID: \"23a483a0-fda5-4bf6-bfea-06a84d8d6533\") " pod="openstack/glance-default-external-api-0" Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.278479 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23a483a0-fda5-4bf6-bfea-06a84d8d6533-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"23a483a0-fda5-4bf6-bfea-06a84d8d6533\") " pod="openstack/glance-default-external-api-0" Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.283603 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/23a483a0-fda5-4bf6-bfea-06a84d8d6533-scripts\") pod \"glance-default-external-api-0\" (UID: \"23a483a0-fda5-4bf6-bfea-06a84d8d6533\") " pod="openstack/glance-default-external-api-0" Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.293794 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4vpb\" (UniqueName: \"kubernetes.io/projected/23a483a0-fda5-4bf6-bfea-06a84d8d6533-kube-api-access-s4vpb\") pod \"glance-default-external-api-0\" (UID: \"23a483a0-fda5-4bf6-bfea-06a84d8d6533\") " pod="openstack/glance-default-external-api-0" Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.295803 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "35011d99-3341-4081-9733-a6f42f69a6d2" (UID: "35011d99-3341-4081-9733-a6f42f69a6d2"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.306059 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"23a483a0-fda5-4bf6-bfea-06a84d8d6533\") " pod="openstack/glance-default-external-api-0" Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.317788 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35011d99-3341-4081-9733-a6f42f69a6d2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "35011d99-3341-4081-9733-a6f42f69a6d2" (UID: "35011d99-3341-4081-9733-a6f42f69a6d2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.324713 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35011d99-3341-4081-9733-a6f42f69a6d2-config-data" (OuterVolumeSpecName: "config-data") pod "35011d99-3341-4081-9733-a6f42f69a6d2" (UID: "35011d99-3341-4081-9733-a6f42f69a6d2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.329098 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35011d99-3341-4081-9733-a6f42f69a6d2-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "35011d99-3341-4081-9733-a6f42f69a6d2" (UID: "35011d99-3341-4081-9733-a6f42f69a6d2"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.359722 4556 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35011d99-3341-4081-9733-a6f42f69a6d2-config-data\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.359754 4556 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/35011d99-3341-4081-9733-a6f42f69a6d2-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.359766 4556 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35011d99-3341-4081-9733-a6f42f69a6d2-scripts\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.359809 4556 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.359819 4556 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35011d99-3341-4081-9733-a6f42f69a6d2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.359828 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h98ck\" (UniqueName: \"kubernetes.io/projected/35011d99-3341-4081-9733-a6f42f69a6d2-kube-api-access-h98ck\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.373338 4556 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.385448 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.469117 4556 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.869826 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 18 09:19:18 crc kubenswrapper[4556]: W0218 09:19:18.883061 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod23a483a0_fda5_4bf6_bfea_06a84d8d6533.slice/crio-a23bd29ce39a73ac2ab3f83ae8853bb0acd211a313729bb93c6286c8e650bd11 WatchSource:0}: Error finding container a23bd29ce39a73ac2ab3f83ae8853bb0acd211a313729bb93c6286c8e650bd11: Status 404 returned error can't find the container with id a23bd29ce39a73ac2ab3f83ae8853bb0acd211a313729bb93c6286c8e650bd11 Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.974960 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"23a483a0-fda5-4bf6-bfea-06a84d8d6533","Type":"ContainerStarted","Data":"a23bd29ce39a73ac2ab3f83ae8853bb0acd211a313729bb93c6286c8e650bd11"} Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.976303 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-wxltl" event={"ID":"f4a792c8-97e0-46f2-8a40-747a7cae58f5","Type":"ContainerStarted","Data":"731029b0942c84e8e5914118f40931c2582da401dfac74804a29e88d44f46c29"} Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.981384 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"35011d99-3341-4081-9733-a6f42f69a6d2","Type":"ContainerDied","Data":"fbc7dd6512ba3c836f9289719c03dca46f287d8dfca505ce872676d47b0cab7f"} Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.981421 4556 scope.go:117] "RemoveContainer" containerID="e18dd7ef339961839ac3bcbecba5e030be1878f9424536354a1a994260d68983" Feb 18 09:19:18 crc kubenswrapper[4556]: I0218 09:19:18.981475 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.072344 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.078046 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.081425 4556 scope.go:117] "RemoveContainer" containerID="4f6a9471fabb777523494b87d377f21524f2dc8cb583c2a3116675d5d8900f75" Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.099233 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 18 09:19:19 crc kubenswrapper[4556]: E0218 09:19:19.099754 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35011d99-3341-4081-9733-a6f42f69a6d2" containerName="glance-httpd" Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.099770 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="35011d99-3341-4081-9733-a6f42f69a6d2" containerName="glance-httpd" Feb 18 09:19:19 crc kubenswrapper[4556]: E0218 09:19:19.099801 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35011d99-3341-4081-9733-a6f42f69a6d2" containerName="glance-log" Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.099807 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="35011d99-3341-4081-9733-a6f42f69a6d2" containerName="glance-log" Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.100084 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="35011d99-3341-4081-9733-a6f42f69a6d2" containerName="glance-httpd" Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.100101 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="35011d99-3341-4081-9733-a6f42f69a6d2" containerName="glance-log" Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.127488 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.149869 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.153756 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.164661 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.286763 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f50b4c03-2bce-417a-b01c-aaec09a5c8c3-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f50b4c03-2bce-417a-b01c-aaec09a5c8c3\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.287273 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"f50b4c03-2bce-417a-b01c-aaec09a5c8c3\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.287458 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f50b4c03-2bce-417a-b01c-aaec09a5c8c3-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f50b4c03-2bce-417a-b01c-aaec09a5c8c3\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.287488 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f50b4c03-2bce-417a-b01c-aaec09a5c8c3-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f50b4c03-2bce-417a-b01c-aaec09a5c8c3\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.287523 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pdxr\" (UniqueName: \"kubernetes.io/projected/f50b4c03-2bce-417a-b01c-aaec09a5c8c3-kube-api-access-4pdxr\") pod \"glance-default-internal-api-0\" (UID: \"f50b4c03-2bce-417a-b01c-aaec09a5c8c3\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.287571 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f50b4c03-2bce-417a-b01c-aaec09a5c8c3-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f50b4c03-2bce-417a-b01c-aaec09a5c8c3\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.287596 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f50b4c03-2bce-417a-b01c-aaec09a5c8c3-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f50b4c03-2bce-417a-b01c-aaec09a5c8c3\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.287951 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f50b4c03-2bce-417a-b01c-aaec09a5c8c3-logs\") pod \"glance-default-internal-api-0\" (UID: \"f50b4c03-2bce-417a-b01c-aaec09a5c8c3\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.304488 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9" path="/var/lib/kubelet/pods/2f3c5d32-5a48-43df-a7e8-c65ae39bf8b9/volumes" Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.313673 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35011d99-3341-4081-9733-a6f42f69a6d2" path="/var/lib/kubelet/pods/35011d99-3341-4081-9733-a6f42f69a6d2/volumes" Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.390382 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f50b4c03-2bce-417a-b01c-aaec09a5c8c3-logs\") pod \"glance-default-internal-api-0\" (UID: \"f50b4c03-2bce-417a-b01c-aaec09a5c8c3\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.390613 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f50b4c03-2bce-417a-b01c-aaec09a5c8c3-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f50b4c03-2bce-417a-b01c-aaec09a5c8c3\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.391576 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"f50b4c03-2bce-417a-b01c-aaec09a5c8c3\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.390817 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f50b4c03-2bce-417a-b01c-aaec09a5c8c3-logs\") pod \"glance-default-internal-api-0\" (UID: \"f50b4c03-2bce-417a-b01c-aaec09a5c8c3\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.391869 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f50b4c03-2bce-417a-b01c-aaec09a5c8c3-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f50b4c03-2bce-417a-b01c-aaec09a5c8c3\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.391900 4556 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"f50b4c03-2bce-417a-b01c-aaec09a5c8c3\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.391930 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f50b4c03-2bce-417a-b01c-aaec09a5c8c3-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f50b4c03-2bce-417a-b01c-aaec09a5c8c3\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.391981 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pdxr\" (UniqueName: \"kubernetes.io/projected/f50b4c03-2bce-417a-b01c-aaec09a5c8c3-kube-api-access-4pdxr\") pod \"glance-default-internal-api-0\" (UID: \"f50b4c03-2bce-417a-b01c-aaec09a5c8c3\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.392052 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f50b4c03-2bce-417a-b01c-aaec09a5c8c3-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f50b4c03-2bce-417a-b01c-aaec09a5c8c3\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.392088 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f50b4c03-2bce-417a-b01c-aaec09a5c8c3-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f50b4c03-2bce-417a-b01c-aaec09a5c8c3\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.396123 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f50b4c03-2bce-417a-b01c-aaec09a5c8c3-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f50b4c03-2bce-417a-b01c-aaec09a5c8c3\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.408978 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f50b4c03-2bce-417a-b01c-aaec09a5c8c3-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f50b4c03-2bce-417a-b01c-aaec09a5c8c3\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.409194 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f50b4c03-2bce-417a-b01c-aaec09a5c8c3-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f50b4c03-2bce-417a-b01c-aaec09a5c8c3\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.409667 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f50b4c03-2bce-417a-b01c-aaec09a5c8c3-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f50b4c03-2bce-417a-b01c-aaec09a5c8c3\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.416357 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f50b4c03-2bce-417a-b01c-aaec09a5c8c3-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f50b4c03-2bce-417a-b01c-aaec09a5c8c3\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.416641 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pdxr\" (UniqueName: \"kubernetes.io/projected/f50b4c03-2bce-417a-b01c-aaec09a5c8c3-kube-api-access-4pdxr\") pod \"glance-default-internal-api-0\" (UID: \"f50b4c03-2bce-417a-b01c-aaec09a5c8c3\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.426352 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"f50b4c03-2bce-417a-b01c-aaec09a5c8c3\") " pod="openstack/glance-default-internal-api-0" Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.436333 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.482692 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.596928 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/40b701d5-4495-47c3-bb6e-6ba466958355-log-httpd\") pod \"40b701d5-4495-47c3-bb6e-6ba466958355\" (UID: \"40b701d5-4495-47c3-bb6e-6ba466958355\") " Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.597771 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/40b701d5-4495-47c3-bb6e-6ba466958355-scripts\") pod \"40b701d5-4495-47c3-bb6e-6ba466958355\" (UID: \"40b701d5-4495-47c3-bb6e-6ba466958355\") " Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.597894 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p8blb\" (UniqueName: \"kubernetes.io/projected/40b701d5-4495-47c3-bb6e-6ba466958355-kube-api-access-p8blb\") pod \"40b701d5-4495-47c3-bb6e-6ba466958355\" (UID: \"40b701d5-4495-47c3-bb6e-6ba466958355\") " Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.598452 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40b701d5-4495-47c3-bb6e-6ba466958355-config-data\") pod \"40b701d5-4495-47c3-bb6e-6ba466958355\" (UID: \"40b701d5-4495-47c3-bb6e-6ba466958355\") " Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.598681 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40b701d5-4495-47c3-bb6e-6ba466958355-combined-ca-bundle\") pod \"40b701d5-4495-47c3-bb6e-6ba466958355\" (UID: \"40b701d5-4495-47c3-bb6e-6ba466958355\") " Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.598810 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/40b701d5-4495-47c3-bb6e-6ba466958355-sg-core-conf-yaml\") pod \"40b701d5-4495-47c3-bb6e-6ba466958355\" (UID: \"40b701d5-4495-47c3-bb6e-6ba466958355\") " Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.598965 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/40b701d5-4495-47c3-bb6e-6ba466958355-run-httpd\") pod \"40b701d5-4495-47c3-bb6e-6ba466958355\" (UID: \"40b701d5-4495-47c3-bb6e-6ba466958355\") " Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.597718 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40b701d5-4495-47c3-bb6e-6ba466958355-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "40b701d5-4495-47c3-bb6e-6ba466958355" (UID: "40b701d5-4495-47c3-bb6e-6ba466958355"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.600296 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40b701d5-4495-47c3-bb6e-6ba466958355-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "40b701d5-4495-47c3-bb6e-6ba466958355" (UID: "40b701d5-4495-47c3-bb6e-6ba466958355"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.607027 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40b701d5-4495-47c3-bb6e-6ba466958355-kube-api-access-p8blb" (OuterVolumeSpecName: "kube-api-access-p8blb") pod "40b701d5-4495-47c3-bb6e-6ba466958355" (UID: "40b701d5-4495-47c3-bb6e-6ba466958355"). InnerVolumeSpecName "kube-api-access-p8blb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.611529 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40b701d5-4495-47c3-bb6e-6ba466958355-scripts" (OuterVolumeSpecName: "scripts") pod "40b701d5-4495-47c3-bb6e-6ba466958355" (UID: "40b701d5-4495-47c3-bb6e-6ba466958355"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.685995 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40b701d5-4495-47c3-bb6e-6ba466958355-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "40b701d5-4495-47c3-bb6e-6ba466958355" (UID: "40b701d5-4495-47c3-bb6e-6ba466958355"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.703470 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p8blb\" (UniqueName: \"kubernetes.io/projected/40b701d5-4495-47c3-bb6e-6ba466958355-kube-api-access-p8blb\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.703607 4556 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/40b701d5-4495-47c3-bb6e-6ba466958355-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.703672 4556 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/40b701d5-4495-47c3-bb6e-6ba466958355-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.703726 4556 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/40b701d5-4495-47c3-bb6e-6ba466958355-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.703787 4556 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/40b701d5-4495-47c3-bb6e-6ba466958355-scripts\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.721076 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40b701d5-4495-47c3-bb6e-6ba466958355-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "40b701d5-4495-47c3-bb6e-6ba466958355" (UID: "40b701d5-4495-47c3-bb6e-6ba466958355"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.736208 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40b701d5-4495-47c3-bb6e-6ba466958355-config-data" (OuterVolumeSpecName: "config-data") pod "40b701d5-4495-47c3-bb6e-6ba466958355" (UID: "40b701d5-4495-47c3-bb6e-6ba466958355"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.806550 4556 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40b701d5-4495-47c3-bb6e-6ba466958355-config-data\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:19 crc kubenswrapper[4556]: I0218 09:19:19.806584 4556 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40b701d5-4495-47c3-bb6e-6ba466958355-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:20 crc kubenswrapper[4556]: I0218 09:19:20.046788 4556 generic.go:334] "Generic (PLEG): container finished" podID="40b701d5-4495-47c3-bb6e-6ba466958355" containerID="b0838fd427fe9647da3a2c765ef31d6f5e3b1d99df1a0a099104c13db47316a4" exitCode=0 Feb 18 09:19:20 crc kubenswrapper[4556]: I0218 09:19:20.046867 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"40b701d5-4495-47c3-bb6e-6ba466958355","Type":"ContainerDied","Data":"b0838fd427fe9647da3a2c765ef31d6f5e3b1d99df1a0a099104c13db47316a4"} Feb 18 09:19:20 crc kubenswrapper[4556]: I0218 09:19:20.046900 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"40b701d5-4495-47c3-bb6e-6ba466958355","Type":"ContainerDied","Data":"e11c71227e2adf3a67dc47bc79ad4548bdf54b4ca48ccc8a77137ac6f5f17844"} Feb 18 09:19:20 crc kubenswrapper[4556]: I0218 09:19:20.046930 4556 scope.go:117] "RemoveContainer" containerID="1dda491fdb344030848ca5df7d5c32e96139ad7022a869bb7c2f0fc02c1bb6ed" Feb 18 09:19:20 crc kubenswrapper[4556]: I0218 09:19:20.047026 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 18 09:19:20 crc kubenswrapper[4556]: I0218 09:19:20.082700 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"23a483a0-fda5-4bf6-bfea-06a84d8d6533","Type":"ContainerStarted","Data":"bd77bb8078024fa73ca7ec5c1d4b7b559fd9726dd746ff3bc3813ec1e2e37fe4"} Feb 18 09:19:20 crc kubenswrapper[4556]: I0218 09:19:20.099561 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 18 09:19:20 crc kubenswrapper[4556]: I0218 09:19:20.107788 4556 scope.go:117] "RemoveContainer" containerID="f984aeec06953af9d589b06b93e201997457368d82ecc83ab082489d056e155f" Feb 18 09:19:20 crc kubenswrapper[4556]: I0218 09:19:20.108195 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 18 09:19:20 crc kubenswrapper[4556]: I0218 09:19:20.117040 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 18 09:19:20 crc kubenswrapper[4556]: I0218 09:19:20.144518 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 18 09:19:20 crc kubenswrapper[4556]: E0218 09:19:20.144891 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40b701d5-4495-47c3-bb6e-6ba466958355" containerName="sg-core" Feb 18 09:19:20 crc kubenswrapper[4556]: I0218 09:19:20.144916 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="40b701d5-4495-47c3-bb6e-6ba466958355" containerName="sg-core" Feb 18 09:19:20 crc kubenswrapper[4556]: E0218 09:19:20.144934 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40b701d5-4495-47c3-bb6e-6ba466958355" containerName="ceilometer-notification-agent" Feb 18 09:19:20 crc kubenswrapper[4556]: I0218 09:19:20.144940 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="40b701d5-4495-47c3-bb6e-6ba466958355" containerName="ceilometer-notification-agent" Feb 18 09:19:20 crc kubenswrapper[4556]: E0218 09:19:20.144951 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40b701d5-4495-47c3-bb6e-6ba466958355" containerName="ceilometer-central-agent" Feb 18 09:19:20 crc kubenswrapper[4556]: I0218 09:19:20.144958 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="40b701d5-4495-47c3-bb6e-6ba466958355" containerName="ceilometer-central-agent" Feb 18 09:19:20 crc kubenswrapper[4556]: E0218 09:19:20.144997 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40b701d5-4495-47c3-bb6e-6ba466958355" containerName="proxy-httpd" Feb 18 09:19:20 crc kubenswrapper[4556]: I0218 09:19:20.145096 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="40b701d5-4495-47c3-bb6e-6ba466958355" containerName="proxy-httpd" Feb 18 09:19:20 crc kubenswrapper[4556]: I0218 09:19:20.145408 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="40b701d5-4495-47c3-bb6e-6ba466958355" containerName="ceilometer-notification-agent" Feb 18 09:19:20 crc kubenswrapper[4556]: I0218 09:19:20.145429 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="40b701d5-4495-47c3-bb6e-6ba466958355" containerName="sg-core" Feb 18 09:19:20 crc kubenswrapper[4556]: I0218 09:19:20.145444 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="40b701d5-4495-47c3-bb6e-6ba466958355" containerName="proxy-httpd" Feb 18 09:19:20 crc kubenswrapper[4556]: I0218 09:19:20.145458 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="40b701d5-4495-47c3-bb6e-6ba466958355" containerName="ceilometer-central-agent" Feb 18 09:19:20 crc kubenswrapper[4556]: I0218 09:19:20.151724 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 18 09:19:20 crc kubenswrapper[4556]: I0218 09:19:20.156136 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 18 09:19:20 crc kubenswrapper[4556]: I0218 09:19:20.158000 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 18 09:19:20 crc kubenswrapper[4556]: I0218 09:19:20.158460 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 18 09:19:20 crc kubenswrapper[4556]: I0218 09:19:20.175588 4556 scope.go:117] "RemoveContainer" containerID="471e8d3bb0e3ae1e40387a1ab945b5b86c109e0aae7d0c0de060669dafb0b2e1" Feb 18 09:19:20 crc kubenswrapper[4556]: W0218 09:19:20.198489 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf50b4c03_2bce_417a_b01c_aaec09a5c8c3.slice/crio-7c7bad1296142a168aeb4d9f757f781c8ae814f45823399acc83c1be3e3bb2f0 WatchSource:0}: Error finding container 7c7bad1296142a168aeb4d9f757f781c8ae814f45823399acc83c1be3e3bb2f0: Status 404 returned error can't find the container with id 7c7bad1296142a168aeb4d9f757f781c8ae814f45823399acc83c1be3e3bb2f0 Feb 18 09:19:20 crc kubenswrapper[4556]: I0218 09:19:20.212329 4556 scope.go:117] "RemoveContainer" containerID="b0838fd427fe9647da3a2c765ef31d6f5e3b1d99df1a0a099104c13db47316a4" Feb 18 09:19:20 crc kubenswrapper[4556]: I0218 09:19:20.215690 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/33ce5ec8-44c2-4b8c-865f-36ee8722d716-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"33ce5ec8-44c2-4b8c-865f-36ee8722d716\") " pod="openstack/ceilometer-0" Feb 18 09:19:20 crc kubenswrapper[4556]: I0218 09:19:20.215764 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gxl4\" (UniqueName: \"kubernetes.io/projected/33ce5ec8-44c2-4b8c-865f-36ee8722d716-kube-api-access-5gxl4\") pod \"ceilometer-0\" (UID: \"33ce5ec8-44c2-4b8c-865f-36ee8722d716\") " pod="openstack/ceilometer-0" Feb 18 09:19:20 crc kubenswrapper[4556]: I0218 09:19:20.215880 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33ce5ec8-44c2-4b8c-865f-36ee8722d716-log-httpd\") pod \"ceilometer-0\" (UID: \"33ce5ec8-44c2-4b8c-865f-36ee8722d716\") " pod="openstack/ceilometer-0" Feb 18 09:19:20 crc kubenswrapper[4556]: I0218 09:19:20.216004 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33ce5ec8-44c2-4b8c-865f-36ee8722d716-scripts\") pod \"ceilometer-0\" (UID: \"33ce5ec8-44c2-4b8c-865f-36ee8722d716\") " pod="openstack/ceilometer-0" Feb 18 09:19:20 crc kubenswrapper[4556]: I0218 09:19:20.216027 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33ce5ec8-44c2-4b8c-865f-36ee8722d716-config-data\") pod \"ceilometer-0\" (UID: \"33ce5ec8-44c2-4b8c-865f-36ee8722d716\") " pod="openstack/ceilometer-0" Feb 18 09:19:20 crc kubenswrapper[4556]: I0218 09:19:20.216042 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33ce5ec8-44c2-4b8c-865f-36ee8722d716-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"33ce5ec8-44c2-4b8c-865f-36ee8722d716\") " pod="openstack/ceilometer-0" Feb 18 09:19:20 crc kubenswrapper[4556]: I0218 09:19:20.216077 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33ce5ec8-44c2-4b8c-865f-36ee8722d716-run-httpd\") pod \"ceilometer-0\" (UID: \"33ce5ec8-44c2-4b8c-865f-36ee8722d716\") " pod="openstack/ceilometer-0" Feb 18 09:19:20 crc kubenswrapper[4556]: I0218 09:19:20.260471 4556 scope.go:117] "RemoveContainer" containerID="1dda491fdb344030848ca5df7d5c32e96139ad7022a869bb7c2f0fc02c1bb6ed" Feb 18 09:19:20 crc kubenswrapper[4556]: E0218 09:19:20.261141 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1dda491fdb344030848ca5df7d5c32e96139ad7022a869bb7c2f0fc02c1bb6ed\": container with ID starting with 1dda491fdb344030848ca5df7d5c32e96139ad7022a869bb7c2f0fc02c1bb6ed not found: ID does not exist" containerID="1dda491fdb344030848ca5df7d5c32e96139ad7022a869bb7c2f0fc02c1bb6ed" Feb 18 09:19:20 crc kubenswrapper[4556]: I0218 09:19:20.261192 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1dda491fdb344030848ca5df7d5c32e96139ad7022a869bb7c2f0fc02c1bb6ed"} err="failed to get container status \"1dda491fdb344030848ca5df7d5c32e96139ad7022a869bb7c2f0fc02c1bb6ed\": rpc error: code = NotFound desc = could not find container \"1dda491fdb344030848ca5df7d5c32e96139ad7022a869bb7c2f0fc02c1bb6ed\": container with ID starting with 1dda491fdb344030848ca5df7d5c32e96139ad7022a869bb7c2f0fc02c1bb6ed not found: ID does not exist" Feb 18 09:19:20 crc kubenswrapper[4556]: I0218 09:19:20.261216 4556 scope.go:117] "RemoveContainer" containerID="f984aeec06953af9d589b06b93e201997457368d82ecc83ab082489d056e155f" Feb 18 09:19:20 crc kubenswrapper[4556]: E0218 09:19:20.261520 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f984aeec06953af9d589b06b93e201997457368d82ecc83ab082489d056e155f\": container with ID starting with f984aeec06953af9d589b06b93e201997457368d82ecc83ab082489d056e155f not found: ID does not exist" containerID="f984aeec06953af9d589b06b93e201997457368d82ecc83ab082489d056e155f" Feb 18 09:19:20 crc kubenswrapper[4556]: I0218 09:19:20.261538 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f984aeec06953af9d589b06b93e201997457368d82ecc83ab082489d056e155f"} err="failed to get container status \"f984aeec06953af9d589b06b93e201997457368d82ecc83ab082489d056e155f\": rpc error: code = NotFound desc = could not find container \"f984aeec06953af9d589b06b93e201997457368d82ecc83ab082489d056e155f\": container with ID starting with f984aeec06953af9d589b06b93e201997457368d82ecc83ab082489d056e155f not found: ID does not exist" Feb 18 09:19:20 crc kubenswrapper[4556]: I0218 09:19:20.261572 4556 scope.go:117] "RemoveContainer" containerID="471e8d3bb0e3ae1e40387a1ab945b5b86c109e0aae7d0c0de060669dafb0b2e1" Feb 18 09:19:20 crc kubenswrapper[4556]: E0218 09:19:20.261821 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"471e8d3bb0e3ae1e40387a1ab945b5b86c109e0aae7d0c0de060669dafb0b2e1\": container with ID starting with 471e8d3bb0e3ae1e40387a1ab945b5b86c109e0aae7d0c0de060669dafb0b2e1 not found: ID does not exist" containerID="471e8d3bb0e3ae1e40387a1ab945b5b86c109e0aae7d0c0de060669dafb0b2e1" Feb 18 09:19:20 crc kubenswrapper[4556]: I0218 09:19:20.261838 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"471e8d3bb0e3ae1e40387a1ab945b5b86c109e0aae7d0c0de060669dafb0b2e1"} err="failed to get container status \"471e8d3bb0e3ae1e40387a1ab945b5b86c109e0aae7d0c0de060669dafb0b2e1\": rpc error: code = NotFound desc = could not find container \"471e8d3bb0e3ae1e40387a1ab945b5b86c109e0aae7d0c0de060669dafb0b2e1\": container with ID starting with 471e8d3bb0e3ae1e40387a1ab945b5b86c109e0aae7d0c0de060669dafb0b2e1 not found: ID does not exist" Feb 18 09:19:20 crc kubenswrapper[4556]: I0218 09:19:20.261870 4556 scope.go:117] "RemoveContainer" containerID="b0838fd427fe9647da3a2c765ef31d6f5e3b1d99df1a0a099104c13db47316a4" Feb 18 09:19:20 crc kubenswrapper[4556]: E0218 09:19:20.262121 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0838fd427fe9647da3a2c765ef31d6f5e3b1d99df1a0a099104c13db47316a4\": container with ID starting with b0838fd427fe9647da3a2c765ef31d6f5e3b1d99df1a0a099104c13db47316a4 not found: ID does not exist" containerID="b0838fd427fe9647da3a2c765ef31d6f5e3b1d99df1a0a099104c13db47316a4" Feb 18 09:19:20 crc kubenswrapper[4556]: I0218 09:19:20.262136 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0838fd427fe9647da3a2c765ef31d6f5e3b1d99df1a0a099104c13db47316a4"} err="failed to get container status \"b0838fd427fe9647da3a2c765ef31d6f5e3b1d99df1a0a099104c13db47316a4\": rpc error: code = NotFound desc = could not find container \"b0838fd427fe9647da3a2c765ef31d6f5e3b1d99df1a0a099104c13db47316a4\": container with ID starting with b0838fd427fe9647da3a2c765ef31d6f5e3b1d99df1a0a099104c13db47316a4 not found: ID does not exist" Feb 18 09:19:20 crc kubenswrapper[4556]: I0218 09:19:20.289103 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Feb 18 09:19:20 crc kubenswrapper[4556]: I0218 09:19:20.318052 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/33ce5ec8-44c2-4b8c-865f-36ee8722d716-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"33ce5ec8-44c2-4b8c-865f-36ee8722d716\") " pod="openstack/ceilometer-0" Feb 18 09:19:20 crc kubenswrapper[4556]: I0218 09:19:20.318136 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gxl4\" (UniqueName: \"kubernetes.io/projected/33ce5ec8-44c2-4b8c-865f-36ee8722d716-kube-api-access-5gxl4\") pod \"ceilometer-0\" (UID: \"33ce5ec8-44c2-4b8c-865f-36ee8722d716\") " pod="openstack/ceilometer-0" Feb 18 09:19:20 crc kubenswrapper[4556]: I0218 09:19:20.318273 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33ce5ec8-44c2-4b8c-865f-36ee8722d716-log-httpd\") pod \"ceilometer-0\" (UID: \"33ce5ec8-44c2-4b8c-865f-36ee8722d716\") " pod="openstack/ceilometer-0" Feb 18 09:19:20 crc kubenswrapper[4556]: I0218 09:19:20.318443 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33ce5ec8-44c2-4b8c-865f-36ee8722d716-scripts\") pod \"ceilometer-0\" (UID: \"33ce5ec8-44c2-4b8c-865f-36ee8722d716\") " pod="openstack/ceilometer-0" Feb 18 09:19:20 crc kubenswrapper[4556]: I0218 09:19:20.318464 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33ce5ec8-44c2-4b8c-865f-36ee8722d716-config-data\") pod \"ceilometer-0\" (UID: \"33ce5ec8-44c2-4b8c-865f-36ee8722d716\") " pod="openstack/ceilometer-0" Feb 18 09:19:20 crc kubenswrapper[4556]: I0218 09:19:20.318480 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33ce5ec8-44c2-4b8c-865f-36ee8722d716-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"33ce5ec8-44c2-4b8c-865f-36ee8722d716\") " pod="openstack/ceilometer-0" Feb 18 09:19:20 crc kubenswrapper[4556]: I0218 09:19:20.318512 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33ce5ec8-44c2-4b8c-865f-36ee8722d716-run-httpd\") pod \"ceilometer-0\" (UID: \"33ce5ec8-44c2-4b8c-865f-36ee8722d716\") " pod="openstack/ceilometer-0" Feb 18 09:19:20 crc kubenswrapper[4556]: I0218 09:19:20.322597 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33ce5ec8-44c2-4b8c-865f-36ee8722d716-log-httpd\") pod \"ceilometer-0\" (UID: \"33ce5ec8-44c2-4b8c-865f-36ee8722d716\") " pod="openstack/ceilometer-0" Feb 18 09:19:20 crc kubenswrapper[4556]: I0218 09:19:20.323478 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/33ce5ec8-44c2-4b8c-865f-36ee8722d716-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"33ce5ec8-44c2-4b8c-865f-36ee8722d716\") " pod="openstack/ceilometer-0" Feb 18 09:19:20 crc kubenswrapper[4556]: I0218 09:19:20.323846 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33ce5ec8-44c2-4b8c-865f-36ee8722d716-run-httpd\") pod \"ceilometer-0\" (UID: \"33ce5ec8-44c2-4b8c-865f-36ee8722d716\") " pod="openstack/ceilometer-0" Feb 18 09:19:20 crc kubenswrapper[4556]: I0218 09:19:20.337598 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33ce5ec8-44c2-4b8c-865f-36ee8722d716-scripts\") pod \"ceilometer-0\" (UID: \"33ce5ec8-44c2-4b8c-865f-36ee8722d716\") " pod="openstack/ceilometer-0" Feb 18 09:19:20 crc kubenswrapper[4556]: I0218 09:19:20.339189 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gxl4\" (UniqueName: \"kubernetes.io/projected/33ce5ec8-44c2-4b8c-865f-36ee8722d716-kube-api-access-5gxl4\") pod \"ceilometer-0\" (UID: \"33ce5ec8-44c2-4b8c-865f-36ee8722d716\") " pod="openstack/ceilometer-0" Feb 18 09:19:20 crc kubenswrapper[4556]: I0218 09:19:20.340374 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33ce5ec8-44c2-4b8c-865f-36ee8722d716-config-data\") pod \"ceilometer-0\" (UID: \"33ce5ec8-44c2-4b8c-865f-36ee8722d716\") " pod="openstack/ceilometer-0" Feb 18 09:19:20 crc kubenswrapper[4556]: I0218 09:19:20.344753 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33ce5ec8-44c2-4b8c-865f-36ee8722d716-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"33ce5ec8-44c2-4b8c-865f-36ee8722d716\") " pod="openstack/ceilometer-0" Feb 18 09:19:20 crc kubenswrapper[4556]: I0218 09:19:20.511493 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 18 09:19:21 crc kubenswrapper[4556]: I0218 09:19:21.065599 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 18 09:19:21 crc kubenswrapper[4556]: I0218 09:19:21.092184 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33ce5ec8-44c2-4b8c-865f-36ee8722d716","Type":"ContainerStarted","Data":"404d04226b2299a9a4f3e0aa73e0e9bc8b4f17c70cae76cb1ca181c3af19d562"} Feb 18 09:19:21 crc kubenswrapper[4556]: I0218 09:19:21.095090 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"23a483a0-fda5-4bf6-bfea-06a84d8d6533","Type":"ContainerStarted","Data":"32bda9aab96112e0beff6cb33f766822b6ddc368472f774f8d0e198cd8a9a1c2"} Feb 18 09:19:21 crc kubenswrapper[4556]: I0218 09:19:21.099208 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f50b4c03-2bce-417a-b01c-aaec09a5c8c3","Type":"ContainerStarted","Data":"72a10ad8e2cb9d19616bb8f5e593c2374a6ba39cbd73900efaccfd0f1d148cb7"} Feb 18 09:19:21 crc kubenswrapper[4556]: I0218 09:19:21.099243 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f50b4c03-2bce-417a-b01c-aaec09a5c8c3","Type":"ContainerStarted","Data":"7c7bad1296142a168aeb4d9f757f781c8ae814f45823399acc83c1be3e3bb2f0"} Feb 18 09:19:21 crc kubenswrapper[4556]: I0218 09:19:21.115177 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.115167549 podStartE2EDuration="4.115167549s" podCreationTimestamp="2026-02-18 09:19:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:19:21.10956054 +0000 UTC m=+918.126521521" watchObservedRunningTime="2026-02-18 09:19:21.115167549 +0000 UTC m=+918.132128519" Feb 18 09:19:21 crc kubenswrapper[4556]: I0218 09:19:21.298557 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40b701d5-4495-47c3-bb6e-6ba466958355" path="/var/lib/kubelet/pods/40b701d5-4495-47c3-bb6e-6ba466958355/volumes" Feb 18 09:19:21 crc kubenswrapper[4556]: I0218 09:19:21.434341 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rmt7n" Feb 18 09:19:21 crc kubenswrapper[4556]: I0218 09:19:21.475629 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rmt7n" Feb 18 09:19:22 crc kubenswrapper[4556]: I0218 09:19:22.114384 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f50b4c03-2bce-417a-b01c-aaec09a5c8c3","Type":"ContainerStarted","Data":"ff964dda1da4a658ef33e7caa5d3ba74f2e5fc06f8d120b4b212f1de57171d3f"} Feb 18 09:19:22 crc kubenswrapper[4556]: I0218 09:19:22.118705 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33ce5ec8-44c2-4b8c-865f-36ee8722d716","Type":"ContainerStarted","Data":"4d4dafcb48311a7d568ad32ddb27501224c2e007d1e076c372c237e47a26a39c"} Feb 18 09:19:22 crc kubenswrapper[4556]: I0218 09:19:22.135104 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.135091463 podStartE2EDuration="3.135091463s" podCreationTimestamp="2026-02-18 09:19:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:19:22.13289068 +0000 UTC m=+919.149851661" watchObservedRunningTime="2026-02-18 09:19:22.135091463 +0000 UTC m=+919.152052443" Feb 18 09:19:22 crc kubenswrapper[4556]: I0218 09:19:22.275536 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rmt7n"] Feb 18 09:19:23 crc kubenswrapper[4556]: I0218 09:19:23.128450 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33ce5ec8-44c2-4b8c-865f-36ee8722d716","Type":"ContainerStarted","Data":"d5a63205c3d5881f590fd4f8fd3992733d92abffbb8653897c64b67729fa0b65"} Feb 18 09:19:23 crc kubenswrapper[4556]: I0218 09:19:23.129092 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rmt7n" podUID="0a557635-d67d-4e93-a82c-990c34fe7d48" containerName="registry-server" containerID="cri-o://e9d99eae194d31abfd1573901538558b29b3ebf6244cb332929af9b0130f99f8" gracePeriod=2 Feb 18 09:19:24 crc kubenswrapper[4556]: I0218 09:19:24.140872 4556 generic.go:334] "Generic (PLEG): container finished" podID="0a557635-d67d-4e93-a82c-990c34fe7d48" containerID="e9d99eae194d31abfd1573901538558b29b3ebf6244cb332929af9b0130f99f8" exitCode=0 Feb 18 09:19:24 crc kubenswrapper[4556]: I0218 09:19:24.141076 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rmt7n" event={"ID":"0a557635-d67d-4e93-a82c-990c34fe7d48","Type":"ContainerDied","Data":"e9d99eae194d31abfd1573901538558b29b3ebf6244cb332929af9b0130f99f8"} Feb 18 09:19:27 crc kubenswrapper[4556]: I0218 09:19:27.252267 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rmt7n" Feb 18 09:19:27 crc kubenswrapper[4556]: I0218 09:19:27.408434 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a557635-d67d-4e93-a82c-990c34fe7d48-utilities\") pod \"0a557635-d67d-4e93-a82c-990c34fe7d48\" (UID: \"0a557635-d67d-4e93-a82c-990c34fe7d48\") " Feb 18 09:19:27 crc kubenswrapper[4556]: I0218 09:19:27.408685 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rqss7\" (UniqueName: \"kubernetes.io/projected/0a557635-d67d-4e93-a82c-990c34fe7d48-kube-api-access-rqss7\") pod \"0a557635-d67d-4e93-a82c-990c34fe7d48\" (UID: \"0a557635-d67d-4e93-a82c-990c34fe7d48\") " Feb 18 09:19:27 crc kubenswrapper[4556]: I0218 09:19:27.408848 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a557635-d67d-4e93-a82c-990c34fe7d48-catalog-content\") pod \"0a557635-d67d-4e93-a82c-990c34fe7d48\" (UID: \"0a557635-d67d-4e93-a82c-990c34fe7d48\") " Feb 18 09:19:27 crc kubenswrapper[4556]: I0218 09:19:27.410721 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a557635-d67d-4e93-a82c-990c34fe7d48-utilities" (OuterVolumeSpecName: "utilities") pod "0a557635-d67d-4e93-a82c-990c34fe7d48" (UID: "0a557635-d67d-4e93-a82c-990c34fe7d48"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:19:27 crc kubenswrapper[4556]: I0218 09:19:27.415592 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a557635-d67d-4e93-a82c-990c34fe7d48-kube-api-access-rqss7" (OuterVolumeSpecName: "kube-api-access-rqss7") pod "0a557635-d67d-4e93-a82c-990c34fe7d48" (UID: "0a557635-d67d-4e93-a82c-990c34fe7d48"). InnerVolumeSpecName "kube-api-access-rqss7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:19:27 crc kubenswrapper[4556]: I0218 09:19:27.511313 4556 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a557635-d67d-4e93-a82c-990c34fe7d48-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:27 crc kubenswrapper[4556]: I0218 09:19:27.511340 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rqss7\" (UniqueName: \"kubernetes.io/projected/0a557635-d67d-4e93-a82c-990c34fe7d48-kube-api-access-rqss7\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:27 crc kubenswrapper[4556]: I0218 09:19:27.514941 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a557635-d67d-4e93-a82c-990c34fe7d48-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0a557635-d67d-4e93-a82c-990c34fe7d48" (UID: "0a557635-d67d-4e93-a82c-990c34fe7d48"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:19:27 crc kubenswrapper[4556]: I0218 09:19:27.613554 4556 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a557635-d67d-4e93-a82c-990c34fe7d48-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:28 crc kubenswrapper[4556]: I0218 09:19:28.182948 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rmt7n" event={"ID":"0a557635-d67d-4e93-a82c-990c34fe7d48","Type":"ContainerDied","Data":"7a953807970cdb89372bde18f3bdab42a137614c5d4ce0ea1c1d2f7f0b718a42"} Feb 18 09:19:28 crc kubenswrapper[4556]: I0218 09:19:28.182997 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rmt7n" Feb 18 09:19:28 crc kubenswrapper[4556]: I0218 09:19:28.183011 4556 scope.go:117] "RemoveContainer" containerID="e9d99eae194d31abfd1573901538558b29b3ebf6244cb332929af9b0130f99f8" Feb 18 09:19:28 crc kubenswrapper[4556]: I0218 09:19:28.184766 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33ce5ec8-44c2-4b8c-865f-36ee8722d716","Type":"ContainerStarted","Data":"4ba210fb11d2410c63309605c467e2da8848df4407665ca313864332e3878227"} Feb 18 09:19:28 crc kubenswrapper[4556]: I0218 09:19:28.185935 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-wxltl" event={"ID":"f4a792c8-97e0-46f2-8a40-747a7cae58f5","Type":"ContainerStarted","Data":"69e41aca003b0c33c159e3090331559d090f11ef66400ffc6495e1d73d0945a8"} Feb 18 09:19:28 crc kubenswrapper[4556]: I0218 09:19:28.199040 4556 scope.go:117] "RemoveContainer" containerID="aa6c9b508b068a81e21c5d33efe4beaa37c6cc99d038679316b57fc58dddfbe3" Feb 18 09:19:28 crc kubenswrapper[4556]: I0218 09:19:28.214072 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-wxltl" podStartSLOduration=2.270714168 podStartE2EDuration="11.214058718s" podCreationTimestamp="2026-02-18 09:19:17 +0000 UTC" firstStartedPulling="2026-02-18 09:19:18.036718787 +0000 UTC m=+915.053679767" lastFinishedPulling="2026-02-18 09:19:26.980063337 +0000 UTC m=+923.997024317" observedRunningTime="2026-02-18 09:19:28.208894724 +0000 UTC m=+925.225855703" watchObservedRunningTime="2026-02-18 09:19:28.214058718 +0000 UTC m=+925.231019698" Feb 18 09:19:28 crc kubenswrapper[4556]: I0218 09:19:28.246051 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rmt7n"] Feb 18 09:19:28 crc kubenswrapper[4556]: I0218 09:19:28.247064 4556 scope.go:117] "RemoveContainer" containerID="920ec97ab27b16c8945b61d64a2df962837b1fd7a6881c99c07a5be5234c0d07" Feb 18 09:19:28 crc kubenswrapper[4556]: I0218 09:19:28.254324 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rmt7n"] Feb 18 09:19:28 crc kubenswrapper[4556]: I0218 09:19:28.386203 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 18 09:19:28 crc kubenswrapper[4556]: I0218 09:19:28.386250 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 18 09:19:28 crc kubenswrapper[4556]: I0218 09:19:28.413716 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 18 09:19:28 crc kubenswrapper[4556]: I0218 09:19:28.418358 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 18 09:19:29 crc kubenswrapper[4556]: I0218 09:19:29.198416 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 18 09:19:29 crc kubenswrapper[4556]: I0218 09:19:29.198569 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 18 09:19:29 crc kubenswrapper[4556]: I0218 09:19:29.293387 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a557635-d67d-4e93-a82c-990c34fe7d48" path="/var/lib/kubelet/pods/0a557635-d67d-4e93-a82c-990c34fe7d48/volumes" Feb 18 09:19:29 crc kubenswrapper[4556]: I0218 09:19:29.483925 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 18 09:19:29 crc kubenswrapper[4556]: I0218 09:19:29.484061 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 18 09:19:29 crc kubenswrapper[4556]: I0218 09:19:29.515054 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 18 09:19:29 crc kubenswrapper[4556]: I0218 09:19:29.520656 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 18 09:19:30 crc kubenswrapper[4556]: I0218 09:19:30.219935 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33ce5ec8-44c2-4b8c-865f-36ee8722d716","Type":"ContainerStarted","Data":"965d06552f33c2639ae14fe6a55d0ebb59be0c8cce9bdb7d929e282c207a9b9c"} Feb 18 09:19:30 crc kubenswrapper[4556]: I0218 09:19:30.220639 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 18 09:19:30 crc kubenswrapper[4556]: I0218 09:19:30.220659 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 18 09:19:30 crc kubenswrapper[4556]: I0218 09:19:30.246981 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.152883694 podStartE2EDuration="10.24696645s" podCreationTimestamp="2026-02-18 09:19:20 +0000 UTC" firstStartedPulling="2026-02-18 09:19:21.081167147 +0000 UTC m=+918.098128127" lastFinishedPulling="2026-02-18 09:19:29.175249893 +0000 UTC m=+926.192210883" observedRunningTime="2026-02-18 09:19:30.241402422 +0000 UTC m=+927.258363402" watchObservedRunningTime="2026-02-18 09:19:30.24696645 +0000 UTC m=+927.263927430" Feb 18 09:19:30 crc kubenswrapper[4556]: I0218 09:19:30.866941 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 18 09:19:30 crc kubenswrapper[4556]: I0218 09:19:30.868827 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 18 09:19:31 crc kubenswrapper[4556]: I0218 09:19:31.245724 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 18 09:19:31 crc kubenswrapper[4556]: I0218 09:19:31.713468 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 18 09:19:31 crc kubenswrapper[4556]: I0218 09:19:31.727459 4556 patch_prober.go:28] interesting pod/machine-config-daemon-f76hs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 09:19:31 crc kubenswrapper[4556]: I0218 09:19:31.727547 4556 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 09:19:31 crc kubenswrapper[4556]: I0218 09:19:31.727620 4556 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" Feb 18 09:19:31 crc kubenswrapper[4556]: I0218 09:19:31.728703 4556 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7138060d5bb9e6802973b6e4fec551cb2729ba7d9b32c912776fd108e1b87ace"} pod="openshift-machine-config-operator/machine-config-daemon-f76hs" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 18 09:19:31 crc kubenswrapper[4556]: I0218 09:19:31.728778 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerName="machine-config-daemon" containerID="cri-o://7138060d5bb9e6802973b6e4fec551cb2729ba7d9b32c912776fd108e1b87ace" gracePeriod=600 Feb 18 09:19:31 crc kubenswrapper[4556]: I0218 09:19:31.878742 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 18 09:19:31 crc kubenswrapper[4556]: I0218 09:19:31.928641 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 18 09:19:32 crc kubenswrapper[4556]: I0218 09:19:32.257104 4556 generic.go:334] "Generic (PLEG): container finished" podID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerID="7138060d5bb9e6802973b6e4fec551cb2729ba7d9b32c912776fd108e1b87ace" exitCode=0 Feb 18 09:19:32 crc kubenswrapper[4556]: I0218 09:19:32.258275 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" event={"ID":"8dac7f27-d3d1-4778-9e54-f273035a1d37","Type":"ContainerDied","Data":"7138060d5bb9e6802973b6e4fec551cb2729ba7d9b32c912776fd108e1b87ace"} Feb 18 09:19:32 crc kubenswrapper[4556]: I0218 09:19:32.258378 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" event={"ID":"8dac7f27-d3d1-4778-9e54-f273035a1d37","Type":"ContainerStarted","Data":"16c417a1a589105242e4b889223fc0777a5d7312ffccba6c3cb31cf53da6f65a"} Feb 18 09:19:32 crc kubenswrapper[4556]: I0218 09:19:32.258412 4556 scope.go:117] "RemoveContainer" containerID="5eaeb58ab20518b6122148754032f26a9e6729291f4bb5d77b32e55450f0b0ee" Feb 18 09:19:33 crc kubenswrapper[4556]: I0218 09:19:33.267101 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="33ce5ec8-44c2-4b8c-865f-36ee8722d716" containerName="ceilometer-central-agent" containerID="cri-o://4d4dafcb48311a7d568ad32ddb27501224c2e007d1e076c372c237e47a26a39c" gracePeriod=30 Feb 18 09:19:33 crc kubenswrapper[4556]: I0218 09:19:33.267212 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="33ce5ec8-44c2-4b8c-865f-36ee8722d716" containerName="ceilometer-notification-agent" containerID="cri-o://d5a63205c3d5881f590fd4f8fd3992733d92abffbb8653897c64b67729fa0b65" gracePeriod=30 Feb 18 09:19:33 crc kubenswrapper[4556]: I0218 09:19:33.267126 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="33ce5ec8-44c2-4b8c-865f-36ee8722d716" containerName="sg-core" containerID="cri-o://4ba210fb11d2410c63309605c467e2da8848df4407665ca313864332e3878227" gracePeriod=30 Feb 18 09:19:33 crc kubenswrapper[4556]: I0218 09:19:33.267128 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="33ce5ec8-44c2-4b8c-865f-36ee8722d716" containerName="proxy-httpd" containerID="cri-o://965d06552f33c2639ae14fe6a55d0ebb59be0c8cce9bdb7d929e282c207a9b9c" gracePeriod=30 Feb 18 09:19:34 crc kubenswrapper[4556]: I0218 09:19:34.280502 4556 generic.go:334] "Generic (PLEG): container finished" podID="33ce5ec8-44c2-4b8c-865f-36ee8722d716" containerID="965d06552f33c2639ae14fe6a55d0ebb59be0c8cce9bdb7d929e282c207a9b9c" exitCode=0 Feb 18 09:19:34 crc kubenswrapper[4556]: I0218 09:19:34.280746 4556 generic.go:334] "Generic (PLEG): container finished" podID="33ce5ec8-44c2-4b8c-865f-36ee8722d716" containerID="4ba210fb11d2410c63309605c467e2da8848df4407665ca313864332e3878227" exitCode=2 Feb 18 09:19:34 crc kubenswrapper[4556]: I0218 09:19:34.280584 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33ce5ec8-44c2-4b8c-865f-36ee8722d716","Type":"ContainerDied","Data":"965d06552f33c2639ae14fe6a55d0ebb59be0c8cce9bdb7d929e282c207a9b9c"} Feb 18 09:19:34 crc kubenswrapper[4556]: I0218 09:19:34.280796 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33ce5ec8-44c2-4b8c-865f-36ee8722d716","Type":"ContainerDied","Data":"4ba210fb11d2410c63309605c467e2da8848df4407665ca313864332e3878227"} Feb 18 09:19:34 crc kubenswrapper[4556]: I0218 09:19:34.280808 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33ce5ec8-44c2-4b8c-865f-36ee8722d716","Type":"ContainerDied","Data":"4d4dafcb48311a7d568ad32ddb27501224c2e007d1e076c372c237e47a26a39c"} Feb 18 09:19:34 crc kubenswrapper[4556]: I0218 09:19:34.280757 4556 generic.go:334] "Generic (PLEG): container finished" podID="33ce5ec8-44c2-4b8c-865f-36ee8722d716" containerID="4d4dafcb48311a7d568ad32ddb27501224c2e007d1e076c372c237e47a26a39c" exitCode=0 Feb 18 09:19:34 crc kubenswrapper[4556]: I0218 09:19:34.282751 4556 generic.go:334] "Generic (PLEG): container finished" podID="f4a792c8-97e0-46f2-8a40-747a7cae58f5" containerID="69e41aca003b0c33c159e3090331559d090f11ef66400ffc6495e1d73d0945a8" exitCode=0 Feb 18 09:19:34 crc kubenswrapper[4556]: I0218 09:19:34.282781 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-wxltl" event={"ID":"f4a792c8-97e0-46f2-8a40-747a7cae58f5","Type":"ContainerDied","Data":"69e41aca003b0c33c159e3090331559d090f11ef66400ffc6495e1d73d0945a8"} Feb 18 09:19:35 crc kubenswrapper[4556]: I0218 09:19:35.573612 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-wxltl" Feb 18 09:19:35 crc kubenswrapper[4556]: I0218 09:19:35.674691 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4a792c8-97e0-46f2-8a40-747a7cae58f5-config-data\") pod \"f4a792c8-97e0-46f2-8a40-747a7cae58f5\" (UID: \"f4a792c8-97e0-46f2-8a40-747a7cae58f5\") " Feb 18 09:19:35 crc kubenswrapper[4556]: I0218 09:19:35.674986 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kkkw5\" (UniqueName: \"kubernetes.io/projected/f4a792c8-97e0-46f2-8a40-747a7cae58f5-kube-api-access-kkkw5\") pod \"f4a792c8-97e0-46f2-8a40-747a7cae58f5\" (UID: \"f4a792c8-97e0-46f2-8a40-747a7cae58f5\") " Feb 18 09:19:35 crc kubenswrapper[4556]: I0218 09:19:35.675091 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4a792c8-97e0-46f2-8a40-747a7cae58f5-combined-ca-bundle\") pod \"f4a792c8-97e0-46f2-8a40-747a7cae58f5\" (UID: \"f4a792c8-97e0-46f2-8a40-747a7cae58f5\") " Feb 18 09:19:35 crc kubenswrapper[4556]: I0218 09:19:35.675207 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4a792c8-97e0-46f2-8a40-747a7cae58f5-scripts\") pod \"f4a792c8-97e0-46f2-8a40-747a7cae58f5\" (UID: \"f4a792c8-97e0-46f2-8a40-747a7cae58f5\") " Feb 18 09:19:35 crc kubenswrapper[4556]: I0218 09:19:35.683210 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4a792c8-97e0-46f2-8a40-747a7cae58f5-kube-api-access-kkkw5" (OuterVolumeSpecName: "kube-api-access-kkkw5") pod "f4a792c8-97e0-46f2-8a40-747a7cae58f5" (UID: "f4a792c8-97e0-46f2-8a40-747a7cae58f5"). InnerVolumeSpecName "kube-api-access-kkkw5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:19:35 crc kubenswrapper[4556]: I0218 09:19:35.683231 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4a792c8-97e0-46f2-8a40-747a7cae58f5-scripts" (OuterVolumeSpecName: "scripts") pod "f4a792c8-97e0-46f2-8a40-747a7cae58f5" (UID: "f4a792c8-97e0-46f2-8a40-747a7cae58f5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:19:35 crc kubenswrapper[4556]: I0218 09:19:35.695403 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4a792c8-97e0-46f2-8a40-747a7cae58f5-config-data" (OuterVolumeSpecName: "config-data") pod "f4a792c8-97e0-46f2-8a40-747a7cae58f5" (UID: "f4a792c8-97e0-46f2-8a40-747a7cae58f5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:19:35 crc kubenswrapper[4556]: I0218 09:19:35.697323 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4a792c8-97e0-46f2-8a40-747a7cae58f5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f4a792c8-97e0-46f2-8a40-747a7cae58f5" (UID: "f4a792c8-97e0-46f2-8a40-747a7cae58f5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:19:35 crc kubenswrapper[4556]: I0218 09:19:35.777959 4556 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4a792c8-97e0-46f2-8a40-747a7cae58f5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:35 crc kubenswrapper[4556]: I0218 09:19:35.777999 4556 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4a792c8-97e0-46f2-8a40-747a7cae58f5-scripts\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:35 crc kubenswrapper[4556]: I0218 09:19:35.778010 4556 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4a792c8-97e0-46f2-8a40-747a7cae58f5-config-data\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:35 crc kubenswrapper[4556]: I0218 09:19:35.778022 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kkkw5\" (UniqueName: \"kubernetes.io/projected/f4a792c8-97e0-46f2-8a40-747a7cae58f5-kube-api-access-kkkw5\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:36 crc kubenswrapper[4556]: I0218 09:19:36.297408 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-wxltl" event={"ID":"f4a792c8-97e0-46f2-8a40-747a7cae58f5","Type":"ContainerDied","Data":"731029b0942c84e8e5914118f40931c2582da401dfac74804a29e88d44f46c29"} Feb 18 09:19:36 crc kubenswrapper[4556]: I0218 09:19:36.297708 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="731029b0942c84e8e5914118f40931c2582da401dfac74804a29e88d44f46c29" Feb 18 09:19:36 crc kubenswrapper[4556]: I0218 09:19:36.297449 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-wxltl" Feb 18 09:19:36 crc kubenswrapper[4556]: I0218 09:19:36.397540 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 18 09:19:36 crc kubenswrapper[4556]: E0218 09:19:36.398741 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a557635-d67d-4e93-a82c-990c34fe7d48" containerName="extract-utilities" Feb 18 09:19:36 crc kubenswrapper[4556]: I0218 09:19:36.398767 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a557635-d67d-4e93-a82c-990c34fe7d48" containerName="extract-utilities" Feb 18 09:19:36 crc kubenswrapper[4556]: E0218 09:19:36.398787 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a557635-d67d-4e93-a82c-990c34fe7d48" containerName="extract-content" Feb 18 09:19:36 crc kubenswrapper[4556]: I0218 09:19:36.398793 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a557635-d67d-4e93-a82c-990c34fe7d48" containerName="extract-content" Feb 18 09:19:36 crc kubenswrapper[4556]: E0218 09:19:36.398821 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a557635-d67d-4e93-a82c-990c34fe7d48" containerName="registry-server" Feb 18 09:19:36 crc kubenswrapper[4556]: I0218 09:19:36.398829 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a557635-d67d-4e93-a82c-990c34fe7d48" containerName="registry-server" Feb 18 09:19:36 crc kubenswrapper[4556]: E0218 09:19:36.398845 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4a792c8-97e0-46f2-8a40-747a7cae58f5" containerName="nova-cell0-conductor-db-sync" Feb 18 09:19:36 crc kubenswrapper[4556]: I0218 09:19:36.398851 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4a792c8-97e0-46f2-8a40-747a7cae58f5" containerName="nova-cell0-conductor-db-sync" Feb 18 09:19:36 crc kubenswrapper[4556]: I0218 09:19:36.399268 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4a792c8-97e0-46f2-8a40-747a7cae58f5" containerName="nova-cell0-conductor-db-sync" Feb 18 09:19:36 crc kubenswrapper[4556]: I0218 09:19:36.399296 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a557635-d67d-4e93-a82c-990c34fe7d48" containerName="registry-server" Feb 18 09:19:36 crc kubenswrapper[4556]: I0218 09:19:36.401622 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Feb 18 09:19:36 crc kubenswrapper[4556]: I0218 09:19:36.404427 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-wwj98" Feb 18 09:19:36 crc kubenswrapper[4556]: I0218 09:19:36.404834 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Feb 18 09:19:36 crc kubenswrapper[4556]: I0218 09:19:36.411988 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 18 09:19:36 crc kubenswrapper[4556]: I0218 09:19:36.494070 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64bcfb93-2db3-4350-97ee-671806ff7d00-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"64bcfb93-2db3-4350-97ee-671806ff7d00\") " pod="openstack/nova-cell0-conductor-0" Feb 18 09:19:36 crc kubenswrapper[4556]: I0218 09:19:36.494293 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nk9xf\" (UniqueName: \"kubernetes.io/projected/64bcfb93-2db3-4350-97ee-671806ff7d00-kube-api-access-nk9xf\") pod \"nova-cell0-conductor-0\" (UID: \"64bcfb93-2db3-4350-97ee-671806ff7d00\") " pod="openstack/nova-cell0-conductor-0" Feb 18 09:19:36 crc kubenswrapper[4556]: I0218 09:19:36.494358 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64bcfb93-2db3-4350-97ee-671806ff7d00-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"64bcfb93-2db3-4350-97ee-671806ff7d00\") " pod="openstack/nova-cell0-conductor-0" Feb 18 09:19:36 crc kubenswrapper[4556]: I0218 09:19:36.596092 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nk9xf\" (UniqueName: \"kubernetes.io/projected/64bcfb93-2db3-4350-97ee-671806ff7d00-kube-api-access-nk9xf\") pod \"nova-cell0-conductor-0\" (UID: \"64bcfb93-2db3-4350-97ee-671806ff7d00\") " pod="openstack/nova-cell0-conductor-0" Feb 18 09:19:36 crc kubenswrapper[4556]: I0218 09:19:36.596172 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64bcfb93-2db3-4350-97ee-671806ff7d00-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"64bcfb93-2db3-4350-97ee-671806ff7d00\") " pod="openstack/nova-cell0-conductor-0" Feb 18 09:19:36 crc kubenswrapper[4556]: I0218 09:19:36.596258 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64bcfb93-2db3-4350-97ee-671806ff7d00-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"64bcfb93-2db3-4350-97ee-671806ff7d00\") " pod="openstack/nova-cell0-conductor-0" Feb 18 09:19:36 crc kubenswrapper[4556]: I0218 09:19:36.602207 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64bcfb93-2db3-4350-97ee-671806ff7d00-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"64bcfb93-2db3-4350-97ee-671806ff7d00\") " pod="openstack/nova-cell0-conductor-0" Feb 18 09:19:36 crc kubenswrapper[4556]: I0218 09:19:36.603662 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64bcfb93-2db3-4350-97ee-671806ff7d00-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"64bcfb93-2db3-4350-97ee-671806ff7d00\") " pod="openstack/nova-cell0-conductor-0" Feb 18 09:19:36 crc kubenswrapper[4556]: I0218 09:19:36.610887 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nk9xf\" (UniqueName: \"kubernetes.io/projected/64bcfb93-2db3-4350-97ee-671806ff7d00-kube-api-access-nk9xf\") pod \"nova-cell0-conductor-0\" (UID: \"64bcfb93-2db3-4350-97ee-671806ff7d00\") " pod="openstack/nova-cell0-conductor-0" Feb 18 09:19:36 crc kubenswrapper[4556]: I0218 09:19:36.726056 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Feb 18 09:19:37 crc kubenswrapper[4556]: I0218 09:19:37.107036 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 18 09:19:37 crc kubenswrapper[4556]: W0218 09:19:37.113011 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod64bcfb93_2db3_4350_97ee_671806ff7d00.slice/crio-545822f176a1e7adbcf660902c9318ef3cfae25bb8bcb7a25c3a924990cf95d0 WatchSource:0}: Error finding container 545822f176a1e7adbcf660902c9318ef3cfae25bb8bcb7a25c3a924990cf95d0: Status 404 returned error can't find the container with id 545822f176a1e7adbcf660902c9318ef3cfae25bb8bcb7a25c3a924990cf95d0 Feb 18 09:19:37 crc kubenswrapper[4556]: I0218 09:19:37.306731 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"64bcfb93-2db3-4350-97ee-671806ff7d00","Type":"ContainerStarted","Data":"b84cc6dc97caa2e7551d51cf5a1d6790829ba11ded228af0892bd60c1b92df26"} Feb 18 09:19:37 crc kubenswrapper[4556]: I0218 09:19:37.306777 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"64bcfb93-2db3-4350-97ee-671806ff7d00","Type":"ContainerStarted","Data":"545822f176a1e7adbcf660902c9318ef3cfae25bb8bcb7a25c3a924990cf95d0"} Feb 18 09:19:37 crc kubenswrapper[4556]: I0218 09:19:37.306870 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Feb 18 09:19:37 crc kubenswrapper[4556]: I0218 09:19:37.325817 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=1.3258034699999999 podStartE2EDuration="1.32580347s" podCreationTimestamp="2026-02-18 09:19:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:19:37.323293214 +0000 UTC m=+934.340254193" watchObservedRunningTime="2026-02-18 09:19:37.32580347 +0000 UTC m=+934.342764449" Feb 18 09:19:37 crc kubenswrapper[4556]: I0218 09:19:37.784138 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 18 09:19:37 crc kubenswrapper[4556]: I0218 09:19:37.924074 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33ce5ec8-44c2-4b8c-865f-36ee8722d716-config-data\") pod \"33ce5ec8-44c2-4b8c-865f-36ee8722d716\" (UID: \"33ce5ec8-44c2-4b8c-865f-36ee8722d716\") " Feb 18 09:19:37 crc kubenswrapper[4556]: I0218 09:19:37.924204 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33ce5ec8-44c2-4b8c-865f-36ee8722d716-combined-ca-bundle\") pod \"33ce5ec8-44c2-4b8c-865f-36ee8722d716\" (UID: \"33ce5ec8-44c2-4b8c-865f-36ee8722d716\") " Feb 18 09:19:37 crc kubenswrapper[4556]: I0218 09:19:37.924241 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33ce5ec8-44c2-4b8c-865f-36ee8722d716-run-httpd\") pod \"33ce5ec8-44c2-4b8c-865f-36ee8722d716\" (UID: \"33ce5ec8-44c2-4b8c-865f-36ee8722d716\") " Feb 18 09:19:37 crc kubenswrapper[4556]: I0218 09:19:37.924261 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5gxl4\" (UniqueName: \"kubernetes.io/projected/33ce5ec8-44c2-4b8c-865f-36ee8722d716-kube-api-access-5gxl4\") pod \"33ce5ec8-44c2-4b8c-865f-36ee8722d716\" (UID: \"33ce5ec8-44c2-4b8c-865f-36ee8722d716\") " Feb 18 09:19:37 crc kubenswrapper[4556]: I0218 09:19:37.924310 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/33ce5ec8-44c2-4b8c-865f-36ee8722d716-sg-core-conf-yaml\") pod \"33ce5ec8-44c2-4b8c-865f-36ee8722d716\" (UID: \"33ce5ec8-44c2-4b8c-865f-36ee8722d716\") " Feb 18 09:19:37 crc kubenswrapper[4556]: I0218 09:19:37.924396 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33ce5ec8-44c2-4b8c-865f-36ee8722d716-scripts\") pod \"33ce5ec8-44c2-4b8c-865f-36ee8722d716\" (UID: \"33ce5ec8-44c2-4b8c-865f-36ee8722d716\") " Feb 18 09:19:37 crc kubenswrapper[4556]: I0218 09:19:37.924465 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33ce5ec8-44c2-4b8c-865f-36ee8722d716-log-httpd\") pod \"33ce5ec8-44c2-4b8c-865f-36ee8722d716\" (UID: \"33ce5ec8-44c2-4b8c-865f-36ee8722d716\") " Feb 18 09:19:37 crc kubenswrapper[4556]: I0218 09:19:37.925408 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33ce5ec8-44c2-4b8c-865f-36ee8722d716-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "33ce5ec8-44c2-4b8c-865f-36ee8722d716" (UID: "33ce5ec8-44c2-4b8c-865f-36ee8722d716"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:19:37 crc kubenswrapper[4556]: I0218 09:19:37.927920 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33ce5ec8-44c2-4b8c-865f-36ee8722d716-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "33ce5ec8-44c2-4b8c-865f-36ee8722d716" (UID: "33ce5ec8-44c2-4b8c-865f-36ee8722d716"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:19:37 crc kubenswrapper[4556]: I0218 09:19:37.931329 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33ce5ec8-44c2-4b8c-865f-36ee8722d716-kube-api-access-5gxl4" (OuterVolumeSpecName: "kube-api-access-5gxl4") pod "33ce5ec8-44c2-4b8c-865f-36ee8722d716" (UID: "33ce5ec8-44c2-4b8c-865f-36ee8722d716"). InnerVolumeSpecName "kube-api-access-5gxl4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:19:37 crc kubenswrapper[4556]: I0218 09:19:37.932449 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33ce5ec8-44c2-4b8c-865f-36ee8722d716-scripts" (OuterVolumeSpecName: "scripts") pod "33ce5ec8-44c2-4b8c-865f-36ee8722d716" (UID: "33ce5ec8-44c2-4b8c-865f-36ee8722d716"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:19:37 crc kubenswrapper[4556]: I0218 09:19:37.970352 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33ce5ec8-44c2-4b8c-865f-36ee8722d716-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "33ce5ec8-44c2-4b8c-865f-36ee8722d716" (UID: "33ce5ec8-44c2-4b8c-865f-36ee8722d716"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:19:37 crc kubenswrapper[4556]: I0218 09:19:37.994098 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33ce5ec8-44c2-4b8c-865f-36ee8722d716-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "33ce5ec8-44c2-4b8c-865f-36ee8722d716" (UID: "33ce5ec8-44c2-4b8c-865f-36ee8722d716"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.007583 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33ce5ec8-44c2-4b8c-865f-36ee8722d716-config-data" (OuterVolumeSpecName: "config-data") pod "33ce5ec8-44c2-4b8c-865f-36ee8722d716" (UID: "33ce5ec8-44c2-4b8c-865f-36ee8722d716"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.027098 4556 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33ce5ec8-44c2-4b8c-865f-36ee8722d716-scripts\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.027129 4556 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33ce5ec8-44c2-4b8c-865f-36ee8722d716-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.027140 4556 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33ce5ec8-44c2-4b8c-865f-36ee8722d716-config-data\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.027169 4556 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33ce5ec8-44c2-4b8c-865f-36ee8722d716-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.027180 4556 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33ce5ec8-44c2-4b8c-865f-36ee8722d716-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.027190 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5gxl4\" (UniqueName: \"kubernetes.io/projected/33ce5ec8-44c2-4b8c-865f-36ee8722d716-kube-api-access-5gxl4\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.027200 4556 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/33ce5ec8-44c2-4b8c-865f-36ee8722d716-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.317590 4556 generic.go:334] "Generic (PLEG): container finished" podID="33ce5ec8-44c2-4b8c-865f-36ee8722d716" containerID="d5a63205c3d5881f590fd4f8fd3992733d92abffbb8653897c64b67729fa0b65" exitCode=0 Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.317659 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.317687 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33ce5ec8-44c2-4b8c-865f-36ee8722d716","Type":"ContainerDied","Data":"d5a63205c3d5881f590fd4f8fd3992733d92abffbb8653897c64b67729fa0b65"} Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.317726 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33ce5ec8-44c2-4b8c-865f-36ee8722d716","Type":"ContainerDied","Data":"404d04226b2299a9a4f3e0aa73e0e9bc8b4f17c70cae76cb1ca181c3af19d562"} Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.317744 4556 scope.go:117] "RemoveContainer" containerID="965d06552f33c2639ae14fe6a55d0ebb59be0c8cce9bdb7d929e282c207a9b9c" Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.338096 4556 scope.go:117] "RemoveContainer" containerID="4ba210fb11d2410c63309605c467e2da8848df4407665ca313864332e3878227" Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.342921 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.357581 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.359922 4556 scope.go:117] "RemoveContainer" containerID="d5a63205c3d5881f590fd4f8fd3992733d92abffbb8653897c64b67729fa0b65" Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.374858 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 18 09:19:38 crc kubenswrapper[4556]: E0218 09:19:38.375245 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33ce5ec8-44c2-4b8c-865f-36ee8722d716" containerName="proxy-httpd" Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.375274 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="33ce5ec8-44c2-4b8c-865f-36ee8722d716" containerName="proxy-httpd" Feb 18 09:19:38 crc kubenswrapper[4556]: E0218 09:19:38.375289 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33ce5ec8-44c2-4b8c-865f-36ee8722d716" containerName="sg-core" Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.375296 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="33ce5ec8-44c2-4b8c-865f-36ee8722d716" containerName="sg-core" Feb 18 09:19:38 crc kubenswrapper[4556]: E0218 09:19:38.375310 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33ce5ec8-44c2-4b8c-865f-36ee8722d716" containerName="ceilometer-central-agent" Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.375316 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="33ce5ec8-44c2-4b8c-865f-36ee8722d716" containerName="ceilometer-central-agent" Feb 18 09:19:38 crc kubenswrapper[4556]: E0218 09:19:38.376604 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33ce5ec8-44c2-4b8c-865f-36ee8722d716" containerName="ceilometer-notification-agent" Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.376623 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="33ce5ec8-44c2-4b8c-865f-36ee8722d716" containerName="ceilometer-notification-agent" Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.376824 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="33ce5ec8-44c2-4b8c-865f-36ee8722d716" containerName="ceilometer-notification-agent" Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.376841 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="33ce5ec8-44c2-4b8c-865f-36ee8722d716" containerName="proxy-httpd" Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.376851 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="33ce5ec8-44c2-4b8c-865f-36ee8722d716" containerName="sg-core" Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.376873 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="33ce5ec8-44c2-4b8c-865f-36ee8722d716" containerName="ceilometer-central-agent" Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.378284 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.380703 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.380878 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.386092 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.389081 4556 scope.go:117] "RemoveContainer" containerID="4d4dafcb48311a7d568ad32ddb27501224c2e007d1e076c372c237e47a26a39c" Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.422036 4556 scope.go:117] "RemoveContainer" containerID="965d06552f33c2639ae14fe6a55d0ebb59be0c8cce9bdb7d929e282c207a9b9c" Feb 18 09:19:38 crc kubenswrapper[4556]: E0218 09:19:38.424304 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"965d06552f33c2639ae14fe6a55d0ebb59be0c8cce9bdb7d929e282c207a9b9c\": container with ID starting with 965d06552f33c2639ae14fe6a55d0ebb59be0c8cce9bdb7d929e282c207a9b9c not found: ID does not exist" containerID="965d06552f33c2639ae14fe6a55d0ebb59be0c8cce9bdb7d929e282c207a9b9c" Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.424353 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"965d06552f33c2639ae14fe6a55d0ebb59be0c8cce9bdb7d929e282c207a9b9c"} err="failed to get container status \"965d06552f33c2639ae14fe6a55d0ebb59be0c8cce9bdb7d929e282c207a9b9c\": rpc error: code = NotFound desc = could not find container \"965d06552f33c2639ae14fe6a55d0ebb59be0c8cce9bdb7d929e282c207a9b9c\": container with ID starting with 965d06552f33c2639ae14fe6a55d0ebb59be0c8cce9bdb7d929e282c207a9b9c not found: ID does not exist" Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.424378 4556 scope.go:117] "RemoveContainer" containerID="4ba210fb11d2410c63309605c467e2da8848df4407665ca313864332e3878227" Feb 18 09:19:38 crc kubenswrapper[4556]: E0218 09:19:38.424707 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ba210fb11d2410c63309605c467e2da8848df4407665ca313864332e3878227\": container with ID starting with 4ba210fb11d2410c63309605c467e2da8848df4407665ca313864332e3878227 not found: ID does not exist" containerID="4ba210fb11d2410c63309605c467e2da8848df4407665ca313864332e3878227" Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.424728 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ba210fb11d2410c63309605c467e2da8848df4407665ca313864332e3878227"} err="failed to get container status \"4ba210fb11d2410c63309605c467e2da8848df4407665ca313864332e3878227\": rpc error: code = NotFound desc = could not find container \"4ba210fb11d2410c63309605c467e2da8848df4407665ca313864332e3878227\": container with ID starting with 4ba210fb11d2410c63309605c467e2da8848df4407665ca313864332e3878227 not found: ID does not exist" Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.424759 4556 scope.go:117] "RemoveContainer" containerID="d5a63205c3d5881f590fd4f8fd3992733d92abffbb8653897c64b67729fa0b65" Feb 18 09:19:38 crc kubenswrapper[4556]: E0218 09:19:38.425022 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5a63205c3d5881f590fd4f8fd3992733d92abffbb8653897c64b67729fa0b65\": container with ID starting with d5a63205c3d5881f590fd4f8fd3992733d92abffbb8653897c64b67729fa0b65 not found: ID does not exist" containerID="d5a63205c3d5881f590fd4f8fd3992733d92abffbb8653897c64b67729fa0b65" Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.425039 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5a63205c3d5881f590fd4f8fd3992733d92abffbb8653897c64b67729fa0b65"} err="failed to get container status \"d5a63205c3d5881f590fd4f8fd3992733d92abffbb8653897c64b67729fa0b65\": rpc error: code = NotFound desc = could not find container \"d5a63205c3d5881f590fd4f8fd3992733d92abffbb8653897c64b67729fa0b65\": container with ID starting with d5a63205c3d5881f590fd4f8fd3992733d92abffbb8653897c64b67729fa0b65 not found: ID does not exist" Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.425074 4556 scope.go:117] "RemoveContainer" containerID="4d4dafcb48311a7d568ad32ddb27501224c2e007d1e076c372c237e47a26a39c" Feb 18 09:19:38 crc kubenswrapper[4556]: E0218 09:19:38.425420 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d4dafcb48311a7d568ad32ddb27501224c2e007d1e076c372c237e47a26a39c\": container with ID starting with 4d4dafcb48311a7d568ad32ddb27501224c2e007d1e076c372c237e47a26a39c not found: ID does not exist" containerID="4d4dafcb48311a7d568ad32ddb27501224c2e007d1e076c372c237e47a26a39c" Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.425440 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d4dafcb48311a7d568ad32ddb27501224c2e007d1e076c372c237e47a26a39c"} err="failed to get container status \"4d4dafcb48311a7d568ad32ddb27501224c2e007d1e076c372c237e47a26a39c\": rpc error: code = NotFound desc = could not find container \"4d4dafcb48311a7d568ad32ddb27501224c2e007d1e076c372c237e47a26a39c\": container with ID starting with 4d4dafcb48311a7d568ad32ddb27501224c2e007d1e076c372c237e47a26a39c not found: ID does not exist" Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.536706 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f16caa9f-5685-4f11-a112-65d1ed02241b-run-httpd\") pod \"ceilometer-0\" (UID: \"f16caa9f-5685-4f11-a112-65d1ed02241b\") " pod="openstack/ceilometer-0" Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.536962 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f16caa9f-5685-4f11-a112-65d1ed02241b-config-data\") pod \"ceilometer-0\" (UID: \"f16caa9f-5685-4f11-a112-65d1ed02241b\") " pod="openstack/ceilometer-0" Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.537096 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f16caa9f-5685-4f11-a112-65d1ed02241b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f16caa9f-5685-4f11-a112-65d1ed02241b\") " pod="openstack/ceilometer-0" Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.537198 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f16caa9f-5685-4f11-a112-65d1ed02241b-scripts\") pod \"ceilometer-0\" (UID: \"f16caa9f-5685-4f11-a112-65d1ed02241b\") " pod="openstack/ceilometer-0" Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.537279 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f16caa9f-5685-4f11-a112-65d1ed02241b-log-httpd\") pod \"ceilometer-0\" (UID: \"f16caa9f-5685-4f11-a112-65d1ed02241b\") " pod="openstack/ceilometer-0" Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.537373 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hjm9\" (UniqueName: \"kubernetes.io/projected/f16caa9f-5685-4f11-a112-65d1ed02241b-kube-api-access-9hjm9\") pod \"ceilometer-0\" (UID: \"f16caa9f-5685-4f11-a112-65d1ed02241b\") " pod="openstack/ceilometer-0" Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.537469 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f16caa9f-5685-4f11-a112-65d1ed02241b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f16caa9f-5685-4f11-a112-65d1ed02241b\") " pod="openstack/ceilometer-0" Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.638898 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f16caa9f-5685-4f11-a112-65d1ed02241b-config-data\") pod \"ceilometer-0\" (UID: \"f16caa9f-5685-4f11-a112-65d1ed02241b\") " pod="openstack/ceilometer-0" Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.638970 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f16caa9f-5685-4f11-a112-65d1ed02241b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f16caa9f-5685-4f11-a112-65d1ed02241b\") " pod="openstack/ceilometer-0" Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.639000 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f16caa9f-5685-4f11-a112-65d1ed02241b-scripts\") pod \"ceilometer-0\" (UID: \"f16caa9f-5685-4f11-a112-65d1ed02241b\") " pod="openstack/ceilometer-0" Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.639034 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f16caa9f-5685-4f11-a112-65d1ed02241b-log-httpd\") pod \"ceilometer-0\" (UID: \"f16caa9f-5685-4f11-a112-65d1ed02241b\") " pod="openstack/ceilometer-0" Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.639055 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hjm9\" (UniqueName: \"kubernetes.io/projected/f16caa9f-5685-4f11-a112-65d1ed02241b-kube-api-access-9hjm9\") pod \"ceilometer-0\" (UID: \"f16caa9f-5685-4f11-a112-65d1ed02241b\") " pod="openstack/ceilometer-0" Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.639092 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f16caa9f-5685-4f11-a112-65d1ed02241b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f16caa9f-5685-4f11-a112-65d1ed02241b\") " pod="openstack/ceilometer-0" Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.639135 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f16caa9f-5685-4f11-a112-65d1ed02241b-run-httpd\") pod \"ceilometer-0\" (UID: \"f16caa9f-5685-4f11-a112-65d1ed02241b\") " pod="openstack/ceilometer-0" Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.639588 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f16caa9f-5685-4f11-a112-65d1ed02241b-run-httpd\") pod \"ceilometer-0\" (UID: \"f16caa9f-5685-4f11-a112-65d1ed02241b\") " pod="openstack/ceilometer-0" Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.639774 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f16caa9f-5685-4f11-a112-65d1ed02241b-log-httpd\") pod \"ceilometer-0\" (UID: \"f16caa9f-5685-4f11-a112-65d1ed02241b\") " pod="openstack/ceilometer-0" Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.642208 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f16caa9f-5685-4f11-a112-65d1ed02241b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f16caa9f-5685-4f11-a112-65d1ed02241b\") " pod="openstack/ceilometer-0" Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.642884 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f16caa9f-5685-4f11-a112-65d1ed02241b-config-data\") pod \"ceilometer-0\" (UID: \"f16caa9f-5685-4f11-a112-65d1ed02241b\") " pod="openstack/ceilometer-0" Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.643308 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f16caa9f-5685-4f11-a112-65d1ed02241b-scripts\") pod \"ceilometer-0\" (UID: \"f16caa9f-5685-4f11-a112-65d1ed02241b\") " pod="openstack/ceilometer-0" Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.645725 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f16caa9f-5685-4f11-a112-65d1ed02241b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f16caa9f-5685-4f11-a112-65d1ed02241b\") " pod="openstack/ceilometer-0" Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.659659 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hjm9\" (UniqueName: \"kubernetes.io/projected/f16caa9f-5685-4f11-a112-65d1ed02241b-kube-api-access-9hjm9\") pod \"ceilometer-0\" (UID: \"f16caa9f-5685-4f11-a112-65d1ed02241b\") " pod="openstack/ceilometer-0" Feb 18 09:19:38 crc kubenswrapper[4556]: I0218 09:19:38.691811 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 18 09:19:39 crc kubenswrapper[4556]: I0218 09:19:39.082377 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 18 09:19:39 crc kubenswrapper[4556]: W0218 09:19:39.088850 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf16caa9f_5685_4f11_a112_65d1ed02241b.slice/crio-a7ffd15e0245698e2eacf432969bb0cbb3efa327a6e2ed55ec4bdb14ce07ae93 WatchSource:0}: Error finding container a7ffd15e0245698e2eacf432969bb0cbb3efa327a6e2ed55ec4bdb14ce07ae93: Status 404 returned error can't find the container with id a7ffd15e0245698e2eacf432969bb0cbb3efa327a6e2ed55ec4bdb14ce07ae93 Feb 18 09:19:39 crc kubenswrapper[4556]: I0218 09:19:39.299493 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33ce5ec8-44c2-4b8c-865f-36ee8722d716" path="/var/lib/kubelet/pods/33ce5ec8-44c2-4b8c-865f-36ee8722d716/volumes" Feb 18 09:19:39 crc kubenswrapper[4556]: I0218 09:19:39.326777 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f16caa9f-5685-4f11-a112-65d1ed02241b","Type":"ContainerStarted","Data":"a7ffd15e0245698e2eacf432969bb0cbb3efa327a6e2ed55ec4bdb14ce07ae93"} Feb 18 09:19:40 crc kubenswrapper[4556]: I0218 09:19:40.336166 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f16caa9f-5685-4f11-a112-65d1ed02241b","Type":"ContainerStarted","Data":"9bcf089e0dd023a0eb2a47ebef6f90f371b331baf6e3080a9a09414e4ae043b7"} Feb 18 09:19:40 crc kubenswrapper[4556]: I0218 09:19:40.403318 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jsb7c"] Feb 18 09:19:40 crc kubenswrapper[4556]: I0218 09:19:40.405021 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jsb7c" Feb 18 09:19:40 crc kubenswrapper[4556]: I0218 09:19:40.415403 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jsb7c"] Feb 18 09:19:40 crc kubenswrapper[4556]: I0218 09:19:40.575710 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b22db00-3e66-4afe-bce4-20f733982dc8-utilities\") pod \"community-operators-jsb7c\" (UID: \"6b22db00-3e66-4afe-bce4-20f733982dc8\") " pod="openshift-marketplace/community-operators-jsb7c" Feb 18 09:19:40 crc kubenswrapper[4556]: I0218 09:19:40.576039 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6k8dh\" (UniqueName: \"kubernetes.io/projected/6b22db00-3e66-4afe-bce4-20f733982dc8-kube-api-access-6k8dh\") pod \"community-operators-jsb7c\" (UID: \"6b22db00-3e66-4afe-bce4-20f733982dc8\") " pod="openshift-marketplace/community-operators-jsb7c" Feb 18 09:19:40 crc kubenswrapper[4556]: I0218 09:19:40.576093 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b22db00-3e66-4afe-bce4-20f733982dc8-catalog-content\") pod \"community-operators-jsb7c\" (UID: \"6b22db00-3e66-4afe-bce4-20f733982dc8\") " pod="openshift-marketplace/community-operators-jsb7c" Feb 18 09:19:40 crc kubenswrapper[4556]: I0218 09:19:40.677988 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b22db00-3e66-4afe-bce4-20f733982dc8-utilities\") pod \"community-operators-jsb7c\" (UID: \"6b22db00-3e66-4afe-bce4-20f733982dc8\") " pod="openshift-marketplace/community-operators-jsb7c" Feb 18 09:19:40 crc kubenswrapper[4556]: I0218 09:19:40.678076 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6k8dh\" (UniqueName: \"kubernetes.io/projected/6b22db00-3e66-4afe-bce4-20f733982dc8-kube-api-access-6k8dh\") pod \"community-operators-jsb7c\" (UID: \"6b22db00-3e66-4afe-bce4-20f733982dc8\") " pod="openshift-marketplace/community-operators-jsb7c" Feb 18 09:19:40 crc kubenswrapper[4556]: I0218 09:19:40.678126 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b22db00-3e66-4afe-bce4-20f733982dc8-catalog-content\") pod \"community-operators-jsb7c\" (UID: \"6b22db00-3e66-4afe-bce4-20f733982dc8\") " pod="openshift-marketplace/community-operators-jsb7c" Feb 18 09:19:40 crc kubenswrapper[4556]: I0218 09:19:40.678694 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b22db00-3e66-4afe-bce4-20f733982dc8-utilities\") pod \"community-operators-jsb7c\" (UID: \"6b22db00-3e66-4afe-bce4-20f733982dc8\") " pod="openshift-marketplace/community-operators-jsb7c" Feb 18 09:19:40 crc kubenswrapper[4556]: I0218 09:19:40.679035 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b22db00-3e66-4afe-bce4-20f733982dc8-catalog-content\") pod \"community-operators-jsb7c\" (UID: \"6b22db00-3e66-4afe-bce4-20f733982dc8\") " pod="openshift-marketplace/community-operators-jsb7c" Feb 18 09:19:40 crc kubenswrapper[4556]: I0218 09:19:40.700025 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6k8dh\" (UniqueName: \"kubernetes.io/projected/6b22db00-3e66-4afe-bce4-20f733982dc8-kube-api-access-6k8dh\") pod \"community-operators-jsb7c\" (UID: \"6b22db00-3e66-4afe-bce4-20f733982dc8\") " pod="openshift-marketplace/community-operators-jsb7c" Feb 18 09:19:40 crc kubenswrapper[4556]: I0218 09:19:40.760685 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jsb7c" Feb 18 09:19:41 crc kubenswrapper[4556]: I0218 09:19:41.267888 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jsb7c"] Feb 18 09:19:41 crc kubenswrapper[4556]: I0218 09:19:41.349908 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f16caa9f-5685-4f11-a112-65d1ed02241b","Type":"ContainerStarted","Data":"90309d247b8912b399ac4dd0bcf7c92ed86b4ad0f5768d5cc4813794ab056e5a"} Feb 18 09:19:41 crc kubenswrapper[4556]: I0218 09:19:41.349962 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f16caa9f-5685-4f11-a112-65d1ed02241b","Type":"ContainerStarted","Data":"6d63c4d4d746825902bcb72d7e5602ab09335c7b337d7c0363dba47c1d39b045"} Feb 18 09:19:41 crc kubenswrapper[4556]: I0218 09:19:41.350976 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jsb7c" event={"ID":"6b22db00-3e66-4afe-bce4-20f733982dc8","Type":"ContainerStarted","Data":"afad6520f76efafd7eb26488ce9c0436cad42933317dc32cbd9f1f02e40f86ae"} Feb 18 09:19:42 crc kubenswrapper[4556]: I0218 09:19:42.361579 4556 generic.go:334] "Generic (PLEG): container finished" podID="6b22db00-3e66-4afe-bce4-20f733982dc8" containerID="a413f74d1082552e6ab9b8a637cc714fc127f96c55cc673c48ed0415267d8639" exitCode=0 Feb 18 09:19:42 crc kubenswrapper[4556]: I0218 09:19:42.361625 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jsb7c" event={"ID":"6b22db00-3e66-4afe-bce4-20f733982dc8","Type":"ContainerDied","Data":"a413f74d1082552e6ab9b8a637cc714fc127f96c55cc673c48ed0415267d8639"} Feb 18 09:19:43 crc kubenswrapper[4556]: I0218 09:19:43.372093 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jsb7c" event={"ID":"6b22db00-3e66-4afe-bce4-20f733982dc8","Type":"ContainerStarted","Data":"f3a70532a2f42c1494230cf1962658d0f84a2066e8f5712ccb3eb037e0494ada"} Feb 18 09:19:43 crc kubenswrapper[4556]: I0218 09:19:43.378347 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f16caa9f-5685-4f11-a112-65d1ed02241b","Type":"ContainerStarted","Data":"cb1f58413a0abdbb6acb88e4efccd9fa33ff1fcb407f6954739fcde9cbfd29bb"} Feb 18 09:19:43 crc kubenswrapper[4556]: I0218 09:19:43.378907 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 18 09:19:43 crc kubenswrapper[4556]: I0218 09:19:43.417505 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.168734821 podStartE2EDuration="5.417492588s" podCreationTimestamp="2026-02-18 09:19:38 +0000 UTC" firstStartedPulling="2026-02-18 09:19:39.09199752 +0000 UTC m=+936.108958500" lastFinishedPulling="2026-02-18 09:19:42.340755287 +0000 UTC m=+939.357716267" observedRunningTime="2026-02-18 09:19:43.415682783 +0000 UTC m=+940.432643764" watchObservedRunningTime="2026-02-18 09:19:43.417492588 +0000 UTC m=+940.434453568" Feb 18 09:19:44 crc kubenswrapper[4556]: I0218 09:19:44.387522 4556 generic.go:334] "Generic (PLEG): container finished" podID="6b22db00-3e66-4afe-bce4-20f733982dc8" containerID="f3a70532a2f42c1494230cf1962658d0f84a2066e8f5712ccb3eb037e0494ada" exitCode=0 Feb 18 09:19:44 crc kubenswrapper[4556]: I0218 09:19:44.388855 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jsb7c" event={"ID":"6b22db00-3e66-4afe-bce4-20f733982dc8","Type":"ContainerDied","Data":"f3a70532a2f42c1494230cf1962658d0f84a2066e8f5712ccb3eb037e0494ada"} Feb 18 09:19:45 crc kubenswrapper[4556]: I0218 09:19:45.400262 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jsb7c" event={"ID":"6b22db00-3e66-4afe-bce4-20f733982dc8","Type":"ContainerStarted","Data":"307c12968fecf2b16c9f86102e3f089cb02c58d88fe6fb6914c31cecd27abf6d"} Feb 18 09:19:45 crc kubenswrapper[4556]: I0218 09:19:45.423089 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jsb7c" podStartSLOduration=2.92946921 podStartE2EDuration="5.423072669s" podCreationTimestamp="2026-02-18 09:19:40 +0000 UTC" firstStartedPulling="2026-02-18 09:19:42.368276285 +0000 UTC m=+939.385237264" lastFinishedPulling="2026-02-18 09:19:44.861879743 +0000 UTC m=+941.878840723" observedRunningTime="2026-02-18 09:19:45.418577798 +0000 UTC m=+942.435538778" watchObservedRunningTime="2026-02-18 09:19:45.423072669 +0000 UTC m=+942.440033649" Feb 18 09:19:46 crc kubenswrapper[4556]: I0218 09:19:46.767776 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.215542 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-dgmmz"] Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.216810 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-dgmmz" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.218351 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.218579 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.222045 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-dgmmz"] Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.242115 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57bln\" (UniqueName: \"kubernetes.io/projected/24dad0a1-29ed-49a7-be65-d024a02fcff4-kube-api-access-57bln\") pod \"nova-cell0-cell-mapping-dgmmz\" (UID: \"24dad0a1-29ed-49a7-be65-d024a02fcff4\") " pod="openstack/nova-cell0-cell-mapping-dgmmz" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.242184 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24dad0a1-29ed-49a7-be65-d024a02fcff4-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-dgmmz\" (UID: \"24dad0a1-29ed-49a7-be65-d024a02fcff4\") " pod="openstack/nova-cell0-cell-mapping-dgmmz" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.242446 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24dad0a1-29ed-49a7-be65-d024a02fcff4-scripts\") pod \"nova-cell0-cell-mapping-dgmmz\" (UID: \"24dad0a1-29ed-49a7-be65-d024a02fcff4\") " pod="openstack/nova-cell0-cell-mapping-dgmmz" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.242639 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24dad0a1-29ed-49a7-be65-d024a02fcff4-config-data\") pod \"nova-cell0-cell-mapping-dgmmz\" (UID: \"24dad0a1-29ed-49a7-be65-d024a02fcff4\") " pod="openstack/nova-cell0-cell-mapping-dgmmz" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.344857 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24dad0a1-29ed-49a7-be65-d024a02fcff4-config-data\") pod \"nova-cell0-cell-mapping-dgmmz\" (UID: \"24dad0a1-29ed-49a7-be65-d024a02fcff4\") " pod="openstack/nova-cell0-cell-mapping-dgmmz" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.344983 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57bln\" (UniqueName: \"kubernetes.io/projected/24dad0a1-29ed-49a7-be65-d024a02fcff4-kube-api-access-57bln\") pod \"nova-cell0-cell-mapping-dgmmz\" (UID: \"24dad0a1-29ed-49a7-be65-d024a02fcff4\") " pod="openstack/nova-cell0-cell-mapping-dgmmz" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.345020 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24dad0a1-29ed-49a7-be65-d024a02fcff4-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-dgmmz\" (UID: \"24dad0a1-29ed-49a7-be65-d024a02fcff4\") " pod="openstack/nova-cell0-cell-mapping-dgmmz" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.345124 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24dad0a1-29ed-49a7-be65-d024a02fcff4-scripts\") pod \"nova-cell0-cell-mapping-dgmmz\" (UID: \"24dad0a1-29ed-49a7-be65-d024a02fcff4\") " pod="openstack/nova-cell0-cell-mapping-dgmmz" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.350914 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24dad0a1-29ed-49a7-be65-d024a02fcff4-config-data\") pod \"nova-cell0-cell-mapping-dgmmz\" (UID: \"24dad0a1-29ed-49a7-be65-d024a02fcff4\") " pod="openstack/nova-cell0-cell-mapping-dgmmz" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.352624 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24dad0a1-29ed-49a7-be65-d024a02fcff4-scripts\") pod \"nova-cell0-cell-mapping-dgmmz\" (UID: \"24dad0a1-29ed-49a7-be65-d024a02fcff4\") " pod="openstack/nova-cell0-cell-mapping-dgmmz" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.353643 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24dad0a1-29ed-49a7-be65-d024a02fcff4-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-dgmmz\" (UID: \"24dad0a1-29ed-49a7-be65-d024a02fcff4\") " pod="openstack/nova-cell0-cell-mapping-dgmmz" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.367274 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57bln\" (UniqueName: \"kubernetes.io/projected/24dad0a1-29ed-49a7-be65-d024a02fcff4-kube-api-access-57bln\") pod \"nova-cell0-cell-mapping-dgmmz\" (UID: \"24dad0a1-29ed-49a7-be65-d024a02fcff4\") " pod="openstack/nova-cell0-cell-mapping-dgmmz" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.392506 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.393992 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.397871 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.412178 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.476285 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.477439 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.481014 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.484062 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.485840 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.491873 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.493599 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.517708 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.538339 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-dgmmz" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.552972 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d64f15b2-c072-4bf1-b60b-3e73d6df2414-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d64f15b2-c072-4bf1-b60b-3e73d6df2414\") " pod="openstack/nova-api-0" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.554397 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d64f15b2-c072-4bf1-b60b-3e73d6df2414-logs\") pod \"nova-api-0\" (UID: \"d64f15b2-c072-4bf1-b60b-3e73d6df2414\") " pod="openstack/nova-api-0" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.554451 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d64f15b2-c072-4bf1-b60b-3e73d6df2414-config-data\") pod \"nova-api-0\" (UID: \"d64f15b2-c072-4bf1-b60b-3e73d6df2414\") " pod="openstack/nova-api-0" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.554536 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qcslx\" (UniqueName: \"kubernetes.io/projected/d64f15b2-c072-4bf1-b60b-3e73d6df2414-kube-api-access-qcslx\") pod \"nova-api-0\" (UID: \"d64f15b2-c072-4bf1-b60b-3e73d6df2414\") " pod="openstack/nova-api-0" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.623363 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.624594 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.627739 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.642097 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-868bc9dc59-827zj"] Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.643925 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-868bc9dc59-827zj" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.656735 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d64f15b2-c072-4bf1-b60b-3e73d6df2414-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d64f15b2-c072-4bf1-b60b-3e73d6df2414\") " pod="openstack/nova-api-0" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.656815 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d64f15b2-c072-4bf1-b60b-3e73d6df2414-logs\") pod \"nova-api-0\" (UID: \"d64f15b2-c072-4bf1-b60b-3e73d6df2414\") " pod="openstack/nova-api-0" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.656851 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6d6r\" (UniqueName: \"kubernetes.io/projected/2330f995-37c7-4b3e-a465-4e682349346f-kube-api-access-b6d6r\") pod \"nova-metadata-0\" (UID: \"2330f995-37c7-4b3e-a465-4e682349346f\") " pod="openstack/nova-metadata-0" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.656869 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d64f15b2-c072-4bf1-b60b-3e73d6df2414-config-data\") pod \"nova-api-0\" (UID: \"d64f15b2-c072-4bf1-b60b-3e73d6df2414\") " pod="openstack/nova-api-0" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.656885 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2330f995-37c7-4b3e-a465-4e682349346f-config-data\") pod \"nova-metadata-0\" (UID: \"2330f995-37c7-4b3e-a465-4e682349346f\") " pod="openstack/nova-metadata-0" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.656902 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2330f995-37c7-4b3e-a465-4e682349346f-logs\") pod \"nova-metadata-0\" (UID: \"2330f995-37c7-4b3e-a465-4e682349346f\") " pod="openstack/nova-metadata-0" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.656960 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67a3f575-2f76-4447-b777-25b0b93ad11a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"67a3f575-2f76-4447-b777-25b0b93ad11a\") " pod="openstack/nova-scheduler-0" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.656985 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qcslx\" (UniqueName: \"kubernetes.io/projected/d64f15b2-c072-4bf1-b60b-3e73d6df2414-kube-api-access-qcslx\") pod \"nova-api-0\" (UID: \"d64f15b2-c072-4bf1-b60b-3e73d6df2414\") " pod="openstack/nova-api-0" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.657073 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2330f995-37c7-4b3e-a465-4e682349346f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2330f995-37c7-4b3e-a465-4e682349346f\") " pod="openstack/nova-metadata-0" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.658481 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ct5gz\" (UniqueName: \"kubernetes.io/projected/67a3f575-2f76-4447-b777-25b0b93ad11a-kube-api-access-ct5gz\") pod \"nova-scheduler-0\" (UID: \"67a3f575-2f76-4447-b777-25b0b93ad11a\") " pod="openstack/nova-scheduler-0" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.658570 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d64f15b2-c072-4bf1-b60b-3e73d6df2414-logs\") pod \"nova-api-0\" (UID: \"d64f15b2-c072-4bf1-b60b-3e73d6df2414\") " pod="openstack/nova-api-0" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.658647 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67a3f575-2f76-4447-b777-25b0b93ad11a-config-data\") pod \"nova-scheduler-0\" (UID: \"67a3f575-2f76-4447-b777-25b0b93ad11a\") " pod="openstack/nova-scheduler-0" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.664204 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d64f15b2-c072-4bf1-b60b-3e73d6df2414-config-data\") pod \"nova-api-0\" (UID: \"d64f15b2-c072-4bf1-b60b-3e73d6df2414\") " pod="openstack/nova-api-0" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.672492 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d64f15b2-c072-4bf1-b60b-3e73d6df2414-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d64f15b2-c072-4bf1-b60b-3e73d6df2414\") " pod="openstack/nova-api-0" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.676976 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.677463 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qcslx\" (UniqueName: \"kubernetes.io/projected/d64f15b2-c072-4bf1-b60b-3e73d6df2414-kube-api-access-qcslx\") pod \"nova-api-0\" (UID: \"d64f15b2-c072-4bf1-b60b-3e73d6df2414\") " pod="openstack/nova-api-0" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.696652 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-868bc9dc59-827zj"] Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.755772 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.761120 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5c2e5d6b-1b2f-4183-b654-908203d4b1d3-dns-svc\") pod \"dnsmasq-dns-868bc9dc59-827zj\" (UID: \"5c2e5d6b-1b2f-4183-b654-908203d4b1d3\") " pod="openstack/dnsmasq-dns-868bc9dc59-827zj" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.761189 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2330f995-37c7-4b3e-a465-4e682349346f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2330f995-37c7-4b3e-a465-4e682349346f\") " pod="openstack/nova-metadata-0" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.761212 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8jpp\" (UniqueName: \"kubernetes.io/projected/4ea5a5ac-537e-40d2-a5dc-4cc2314bf097-kube-api-access-x8jpp\") pod \"nova-cell1-novncproxy-0\" (UID: \"4ea5a5ac-537e-40d2-a5dc-4cc2314bf097\") " pod="openstack/nova-cell1-novncproxy-0" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.762732 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5c2e5d6b-1b2f-4183-b654-908203d4b1d3-ovsdbserver-nb\") pod \"dnsmasq-dns-868bc9dc59-827zj\" (UID: \"5c2e5d6b-1b2f-4183-b654-908203d4b1d3\") " pod="openstack/dnsmasq-dns-868bc9dc59-827zj" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.762782 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ct5gz\" (UniqueName: \"kubernetes.io/projected/67a3f575-2f76-4447-b777-25b0b93ad11a-kube-api-access-ct5gz\") pod \"nova-scheduler-0\" (UID: \"67a3f575-2f76-4447-b777-25b0b93ad11a\") " pod="openstack/nova-scheduler-0" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.762837 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5c2e5d6b-1b2f-4183-b654-908203d4b1d3-dns-swift-storage-0\") pod \"dnsmasq-dns-868bc9dc59-827zj\" (UID: \"5c2e5d6b-1b2f-4183-b654-908203d4b1d3\") " pod="openstack/dnsmasq-dns-868bc9dc59-827zj" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.762869 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67a3f575-2f76-4447-b777-25b0b93ad11a-config-data\") pod \"nova-scheduler-0\" (UID: \"67a3f575-2f76-4447-b777-25b0b93ad11a\") " pod="openstack/nova-scheduler-0" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.762936 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ea5a5ac-537e-40d2-a5dc-4cc2314bf097-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"4ea5a5ac-537e-40d2-a5dc-4cc2314bf097\") " pod="openstack/nova-cell1-novncproxy-0" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.762966 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5c2e5d6b-1b2f-4183-b654-908203d4b1d3-ovsdbserver-sb\") pod \"dnsmasq-dns-868bc9dc59-827zj\" (UID: \"5c2e5d6b-1b2f-4183-b654-908203d4b1d3\") " pod="openstack/dnsmasq-dns-868bc9dc59-827zj" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.763020 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6d6r\" (UniqueName: \"kubernetes.io/projected/2330f995-37c7-4b3e-a465-4e682349346f-kube-api-access-b6d6r\") pod \"nova-metadata-0\" (UID: \"2330f995-37c7-4b3e-a465-4e682349346f\") " pod="openstack/nova-metadata-0" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.763034 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ea5a5ac-537e-40d2-a5dc-4cc2314bf097-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"4ea5a5ac-537e-40d2-a5dc-4cc2314bf097\") " pod="openstack/nova-cell1-novncproxy-0" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.763050 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2330f995-37c7-4b3e-a465-4e682349346f-config-data\") pod \"nova-metadata-0\" (UID: \"2330f995-37c7-4b3e-a465-4e682349346f\") " pod="openstack/nova-metadata-0" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.763064 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2330f995-37c7-4b3e-a465-4e682349346f-logs\") pod \"nova-metadata-0\" (UID: \"2330f995-37c7-4b3e-a465-4e682349346f\") " pod="openstack/nova-metadata-0" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.763081 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c2e5d6b-1b2f-4183-b654-908203d4b1d3-config\") pod \"dnsmasq-dns-868bc9dc59-827zj\" (UID: \"5c2e5d6b-1b2f-4183-b654-908203d4b1d3\") " pod="openstack/dnsmasq-dns-868bc9dc59-827zj" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.763113 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2n26g\" (UniqueName: \"kubernetes.io/projected/5c2e5d6b-1b2f-4183-b654-908203d4b1d3-kube-api-access-2n26g\") pod \"dnsmasq-dns-868bc9dc59-827zj\" (UID: \"5c2e5d6b-1b2f-4183-b654-908203d4b1d3\") " pod="openstack/dnsmasq-dns-868bc9dc59-827zj" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.763140 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67a3f575-2f76-4447-b777-25b0b93ad11a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"67a3f575-2f76-4447-b777-25b0b93ad11a\") " pod="openstack/nova-scheduler-0" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.764521 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2330f995-37c7-4b3e-a465-4e682349346f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2330f995-37c7-4b3e-a465-4e682349346f\") " pod="openstack/nova-metadata-0" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.766303 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2330f995-37c7-4b3e-a465-4e682349346f-logs\") pod \"nova-metadata-0\" (UID: \"2330f995-37c7-4b3e-a465-4e682349346f\") " pod="openstack/nova-metadata-0" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.767087 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67a3f575-2f76-4447-b777-25b0b93ad11a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"67a3f575-2f76-4447-b777-25b0b93ad11a\") " pod="openstack/nova-scheduler-0" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.772647 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67a3f575-2f76-4447-b777-25b0b93ad11a-config-data\") pod \"nova-scheduler-0\" (UID: \"67a3f575-2f76-4447-b777-25b0b93ad11a\") " pod="openstack/nova-scheduler-0" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.772785 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2330f995-37c7-4b3e-a465-4e682349346f-config-data\") pod \"nova-metadata-0\" (UID: \"2330f995-37c7-4b3e-a465-4e682349346f\") " pod="openstack/nova-metadata-0" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.781121 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6d6r\" (UniqueName: \"kubernetes.io/projected/2330f995-37c7-4b3e-a465-4e682349346f-kube-api-access-b6d6r\") pod \"nova-metadata-0\" (UID: \"2330f995-37c7-4b3e-a465-4e682349346f\") " pod="openstack/nova-metadata-0" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.785255 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ct5gz\" (UniqueName: \"kubernetes.io/projected/67a3f575-2f76-4447-b777-25b0b93ad11a-kube-api-access-ct5gz\") pod \"nova-scheduler-0\" (UID: \"67a3f575-2f76-4447-b777-25b0b93ad11a\") " pod="openstack/nova-scheduler-0" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.794441 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.814925 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.865159 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5c2e5d6b-1b2f-4183-b654-908203d4b1d3-ovsdbserver-nb\") pod \"dnsmasq-dns-868bc9dc59-827zj\" (UID: \"5c2e5d6b-1b2f-4183-b654-908203d4b1d3\") " pod="openstack/dnsmasq-dns-868bc9dc59-827zj" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.865245 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5c2e5d6b-1b2f-4183-b654-908203d4b1d3-dns-swift-storage-0\") pod \"dnsmasq-dns-868bc9dc59-827zj\" (UID: \"5c2e5d6b-1b2f-4183-b654-908203d4b1d3\") " pod="openstack/dnsmasq-dns-868bc9dc59-827zj" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.865322 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ea5a5ac-537e-40d2-a5dc-4cc2314bf097-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"4ea5a5ac-537e-40d2-a5dc-4cc2314bf097\") " pod="openstack/nova-cell1-novncproxy-0" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.865349 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5c2e5d6b-1b2f-4183-b654-908203d4b1d3-ovsdbserver-sb\") pod \"dnsmasq-dns-868bc9dc59-827zj\" (UID: \"5c2e5d6b-1b2f-4183-b654-908203d4b1d3\") " pod="openstack/dnsmasq-dns-868bc9dc59-827zj" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.865399 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ea5a5ac-537e-40d2-a5dc-4cc2314bf097-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"4ea5a5ac-537e-40d2-a5dc-4cc2314bf097\") " pod="openstack/nova-cell1-novncproxy-0" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.865422 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c2e5d6b-1b2f-4183-b654-908203d4b1d3-config\") pod \"dnsmasq-dns-868bc9dc59-827zj\" (UID: \"5c2e5d6b-1b2f-4183-b654-908203d4b1d3\") " pod="openstack/dnsmasq-dns-868bc9dc59-827zj" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.865445 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2n26g\" (UniqueName: \"kubernetes.io/projected/5c2e5d6b-1b2f-4183-b654-908203d4b1d3-kube-api-access-2n26g\") pod \"dnsmasq-dns-868bc9dc59-827zj\" (UID: \"5c2e5d6b-1b2f-4183-b654-908203d4b1d3\") " pod="openstack/dnsmasq-dns-868bc9dc59-827zj" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.865517 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5c2e5d6b-1b2f-4183-b654-908203d4b1d3-dns-svc\") pod \"dnsmasq-dns-868bc9dc59-827zj\" (UID: \"5c2e5d6b-1b2f-4183-b654-908203d4b1d3\") " pod="openstack/dnsmasq-dns-868bc9dc59-827zj" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.865545 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8jpp\" (UniqueName: \"kubernetes.io/projected/4ea5a5ac-537e-40d2-a5dc-4cc2314bf097-kube-api-access-x8jpp\") pod \"nova-cell1-novncproxy-0\" (UID: \"4ea5a5ac-537e-40d2-a5dc-4cc2314bf097\") " pod="openstack/nova-cell1-novncproxy-0" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.865893 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5c2e5d6b-1b2f-4183-b654-908203d4b1d3-ovsdbserver-nb\") pod \"dnsmasq-dns-868bc9dc59-827zj\" (UID: \"5c2e5d6b-1b2f-4183-b654-908203d4b1d3\") " pod="openstack/dnsmasq-dns-868bc9dc59-827zj" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.866619 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5c2e5d6b-1b2f-4183-b654-908203d4b1d3-dns-swift-storage-0\") pod \"dnsmasq-dns-868bc9dc59-827zj\" (UID: \"5c2e5d6b-1b2f-4183-b654-908203d4b1d3\") " pod="openstack/dnsmasq-dns-868bc9dc59-827zj" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.867209 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c2e5d6b-1b2f-4183-b654-908203d4b1d3-config\") pod \"dnsmasq-dns-868bc9dc59-827zj\" (UID: \"5c2e5d6b-1b2f-4183-b654-908203d4b1d3\") " pod="openstack/dnsmasq-dns-868bc9dc59-827zj" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.867594 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5c2e5d6b-1b2f-4183-b654-908203d4b1d3-ovsdbserver-sb\") pod \"dnsmasq-dns-868bc9dc59-827zj\" (UID: \"5c2e5d6b-1b2f-4183-b654-908203d4b1d3\") " pod="openstack/dnsmasq-dns-868bc9dc59-827zj" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.869379 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5c2e5d6b-1b2f-4183-b654-908203d4b1d3-dns-svc\") pod \"dnsmasq-dns-868bc9dc59-827zj\" (UID: \"5c2e5d6b-1b2f-4183-b654-908203d4b1d3\") " pod="openstack/dnsmasq-dns-868bc9dc59-827zj" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.878380 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ea5a5ac-537e-40d2-a5dc-4cc2314bf097-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"4ea5a5ac-537e-40d2-a5dc-4cc2314bf097\") " pod="openstack/nova-cell1-novncproxy-0" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.882844 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2n26g\" (UniqueName: \"kubernetes.io/projected/5c2e5d6b-1b2f-4183-b654-908203d4b1d3-kube-api-access-2n26g\") pod \"dnsmasq-dns-868bc9dc59-827zj\" (UID: \"5c2e5d6b-1b2f-4183-b654-908203d4b1d3\") " pod="openstack/dnsmasq-dns-868bc9dc59-827zj" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.883778 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ea5a5ac-537e-40d2-a5dc-4cc2314bf097-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"4ea5a5ac-537e-40d2-a5dc-4cc2314bf097\") " pod="openstack/nova-cell1-novncproxy-0" Feb 18 09:19:47 crc kubenswrapper[4556]: I0218 09:19:47.892376 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8jpp\" (UniqueName: \"kubernetes.io/projected/4ea5a5ac-537e-40d2-a5dc-4cc2314bf097-kube-api-access-x8jpp\") pod \"nova-cell1-novncproxy-0\" (UID: \"4ea5a5ac-537e-40d2-a5dc-4cc2314bf097\") " pod="openstack/nova-cell1-novncproxy-0" Feb 18 09:19:48 crc kubenswrapper[4556]: I0218 09:19:48.008449 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 18 09:19:48 crc kubenswrapper[4556]: I0218 09:19:48.021637 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-868bc9dc59-827zj" Feb 18 09:19:48 crc kubenswrapper[4556]: I0218 09:19:48.126621 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-dgmmz"] Feb 18 09:19:48 crc kubenswrapper[4556]: I0218 09:19:48.299343 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 18 09:19:48 crc kubenswrapper[4556]: W0218 09:19:48.309737 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd64f15b2_c072_4bf1_b60b_3e73d6df2414.slice/crio-c29843d3a2915e6608ef40777d7dda94e47f77b27730fb500caa53d064c5a9a3 WatchSource:0}: Error finding container c29843d3a2915e6608ef40777d7dda94e47f77b27730fb500caa53d064c5a9a3: Status 404 returned error can't find the container with id c29843d3a2915e6608ef40777d7dda94e47f77b27730fb500caa53d064c5a9a3 Feb 18 09:19:48 crc kubenswrapper[4556]: I0218 09:19:48.368012 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-fj6lv"] Feb 18 09:19:48 crc kubenswrapper[4556]: I0218 09:19:48.369386 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-fj6lv" Feb 18 09:19:48 crc kubenswrapper[4556]: I0218 09:19:48.372882 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Feb 18 09:19:48 crc kubenswrapper[4556]: I0218 09:19:48.373681 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Feb 18 09:19:48 crc kubenswrapper[4556]: I0218 09:19:48.376651 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-fj6lv"] Feb 18 09:19:48 crc kubenswrapper[4556]: I0218 09:19:48.399094 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 18 09:19:48 crc kubenswrapper[4556]: I0218 09:19:48.451572 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d64f15b2-c072-4bf1-b60b-3e73d6df2414","Type":"ContainerStarted","Data":"c29843d3a2915e6608ef40777d7dda94e47f77b27730fb500caa53d064c5a9a3"} Feb 18 09:19:48 crc kubenswrapper[4556]: I0218 09:19:48.455569 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"67a3f575-2f76-4447-b777-25b0b93ad11a","Type":"ContainerStarted","Data":"bfc9b8c15721abf316f82254b9f9e41ade7dfc282a53cc807f4daa2c8ca7cae5"} Feb 18 09:19:48 crc kubenswrapper[4556]: I0218 09:19:48.463174 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-dgmmz" event={"ID":"24dad0a1-29ed-49a7-be65-d024a02fcff4","Type":"ContainerStarted","Data":"fa0e4cd24d874dcbc0d61004056eba071ea491b925058c7e338ea6c08eba387d"} Feb 18 09:19:48 crc kubenswrapper[4556]: I0218 09:19:48.463307 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-dgmmz" event={"ID":"24dad0a1-29ed-49a7-be65-d024a02fcff4","Type":"ContainerStarted","Data":"f0ab33dc9033749e3e8183dbbffb46360d83e3b0b19f98f839b19f9eb275113d"} Feb 18 09:19:48 crc kubenswrapper[4556]: I0218 09:19:48.470880 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 18 09:19:48 crc kubenswrapper[4556]: I0218 09:19:48.495845 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-dgmmz" podStartSLOduration=1.495824642 podStartE2EDuration="1.495824642s" podCreationTimestamp="2026-02-18 09:19:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:19:48.483145839 +0000 UTC m=+945.500106819" watchObservedRunningTime="2026-02-18 09:19:48.495824642 +0000 UTC m=+945.512785622" Feb 18 09:19:48 crc kubenswrapper[4556]: I0218 09:19:48.503458 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjpfq\" (UniqueName: \"kubernetes.io/projected/ef39cb1a-74bf-48cf-8166-5cf97485bd88-kube-api-access-jjpfq\") pod \"nova-cell1-conductor-db-sync-fj6lv\" (UID: \"ef39cb1a-74bf-48cf-8166-5cf97485bd88\") " pod="openstack/nova-cell1-conductor-db-sync-fj6lv" Feb 18 09:19:48 crc kubenswrapper[4556]: I0218 09:19:48.503497 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef39cb1a-74bf-48cf-8166-5cf97485bd88-scripts\") pod \"nova-cell1-conductor-db-sync-fj6lv\" (UID: \"ef39cb1a-74bf-48cf-8166-5cf97485bd88\") " pod="openstack/nova-cell1-conductor-db-sync-fj6lv" Feb 18 09:19:48 crc kubenswrapper[4556]: I0218 09:19:48.503556 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef39cb1a-74bf-48cf-8166-5cf97485bd88-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-fj6lv\" (UID: \"ef39cb1a-74bf-48cf-8166-5cf97485bd88\") " pod="openstack/nova-cell1-conductor-db-sync-fj6lv" Feb 18 09:19:48 crc kubenswrapper[4556]: I0218 09:19:48.503580 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef39cb1a-74bf-48cf-8166-5cf97485bd88-config-data\") pod \"nova-cell1-conductor-db-sync-fj6lv\" (UID: \"ef39cb1a-74bf-48cf-8166-5cf97485bd88\") " pod="openstack/nova-cell1-conductor-db-sync-fj6lv" Feb 18 09:19:48 crc kubenswrapper[4556]: I0218 09:19:48.606374 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef39cb1a-74bf-48cf-8166-5cf97485bd88-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-fj6lv\" (UID: \"ef39cb1a-74bf-48cf-8166-5cf97485bd88\") " pod="openstack/nova-cell1-conductor-db-sync-fj6lv" Feb 18 09:19:48 crc kubenswrapper[4556]: I0218 09:19:48.607402 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef39cb1a-74bf-48cf-8166-5cf97485bd88-config-data\") pod \"nova-cell1-conductor-db-sync-fj6lv\" (UID: \"ef39cb1a-74bf-48cf-8166-5cf97485bd88\") " pod="openstack/nova-cell1-conductor-db-sync-fj6lv" Feb 18 09:19:48 crc kubenswrapper[4556]: I0218 09:19:48.607960 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjpfq\" (UniqueName: \"kubernetes.io/projected/ef39cb1a-74bf-48cf-8166-5cf97485bd88-kube-api-access-jjpfq\") pod \"nova-cell1-conductor-db-sync-fj6lv\" (UID: \"ef39cb1a-74bf-48cf-8166-5cf97485bd88\") " pod="openstack/nova-cell1-conductor-db-sync-fj6lv" Feb 18 09:19:48 crc kubenswrapper[4556]: I0218 09:19:48.607994 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef39cb1a-74bf-48cf-8166-5cf97485bd88-scripts\") pod \"nova-cell1-conductor-db-sync-fj6lv\" (UID: \"ef39cb1a-74bf-48cf-8166-5cf97485bd88\") " pod="openstack/nova-cell1-conductor-db-sync-fj6lv" Feb 18 09:19:48 crc kubenswrapper[4556]: I0218 09:19:48.612073 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef39cb1a-74bf-48cf-8166-5cf97485bd88-scripts\") pod \"nova-cell1-conductor-db-sync-fj6lv\" (UID: \"ef39cb1a-74bf-48cf-8166-5cf97485bd88\") " pod="openstack/nova-cell1-conductor-db-sync-fj6lv" Feb 18 09:19:48 crc kubenswrapper[4556]: I0218 09:19:48.616669 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef39cb1a-74bf-48cf-8166-5cf97485bd88-config-data\") pod \"nova-cell1-conductor-db-sync-fj6lv\" (UID: \"ef39cb1a-74bf-48cf-8166-5cf97485bd88\") " pod="openstack/nova-cell1-conductor-db-sync-fj6lv" Feb 18 09:19:48 crc kubenswrapper[4556]: I0218 09:19:48.622119 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef39cb1a-74bf-48cf-8166-5cf97485bd88-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-fj6lv\" (UID: \"ef39cb1a-74bf-48cf-8166-5cf97485bd88\") " pod="openstack/nova-cell1-conductor-db-sync-fj6lv" Feb 18 09:19:48 crc kubenswrapper[4556]: I0218 09:19:48.628681 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjpfq\" (UniqueName: \"kubernetes.io/projected/ef39cb1a-74bf-48cf-8166-5cf97485bd88-kube-api-access-jjpfq\") pod \"nova-cell1-conductor-db-sync-fj6lv\" (UID: \"ef39cb1a-74bf-48cf-8166-5cf97485bd88\") " pod="openstack/nova-cell1-conductor-db-sync-fj6lv" Feb 18 09:19:48 crc kubenswrapper[4556]: I0218 09:19:48.659957 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-868bc9dc59-827zj"] Feb 18 09:19:48 crc kubenswrapper[4556]: I0218 09:19:48.755000 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 18 09:19:48 crc kubenswrapper[4556]: W0218 09:19:48.756045 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4ea5a5ac_537e_40d2_a5dc_4cc2314bf097.slice/crio-cb28819905c553b8f23d96fcc570350621ef2ca814d42ff1c3f0ef0768480d07 WatchSource:0}: Error finding container cb28819905c553b8f23d96fcc570350621ef2ca814d42ff1c3f0ef0768480d07: Status 404 returned error can't find the container with id cb28819905c553b8f23d96fcc570350621ef2ca814d42ff1c3f0ef0768480d07 Feb 18 09:19:48 crc kubenswrapper[4556]: I0218 09:19:48.817557 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-fj6lv" Feb 18 09:19:49 crc kubenswrapper[4556]: I0218 09:19:49.327817 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-fj6lv"] Feb 18 09:19:49 crc kubenswrapper[4556]: W0218 09:19:49.328399 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podef39cb1a_74bf_48cf_8166_5cf97485bd88.slice/crio-c0168528a4255674aa92ca6d9180d374ebc93feeb556a1627d6dee4ca3cb4269 WatchSource:0}: Error finding container c0168528a4255674aa92ca6d9180d374ebc93feeb556a1627d6dee4ca3cb4269: Status 404 returned error can't find the container with id c0168528a4255674aa92ca6d9180d374ebc93feeb556a1627d6dee4ca3cb4269 Feb 18 09:19:49 crc kubenswrapper[4556]: I0218 09:19:49.473862 4556 generic.go:334] "Generic (PLEG): container finished" podID="5c2e5d6b-1b2f-4183-b654-908203d4b1d3" containerID="bd038c2f8b47818e7f3f6eefba0f29d206ced733a70b887189020a9b0de9f33b" exitCode=0 Feb 18 09:19:49 crc kubenswrapper[4556]: I0218 09:19:49.474142 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-868bc9dc59-827zj" event={"ID":"5c2e5d6b-1b2f-4183-b654-908203d4b1d3","Type":"ContainerDied","Data":"bd038c2f8b47818e7f3f6eefba0f29d206ced733a70b887189020a9b0de9f33b"} Feb 18 09:19:49 crc kubenswrapper[4556]: I0218 09:19:49.474202 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-868bc9dc59-827zj" event={"ID":"5c2e5d6b-1b2f-4183-b654-908203d4b1d3","Type":"ContainerStarted","Data":"e8c414f28520923e6fc1777f21d7e17726a1e57d0e9c1fa2e64991e92ebe9c62"} Feb 18 09:19:49 crc kubenswrapper[4556]: I0218 09:19:49.477898 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-fj6lv" event={"ID":"ef39cb1a-74bf-48cf-8166-5cf97485bd88","Type":"ContainerStarted","Data":"c0168528a4255674aa92ca6d9180d374ebc93feeb556a1627d6dee4ca3cb4269"} Feb 18 09:19:49 crc kubenswrapper[4556]: I0218 09:19:49.481721 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2330f995-37c7-4b3e-a465-4e682349346f","Type":"ContainerStarted","Data":"408bda8acc5dc2e8cf10b40b04b6699c74680ecfb14fa057bd58ab263a48b395"} Feb 18 09:19:49 crc kubenswrapper[4556]: I0218 09:19:49.483797 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"4ea5a5ac-537e-40d2-a5dc-4cc2314bf097","Type":"ContainerStarted","Data":"cb28819905c553b8f23d96fcc570350621ef2ca814d42ff1c3f0ef0768480d07"} Feb 18 09:19:50 crc kubenswrapper[4556]: I0218 09:19:50.515242 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-868bc9dc59-827zj" event={"ID":"5c2e5d6b-1b2f-4183-b654-908203d4b1d3","Type":"ContainerStarted","Data":"557d6c6a6dd07fd9b0ef587eb11d08bfd8f3521c738d7bf4463bac456d3d22b0"} Feb 18 09:19:50 crc kubenswrapper[4556]: I0218 09:19:50.516685 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-868bc9dc59-827zj" Feb 18 09:19:50 crc kubenswrapper[4556]: I0218 09:19:50.523767 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-fj6lv" event={"ID":"ef39cb1a-74bf-48cf-8166-5cf97485bd88","Type":"ContainerStarted","Data":"eed458cdccc5b601f881f1ea893deb67f0c201b5b99dc68921127d9a44b49b65"} Feb 18 09:19:50 crc kubenswrapper[4556]: I0218 09:19:50.551409 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-868bc9dc59-827zj" podStartSLOduration=3.5513839750000002 podStartE2EDuration="3.551383975s" podCreationTimestamp="2026-02-18 09:19:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:19:50.530818392 +0000 UTC m=+947.547779373" watchObservedRunningTime="2026-02-18 09:19:50.551383975 +0000 UTC m=+947.568344956" Feb 18 09:19:50 crc kubenswrapper[4556]: I0218 09:19:50.573895 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-fj6lv" podStartSLOduration=2.57386829 podStartE2EDuration="2.57386829s" podCreationTimestamp="2026-02-18 09:19:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:19:50.553780136 +0000 UTC m=+947.570741126" watchObservedRunningTime="2026-02-18 09:19:50.57386829 +0000 UTC m=+947.590829270" Feb 18 09:19:50 crc kubenswrapper[4556]: I0218 09:19:50.761673 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jsb7c" Feb 18 09:19:50 crc kubenswrapper[4556]: I0218 09:19:50.761865 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jsb7c" Feb 18 09:19:50 crc kubenswrapper[4556]: I0218 09:19:50.809640 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jsb7c" Feb 18 09:19:51 crc kubenswrapper[4556]: I0218 09:19:51.163763 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 18 09:19:51 crc kubenswrapper[4556]: I0218 09:19:51.170638 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 18 09:19:51 crc kubenswrapper[4556]: I0218 09:19:51.599046 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jsb7c" Feb 18 09:19:52 crc kubenswrapper[4556]: I0218 09:19:52.045585 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jsb7c"] Feb 18 09:19:52 crc kubenswrapper[4556]: I0218 09:19:52.559192 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"67a3f575-2f76-4447-b777-25b0b93ad11a","Type":"ContainerStarted","Data":"136ed6176137995b75948550d7dc95861646c7cc08f0e807d940a4c6a25153c7"} Feb 18 09:19:52 crc kubenswrapper[4556]: I0218 09:19:52.561695 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"4ea5a5ac-537e-40d2-a5dc-4cc2314bf097","Type":"ContainerStarted","Data":"5b0edfc3e404d52fc9c324525b661c589d816bf646b1ce5470138f85ad4b763a"} Feb 18 09:19:52 crc kubenswrapper[4556]: I0218 09:19:52.561773 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="4ea5a5ac-537e-40d2-a5dc-4cc2314bf097" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://5b0edfc3e404d52fc9c324525b661c589d816bf646b1ce5470138f85ad4b763a" gracePeriod=30 Feb 18 09:19:52 crc kubenswrapper[4556]: I0218 09:19:52.565291 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2330f995-37c7-4b3e-a465-4e682349346f","Type":"ContainerStarted","Data":"f6a2777c15f011057e9eb148ba4d170bda4c52701b24e2f8ccd0d1db588afa4f"} Feb 18 09:19:52 crc kubenswrapper[4556]: I0218 09:19:52.565433 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2330f995-37c7-4b3e-a465-4e682349346f","Type":"ContainerStarted","Data":"1d20caf4c1ff142373c33a04b0efed7d80eb00bca35ccaa547d8f5984917dcef"} Feb 18 09:19:52 crc kubenswrapper[4556]: I0218 09:19:52.565697 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="2330f995-37c7-4b3e-a465-4e682349346f" containerName="nova-metadata-log" containerID="cri-o://1d20caf4c1ff142373c33a04b0efed7d80eb00bca35ccaa547d8f5984917dcef" gracePeriod=30 Feb 18 09:19:52 crc kubenswrapper[4556]: I0218 09:19:52.565925 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="2330f995-37c7-4b3e-a465-4e682349346f" containerName="nova-metadata-metadata" containerID="cri-o://f6a2777c15f011057e9eb148ba4d170bda4c52701b24e2f8ccd0d1db588afa4f" gracePeriod=30 Feb 18 09:19:52 crc kubenswrapper[4556]: I0218 09:19:52.570308 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d64f15b2-c072-4bf1-b60b-3e73d6df2414","Type":"ContainerStarted","Data":"c26e129cdfd10ac7f662bce7f0cd0c203089095198aae4710eadd93dc6092a61"} Feb 18 09:19:52 crc kubenswrapper[4556]: I0218 09:19:52.570366 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d64f15b2-c072-4bf1-b60b-3e73d6df2414","Type":"ContainerStarted","Data":"02727f04dcdb857126a77711e83fdfc2bf3d35b672c828dcea31d7eb9ab89b9e"} Feb 18 09:19:52 crc kubenswrapper[4556]: I0218 09:19:52.589549 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.035184523 podStartE2EDuration="5.589530153s" podCreationTimestamp="2026-02-18 09:19:47 +0000 UTC" firstStartedPulling="2026-02-18 09:19:48.40849672 +0000 UTC m=+945.425457699" lastFinishedPulling="2026-02-18 09:19:51.962842349 +0000 UTC m=+948.979803329" observedRunningTime="2026-02-18 09:19:52.58856938 +0000 UTC m=+949.605530361" watchObservedRunningTime="2026-02-18 09:19:52.589530153 +0000 UTC m=+949.606491134" Feb 18 09:19:52 crc kubenswrapper[4556]: I0218 09:19:52.616678 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=1.972276587 podStartE2EDuration="5.616665503s" podCreationTimestamp="2026-02-18 09:19:47 +0000 UTC" firstStartedPulling="2026-02-18 09:19:48.326176365 +0000 UTC m=+945.343137345" lastFinishedPulling="2026-02-18 09:19:51.97056528 +0000 UTC m=+948.987526261" observedRunningTime="2026-02-18 09:19:52.613634123 +0000 UTC m=+949.630595103" watchObservedRunningTime="2026-02-18 09:19:52.616665503 +0000 UTC m=+949.633626482" Feb 18 09:19:52 crc kubenswrapper[4556]: I0218 09:19:52.638879 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.135444597 podStartE2EDuration="5.638856293s" podCreationTimestamp="2026-02-18 09:19:47 +0000 UTC" firstStartedPulling="2026-02-18 09:19:48.470728581 +0000 UTC m=+945.487689560" lastFinishedPulling="2026-02-18 09:19:51.974140276 +0000 UTC m=+948.991101256" observedRunningTime="2026-02-18 09:19:52.636428021 +0000 UTC m=+949.653389002" watchObservedRunningTime="2026-02-18 09:19:52.638856293 +0000 UTC m=+949.655817273" Feb 18 09:19:52 crc kubenswrapper[4556]: I0218 09:19:52.654982 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.4488533869999998 podStartE2EDuration="5.654966951s" podCreationTimestamp="2026-02-18 09:19:47 +0000 UTC" firstStartedPulling="2026-02-18 09:19:48.764533049 +0000 UTC m=+945.781494030" lastFinishedPulling="2026-02-18 09:19:51.970646614 +0000 UTC m=+948.987607594" observedRunningTime="2026-02-18 09:19:52.650414481 +0000 UTC m=+949.667375461" watchObservedRunningTime="2026-02-18 09:19:52.654966951 +0000 UTC m=+949.671927931" Feb 18 09:19:52 crc kubenswrapper[4556]: I0218 09:19:52.795748 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Feb 18 09:19:52 crc kubenswrapper[4556]: I0218 09:19:52.815602 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 18 09:19:52 crc kubenswrapper[4556]: I0218 09:19:52.815643 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 18 09:19:53 crc kubenswrapper[4556]: I0218 09:19:53.009737 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Feb 18 09:19:53 crc kubenswrapper[4556]: I0218 09:19:53.580860 4556 generic.go:334] "Generic (PLEG): container finished" podID="2330f995-37c7-4b3e-a465-4e682349346f" containerID="1d20caf4c1ff142373c33a04b0efed7d80eb00bca35ccaa547d8f5984917dcef" exitCode=143 Feb 18 09:19:53 crc kubenswrapper[4556]: I0218 09:19:53.582188 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2330f995-37c7-4b3e-a465-4e682349346f","Type":"ContainerDied","Data":"1d20caf4c1ff142373c33a04b0efed7d80eb00bca35ccaa547d8f5984917dcef"} Feb 18 09:19:53 crc kubenswrapper[4556]: I0218 09:19:53.582397 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jsb7c" podUID="6b22db00-3e66-4afe-bce4-20f733982dc8" containerName="registry-server" containerID="cri-o://307c12968fecf2b16c9f86102e3f089cb02c58d88fe6fb6914c31cecd27abf6d" gracePeriod=2 Feb 18 09:19:54 crc kubenswrapper[4556]: I0218 09:19:54.070458 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jsb7c" Feb 18 09:19:54 crc kubenswrapper[4556]: I0218 09:19:54.092305 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b22db00-3e66-4afe-bce4-20f733982dc8-utilities\") pod \"6b22db00-3e66-4afe-bce4-20f733982dc8\" (UID: \"6b22db00-3e66-4afe-bce4-20f733982dc8\") " Feb 18 09:19:54 crc kubenswrapper[4556]: I0218 09:19:54.092456 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6k8dh\" (UniqueName: \"kubernetes.io/projected/6b22db00-3e66-4afe-bce4-20f733982dc8-kube-api-access-6k8dh\") pod \"6b22db00-3e66-4afe-bce4-20f733982dc8\" (UID: \"6b22db00-3e66-4afe-bce4-20f733982dc8\") " Feb 18 09:19:54 crc kubenswrapper[4556]: I0218 09:19:54.092809 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b22db00-3e66-4afe-bce4-20f733982dc8-utilities" (OuterVolumeSpecName: "utilities") pod "6b22db00-3e66-4afe-bce4-20f733982dc8" (UID: "6b22db00-3e66-4afe-bce4-20f733982dc8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:19:54 crc kubenswrapper[4556]: I0218 09:19:54.093185 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b22db00-3e66-4afe-bce4-20f733982dc8-catalog-content\") pod \"6b22db00-3e66-4afe-bce4-20f733982dc8\" (UID: \"6b22db00-3e66-4afe-bce4-20f733982dc8\") " Feb 18 09:19:54 crc kubenswrapper[4556]: I0218 09:19:54.093893 4556 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b22db00-3e66-4afe-bce4-20f733982dc8-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:54 crc kubenswrapper[4556]: I0218 09:19:54.109373 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b22db00-3e66-4afe-bce4-20f733982dc8-kube-api-access-6k8dh" (OuterVolumeSpecName: "kube-api-access-6k8dh") pod "6b22db00-3e66-4afe-bce4-20f733982dc8" (UID: "6b22db00-3e66-4afe-bce4-20f733982dc8"). InnerVolumeSpecName "kube-api-access-6k8dh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:19:54 crc kubenswrapper[4556]: I0218 09:19:54.134003 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b22db00-3e66-4afe-bce4-20f733982dc8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6b22db00-3e66-4afe-bce4-20f733982dc8" (UID: "6b22db00-3e66-4afe-bce4-20f733982dc8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:19:54 crc kubenswrapper[4556]: I0218 09:19:54.196319 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6k8dh\" (UniqueName: \"kubernetes.io/projected/6b22db00-3e66-4afe-bce4-20f733982dc8-kube-api-access-6k8dh\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:54 crc kubenswrapper[4556]: I0218 09:19:54.196359 4556 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b22db00-3e66-4afe-bce4-20f733982dc8-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:54 crc kubenswrapper[4556]: I0218 09:19:54.592211 4556 generic.go:334] "Generic (PLEG): container finished" podID="ef39cb1a-74bf-48cf-8166-5cf97485bd88" containerID="eed458cdccc5b601f881f1ea893deb67f0c201b5b99dc68921127d9a44b49b65" exitCode=0 Feb 18 09:19:54 crc kubenswrapper[4556]: I0218 09:19:54.592309 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-fj6lv" event={"ID":"ef39cb1a-74bf-48cf-8166-5cf97485bd88","Type":"ContainerDied","Data":"eed458cdccc5b601f881f1ea893deb67f0c201b5b99dc68921127d9a44b49b65"} Feb 18 09:19:54 crc kubenswrapper[4556]: I0218 09:19:54.596455 4556 generic.go:334] "Generic (PLEG): container finished" podID="6b22db00-3e66-4afe-bce4-20f733982dc8" containerID="307c12968fecf2b16c9f86102e3f089cb02c58d88fe6fb6914c31cecd27abf6d" exitCode=0 Feb 18 09:19:54 crc kubenswrapper[4556]: I0218 09:19:54.596509 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jsb7c" event={"ID":"6b22db00-3e66-4afe-bce4-20f733982dc8","Type":"ContainerDied","Data":"307c12968fecf2b16c9f86102e3f089cb02c58d88fe6fb6914c31cecd27abf6d"} Feb 18 09:19:54 crc kubenswrapper[4556]: I0218 09:19:54.596542 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jsb7c" event={"ID":"6b22db00-3e66-4afe-bce4-20f733982dc8","Type":"ContainerDied","Data":"afad6520f76efafd7eb26488ce9c0436cad42933317dc32cbd9f1f02e40f86ae"} Feb 18 09:19:54 crc kubenswrapper[4556]: I0218 09:19:54.596562 4556 scope.go:117] "RemoveContainer" containerID="307c12968fecf2b16c9f86102e3f089cb02c58d88fe6fb6914c31cecd27abf6d" Feb 18 09:19:54 crc kubenswrapper[4556]: I0218 09:19:54.596491 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jsb7c" Feb 18 09:19:54 crc kubenswrapper[4556]: I0218 09:19:54.617567 4556 scope.go:117] "RemoveContainer" containerID="f3a70532a2f42c1494230cf1962658d0f84a2066e8f5712ccb3eb037e0494ada" Feb 18 09:19:54 crc kubenswrapper[4556]: I0218 09:19:54.639450 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jsb7c"] Feb 18 09:19:54 crc kubenswrapper[4556]: I0218 09:19:54.647873 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jsb7c"] Feb 18 09:19:54 crc kubenswrapper[4556]: I0218 09:19:54.651756 4556 scope.go:117] "RemoveContainer" containerID="a413f74d1082552e6ab9b8a637cc714fc127f96c55cc673c48ed0415267d8639" Feb 18 09:19:54 crc kubenswrapper[4556]: I0218 09:19:54.672760 4556 scope.go:117] "RemoveContainer" containerID="307c12968fecf2b16c9f86102e3f089cb02c58d88fe6fb6914c31cecd27abf6d" Feb 18 09:19:54 crc kubenswrapper[4556]: E0218 09:19:54.673211 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"307c12968fecf2b16c9f86102e3f089cb02c58d88fe6fb6914c31cecd27abf6d\": container with ID starting with 307c12968fecf2b16c9f86102e3f089cb02c58d88fe6fb6914c31cecd27abf6d not found: ID does not exist" containerID="307c12968fecf2b16c9f86102e3f089cb02c58d88fe6fb6914c31cecd27abf6d" Feb 18 09:19:54 crc kubenswrapper[4556]: I0218 09:19:54.673246 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"307c12968fecf2b16c9f86102e3f089cb02c58d88fe6fb6914c31cecd27abf6d"} err="failed to get container status \"307c12968fecf2b16c9f86102e3f089cb02c58d88fe6fb6914c31cecd27abf6d\": rpc error: code = NotFound desc = could not find container \"307c12968fecf2b16c9f86102e3f089cb02c58d88fe6fb6914c31cecd27abf6d\": container with ID starting with 307c12968fecf2b16c9f86102e3f089cb02c58d88fe6fb6914c31cecd27abf6d not found: ID does not exist" Feb 18 09:19:54 crc kubenswrapper[4556]: I0218 09:19:54.673272 4556 scope.go:117] "RemoveContainer" containerID="f3a70532a2f42c1494230cf1962658d0f84a2066e8f5712ccb3eb037e0494ada" Feb 18 09:19:54 crc kubenswrapper[4556]: E0218 09:19:54.673712 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3a70532a2f42c1494230cf1962658d0f84a2066e8f5712ccb3eb037e0494ada\": container with ID starting with f3a70532a2f42c1494230cf1962658d0f84a2066e8f5712ccb3eb037e0494ada not found: ID does not exist" containerID="f3a70532a2f42c1494230cf1962658d0f84a2066e8f5712ccb3eb037e0494ada" Feb 18 09:19:54 crc kubenswrapper[4556]: I0218 09:19:54.673816 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3a70532a2f42c1494230cf1962658d0f84a2066e8f5712ccb3eb037e0494ada"} err="failed to get container status \"f3a70532a2f42c1494230cf1962658d0f84a2066e8f5712ccb3eb037e0494ada\": rpc error: code = NotFound desc = could not find container \"f3a70532a2f42c1494230cf1962658d0f84a2066e8f5712ccb3eb037e0494ada\": container with ID starting with f3a70532a2f42c1494230cf1962658d0f84a2066e8f5712ccb3eb037e0494ada not found: ID does not exist" Feb 18 09:19:54 crc kubenswrapper[4556]: I0218 09:19:54.673844 4556 scope.go:117] "RemoveContainer" containerID="a413f74d1082552e6ab9b8a637cc714fc127f96c55cc673c48ed0415267d8639" Feb 18 09:19:54 crc kubenswrapper[4556]: E0218 09:19:54.674126 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a413f74d1082552e6ab9b8a637cc714fc127f96c55cc673c48ed0415267d8639\": container with ID starting with a413f74d1082552e6ab9b8a637cc714fc127f96c55cc673c48ed0415267d8639 not found: ID does not exist" containerID="a413f74d1082552e6ab9b8a637cc714fc127f96c55cc673c48ed0415267d8639" Feb 18 09:19:54 crc kubenswrapper[4556]: I0218 09:19:54.674233 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a413f74d1082552e6ab9b8a637cc714fc127f96c55cc673c48ed0415267d8639"} err="failed to get container status \"a413f74d1082552e6ab9b8a637cc714fc127f96c55cc673c48ed0415267d8639\": rpc error: code = NotFound desc = could not find container \"a413f74d1082552e6ab9b8a637cc714fc127f96c55cc673c48ed0415267d8639\": container with ID starting with a413f74d1082552e6ab9b8a637cc714fc127f96c55cc673c48ed0415267d8639 not found: ID does not exist" Feb 18 09:19:55 crc kubenswrapper[4556]: I0218 09:19:55.293130 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b22db00-3e66-4afe-bce4-20f733982dc8" path="/var/lib/kubelet/pods/6b22db00-3e66-4afe-bce4-20f733982dc8/volumes" Feb 18 09:19:55 crc kubenswrapper[4556]: I0218 09:19:55.606593 4556 generic.go:334] "Generic (PLEG): container finished" podID="24dad0a1-29ed-49a7-be65-d024a02fcff4" containerID="fa0e4cd24d874dcbc0d61004056eba071ea491b925058c7e338ea6c08eba387d" exitCode=0 Feb 18 09:19:55 crc kubenswrapper[4556]: I0218 09:19:55.606683 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-dgmmz" event={"ID":"24dad0a1-29ed-49a7-be65-d024a02fcff4","Type":"ContainerDied","Data":"fa0e4cd24d874dcbc0d61004056eba071ea491b925058c7e338ea6c08eba387d"} Feb 18 09:19:55 crc kubenswrapper[4556]: I0218 09:19:55.944958 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-fj6lv" Feb 18 09:19:56 crc kubenswrapper[4556]: I0218 09:19:56.031144 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef39cb1a-74bf-48cf-8166-5cf97485bd88-combined-ca-bundle\") pod \"ef39cb1a-74bf-48cf-8166-5cf97485bd88\" (UID: \"ef39cb1a-74bf-48cf-8166-5cf97485bd88\") " Feb 18 09:19:56 crc kubenswrapper[4556]: I0218 09:19:56.031220 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jjpfq\" (UniqueName: \"kubernetes.io/projected/ef39cb1a-74bf-48cf-8166-5cf97485bd88-kube-api-access-jjpfq\") pod \"ef39cb1a-74bf-48cf-8166-5cf97485bd88\" (UID: \"ef39cb1a-74bf-48cf-8166-5cf97485bd88\") " Feb 18 09:19:56 crc kubenswrapper[4556]: I0218 09:19:56.031354 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef39cb1a-74bf-48cf-8166-5cf97485bd88-scripts\") pod \"ef39cb1a-74bf-48cf-8166-5cf97485bd88\" (UID: \"ef39cb1a-74bf-48cf-8166-5cf97485bd88\") " Feb 18 09:19:56 crc kubenswrapper[4556]: I0218 09:19:56.031420 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef39cb1a-74bf-48cf-8166-5cf97485bd88-config-data\") pod \"ef39cb1a-74bf-48cf-8166-5cf97485bd88\" (UID: \"ef39cb1a-74bf-48cf-8166-5cf97485bd88\") " Feb 18 09:19:56 crc kubenswrapper[4556]: I0218 09:19:56.042304 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef39cb1a-74bf-48cf-8166-5cf97485bd88-scripts" (OuterVolumeSpecName: "scripts") pod "ef39cb1a-74bf-48cf-8166-5cf97485bd88" (UID: "ef39cb1a-74bf-48cf-8166-5cf97485bd88"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:19:56 crc kubenswrapper[4556]: I0218 09:19:56.065270 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef39cb1a-74bf-48cf-8166-5cf97485bd88-kube-api-access-jjpfq" (OuterVolumeSpecName: "kube-api-access-jjpfq") pod "ef39cb1a-74bf-48cf-8166-5cf97485bd88" (UID: "ef39cb1a-74bf-48cf-8166-5cf97485bd88"). InnerVolumeSpecName "kube-api-access-jjpfq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:19:56 crc kubenswrapper[4556]: I0218 09:19:56.085304 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef39cb1a-74bf-48cf-8166-5cf97485bd88-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ef39cb1a-74bf-48cf-8166-5cf97485bd88" (UID: "ef39cb1a-74bf-48cf-8166-5cf97485bd88"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:19:56 crc kubenswrapper[4556]: I0218 09:19:56.093231 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef39cb1a-74bf-48cf-8166-5cf97485bd88-config-data" (OuterVolumeSpecName: "config-data") pod "ef39cb1a-74bf-48cf-8166-5cf97485bd88" (UID: "ef39cb1a-74bf-48cf-8166-5cf97485bd88"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:19:56 crc kubenswrapper[4556]: I0218 09:19:56.136760 4556 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef39cb1a-74bf-48cf-8166-5cf97485bd88-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:56 crc kubenswrapper[4556]: I0218 09:19:56.136793 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jjpfq\" (UniqueName: \"kubernetes.io/projected/ef39cb1a-74bf-48cf-8166-5cf97485bd88-kube-api-access-jjpfq\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:56 crc kubenswrapper[4556]: I0218 09:19:56.136806 4556 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef39cb1a-74bf-48cf-8166-5cf97485bd88-scripts\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:56 crc kubenswrapper[4556]: I0218 09:19:56.136820 4556 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef39cb1a-74bf-48cf-8166-5cf97485bd88-config-data\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:56 crc kubenswrapper[4556]: I0218 09:19:56.634413 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-fj6lv" event={"ID":"ef39cb1a-74bf-48cf-8166-5cf97485bd88","Type":"ContainerDied","Data":"c0168528a4255674aa92ca6d9180d374ebc93feeb556a1627d6dee4ca3cb4269"} Feb 18 09:19:56 crc kubenswrapper[4556]: I0218 09:19:56.634446 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-fj6lv" Feb 18 09:19:56 crc kubenswrapper[4556]: I0218 09:19:56.634470 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c0168528a4255674aa92ca6d9180d374ebc93feeb556a1627d6dee4ca3cb4269" Feb 18 09:19:56 crc kubenswrapper[4556]: I0218 09:19:56.687460 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 18 09:19:56 crc kubenswrapper[4556]: E0218 09:19:56.702068 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b22db00-3e66-4afe-bce4-20f733982dc8" containerName="extract-content" Feb 18 09:19:56 crc kubenswrapper[4556]: I0218 09:19:56.702102 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b22db00-3e66-4afe-bce4-20f733982dc8" containerName="extract-content" Feb 18 09:19:56 crc kubenswrapper[4556]: E0218 09:19:56.702123 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef39cb1a-74bf-48cf-8166-5cf97485bd88" containerName="nova-cell1-conductor-db-sync" Feb 18 09:19:56 crc kubenswrapper[4556]: I0218 09:19:56.702130 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef39cb1a-74bf-48cf-8166-5cf97485bd88" containerName="nova-cell1-conductor-db-sync" Feb 18 09:19:56 crc kubenswrapper[4556]: E0218 09:19:56.702172 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b22db00-3e66-4afe-bce4-20f733982dc8" containerName="extract-utilities" Feb 18 09:19:56 crc kubenswrapper[4556]: I0218 09:19:56.702181 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b22db00-3e66-4afe-bce4-20f733982dc8" containerName="extract-utilities" Feb 18 09:19:56 crc kubenswrapper[4556]: E0218 09:19:56.702190 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b22db00-3e66-4afe-bce4-20f733982dc8" containerName="registry-server" Feb 18 09:19:56 crc kubenswrapper[4556]: I0218 09:19:56.702197 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b22db00-3e66-4afe-bce4-20f733982dc8" containerName="registry-server" Feb 18 09:19:56 crc kubenswrapper[4556]: I0218 09:19:56.702420 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b22db00-3e66-4afe-bce4-20f733982dc8" containerName="registry-server" Feb 18 09:19:56 crc kubenswrapper[4556]: I0218 09:19:56.702442 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef39cb1a-74bf-48cf-8166-5cf97485bd88" containerName="nova-cell1-conductor-db-sync" Feb 18 09:19:56 crc kubenswrapper[4556]: I0218 09:19:56.703099 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 18 09:19:56 crc kubenswrapper[4556]: I0218 09:19:56.703209 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Feb 18 09:19:56 crc kubenswrapper[4556]: I0218 09:19:56.705580 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Feb 18 09:19:56 crc kubenswrapper[4556]: I0218 09:19:56.755065 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0149238-212d-489d-8c94-7b0a8d86b1f5-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"f0149238-212d-489d-8c94-7b0a8d86b1f5\") " pod="openstack/nova-cell1-conductor-0" Feb 18 09:19:56 crc kubenswrapper[4556]: I0218 09:19:56.755275 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0149238-212d-489d-8c94-7b0a8d86b1f5-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"f0149238-212d-489d-8c94-7b0a8d86b1f5\") " pod="openstack/nova-cell1-conductor-0" Feb 18 09:19:56 crc kubenswrapper[4556]: I0218 09:19:56.755435 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfhss\" (UniqueName: \"kubernetes.io/projected/f0149238-212d-489d-8c94-7b0a8d86b1f5-kube-api-access-mfhss\") pod \"nova-cell1-conductor-0\" (UID: \"f0149238-212d-489d-8c94-7b0a8d86b1f5\") " pod="openstack/nova-cell1-conductor-0" Feb 18 09:19:56 crc kubenswrapper[4556]: I0218 09:19:56.857129 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0149238-212d-489d-8c94-7b0a8d86b1f5-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"f0149238-212d-489d-8c94-7b0a8d86b1f5\") " pod="openstack/nova-cell1-conductor-0" Feb 18 09:19:56 crc kubenswrapper[4556]: I0218 09:19:56.857233 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0149238-212d-489d-8c94-7b0a8d86b1f5-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"f0149238-212d-489d-8c94-7b0a8d86b1f5\") " pod="openstack/nova-cell1-conductor-0" Feb 18 09:19:56 crc kubenswrapper[4556]: I0218 09:19:56.857283 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfhss\" (UniqueName: \"kubernetes.io/projected/f0149238-212d-489d-8c94-7b0a8d86b1f5-kube-api-access-mfhss\") pod \"nova-cell1-conductor-0\" (UID: \"f0149238-212d-489d-8c94-7b0a8d86b1f5\") " pod="openstack/nova-cell1-conductor-0" Feb 18 09:19:56 crc kubenswrapper[4556]: I0218 09:19:56.863384 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0149238-212d-489d-8c94-7b0a8d86b1f5-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"f0149238-212d-489d-8c94-7b0a8d86b1f5\") " pod="openstack/nova-cell1-conductor-0" Feb 18 09:19:56 crc kubenswrapper[4556]: I0218 09:19:56.864567 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0149238-212d-489d-8c94-7b0a8d86b1f5-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"f0149238-212d-489d-8c94-7b0a8d86b1f5\") " pod="openstack/nova-cell1-conductor-0" Feb 18 09:19:56 crc kubenswrapper[4556]: I0218 09:19:56.875620 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfhss\" (UniqueName: \"kubernetes.io/projected/f0149238-212d-489d-8c94-7b0a8d86b1f5-kube-api-access-mfhss\") pod \"nova-cell1-conductor-0\" (UID: \"f0149238-212d-489d-8c94-7b0a8d86b1f5\") " pod="openstack/nova-cell1-conductor-0" Feb 18 09:19:56 crc kubenswrapper[4556]: I0218 09:19:56.981268 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-dgmmz" Feb 18 09:19:57 crc kubenswrapper[4556]: I0218 09:19:57.023524 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Feb 18 09:19:57 crc kubenswrapper[4556]: I0218 09:19:57.063739 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-57bln\" (UniqueName: \"kubernetes.io/projected/24dad0a1-29ed-49a7-be65-d024a02fcff4-kube-api-access-57bln\") pod \"24dad0a1-29ed-49a7-be65-d024a02fcff4\" (UID: \"24dad0a1-29ed-49a7-be65-d024a02fcff4\") " Feb 18 09:19:57 crc kubenswrapper[4556]: I0218 09:19:57.067410 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24dad0a1-29ed-49a7-be65-d024a02fcff4-kube-api-access-57bln" (OuterVolumeSpecName: "kube-api-access-57bln") pod "24dad0a1-29ed-49a7-be65-d024a02fcff4" (UID: "24dad0a1-29ed-49a7-be65-d024a02fcff4"). InnerVolumeSpecName "kube-api-access-57bln". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:19:57 crc kubenswrapper[4556]: I0218 09:19:57.165648 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24dad0a1-29ed-49a7-be65-d024a02fcff4-combined-ca-bundle\") pod \"24dad0a1-29ed-49a7-be65-d024a02fcff4\" (UID: \"24dad0a1-29ed-49a7-be65-d024a02fcff4\") " Feb 18 09:19:57 crc kubenswrapper[4556]: I0218 09:19:57.165786 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24dad0a1-29ed-49a7-be65-d024a02fcff4-config-data\") pod \"24dad0a1-29ed-49a7-be65-d024a02fcff4\" (UID: \"24dad0a1-29ed-49a7-be65-d024a02fcff4\") " Feb 18 09:19:57 crc kubenswrapper[4556]: I0218 09:19:57.165941 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24dad0a1-29ed-49a7-be65-d024a02fcff4-scripts\") pod \"24dad0a1-29ed-49a7-be65-d024a02fcff4\" (UID: \"24dad0a1-29ed-49a7-be65-d024a02fcff4\") " Feb 18 09:19:57 crc kubenswrapper[4556]: I0218 09:19:57.166946 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-57bln\" (UniqueName: \"kubernetes.io/projected/24dad0a1-29ed-49a7-be65-d024a02fcff4-kube-api-access-57bln\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:57 crc kubenswrapper[4556]: I0218 09:19:57.169542 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24dad0a1-29ed-49a7-be65-d024a02fcff4-scripts" (OuterVolumeSpecName: "scripts") pod "24dad0a1-29ed-49a7-be65-d024a02fcff4" (UID: "24dad0a1-29ed-49a7-be65-d024a02fcff4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:19:57 crc kubenswrapper[4556]: I0218 09:19:57.194323 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24dad0a1-29ed-49a7-be65-d024a02fcff4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "24dad0a1-29ed-49a7-be65-d024a02fcff4" (UID: "24dad0a1-29ed-49a7-be65-d024a02fcff4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:19:57 crc kubenswrapper[4556]: I0218 09:19:57.198294 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24dad0a1-29ed-49a7-be65-d024a02fcff4-config-data" (OuterVolumeSpecName: "config-data") pod "24dad0a1-29ed-49a7-be65-d024a02fcff4" (UID: "24dad0a1-29ed-49a7-be65-d024a02fcff4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:19:57 crc kubenswrapper[4556]: I0218 09:19:57.269117 4556 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24dad0a1-29ed-49a7-be65-d024a02fcff4-config-data\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:57 crc kubenswrapper[4556]: I0218 09:19:57.269171 4556 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24dad0a1-29ed-49a7-be65-d024a02fcff4-scripts\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:57 crc kubenswrapper[4556]: I0218 09:19:57.269184 4556 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24dad0a1-29ed-49a7-be65-d024a02fcff4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:57 crc kubenswrapper[4556]: I0218 09:19:57.437264 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 18 09:19:57 crc kubenswrapper[4556]: W0218 09:19:57.446971 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf0149238_212d_489d_8c94_7b0a8d86b1f5.slice/crio-cf7bd478f8a395355406e86ed2be3c3abfddc08187c814cecee488a659544819 WatchSource:0}: Error finding container cf7bd478f8a395355406e86ed2be3c3abfddc08187c814cecee488a659544819: Status 404 returned error can't find the container with id cf7bd478f8a395355406e86ed2be3c3abfddc08187c814cecee488a659544819 Feb 18 09:19:57 crc kubenswrapper[4556]: I0218 09:19:57.645043 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-dgmmz" event={"ID":"24dad0a1-29ed-49a7-be65-d024a02fcff4","Type":"ContainerDied","Data":"f0ab33dc9033749e3e8183dbbffb46360d83e3b0b19f98f839b19f9eb275113d"} Feb 18 09:19:57 crc kubenswrapper[4556]: I0218 09:19:57.645087 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f0ab33dc9033749e3e8183dbbffb46360d83e3b0b19f98f839b19f9eb275113d" Feb 18 09:19:57 crc kubenswrapper[4556]: I0218 09:19:57.645125 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-dgmmz" Feb 18 09:19:57 crc kubenswrapper[4556]: I0218 09:19:57.647903 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"f0149238-212d-489d-8c94-7b0a8d86b1f5","Type":"ContainerStarted","Data":"0174623ab95389c4a83f976e1866768ecec252628ad81b5b8c48f7d633f84472"} Feb 18 09:19:57 crc kubenswrapper[4556]: I0218 09:19:57.647944 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"f0149238-212d-489d-8c94-7b0a8d86b1f5","Type":"ContainerStarted","Data":"cf7bd478f8a395355406e86ed2be3c3abfddc08187c814cecee488a659544819"} Feb 18 09:19:57 crc kubenswrapper[4556]: I0218 09:19:57.648000 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Feb 18 09:19:57 crc kubenswrapper[4556]: I0218 09:19:57.667361 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=1.6673440400000001 podStartE2EDuration="1.66734404s" podCreationTimestamp="2026-02-18 09:19:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:19:57.66235647 +0000 UTC m=+954.679317449" watchObservedRunningTime="2026-02-18 09:19:57.66734404 +0000 UTC m=+954.684305019" Feb 18 09:19:57 crc kubenswrapper[4556]: I0218 09:19:57.757074 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 18 09:19:57 crc kubenswrapper[4556]: I0218 09:19:57.757218 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 18 09:19:57 crc kubenswrapper[4556]: I0218 09:19:57.795890 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Feb 18 09:19:57 crc kubenswrapper[4556]: I0218 09:19:57.799468 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 18 09:19:57 crc kubenswrapper[4556]: I0218 09:19:57.810128 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 18 09:19:57 crc kubenswrapper[4556]: I0218 09:19:57.831731 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Feb 18 09:19:58 crc kubenswrapper[4556]: I0218 09:19:58.023327 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-868bc9dc59-827zj" Feb 18 09:19:58 crc kubenswrapper[4556]: I0218 09:19:58.088654 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-849fd69845-bdbc8"] Feb 18 09:19:58 crc kubenswrapper[4556]: I0218 09:19:58.088875 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-849fd69845-bdbc8" podUID="129baf51-56c6-488a-88e2-90afcbb3d880" containerName="dnsmasq-dns" containerID="cri-o://6cbad85d5780b1b4e977194f87ef2f9451600f0524aecf9002beb6201ffcc69f" gracePeriod=10 Feb 18 09:19:58 crc kubenswrapper[4556]: I0218 09:19:58.570003 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-849fd69845-bdbc8" Feb 18 09:19:58 crc kubenswrapper[4556]: I0218 09:19:58.613437 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/129baf51-56c6-488a-88e2-90afcbb3d880-dns-swift-storage-0\") pod \"129baf51-56c6-488a-88e2-90afcbb3d880\" (UID: \"129baf51-56c6-488a-88e2-90afcbb3d880\") " Feb 18 09:19:58 crc kubenswrapper[4556]: I0218 09:19:58.613531 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/129baf51-56c6-488a-88e2-90afcbb3d880-ovsdbserver-nb\") pod \"129baf51-56c6-488a-88e2-90afcbb3d880\" (UID: \"129baf51-56c6-488a-88e2-90afcbb3d880\") " Feb 18 09:19:58 crc kubenswrapper[4556]: I0218 09:19:58.613573 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6zccq\" (UniqueName: \"kubernetes.io/projected/129baf51-56c6-488a-88e2-90afcbb3d880-kube-api-access-6zccq\") pod \"129baf51-56c6-488a-88e2-90afcbb3d880\" (UID: \"129baf51-56c6-488a-88e2-90afcbb3d880\") " Feb 18 09:19:58 crc kubenswrapper[4556]: I0218 09:19:58.613602 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/129baf51-56c6-488a-88e2-90afcbb3d880-ovsdbserver-sb\") pod \"129baf51-56c6-488a-88e2-90afcbb3d880\" (UID: \"129baf51-56c6-488a-88e2-90afcbb3d880\") " Feb 18 09:19:58 crc kubenswrapper[4556]: I0218 09:19:58.613620 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/129baf51-56c6-488a-88e2-90afcbb3d880-dns-svc\") pod \"129baf51-56c6-488a-88e2-90afcbb3d880\" (UID: \"129baf51-56c6-488a-88e2-90afcbb3d880\") " Feb 18 09:19:58 crc kubenswrapper[4556]: I0218 09:19:58.613693 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/129baf51-56c6-488a-88e2-90afcbb3d880-config\") pod \"129baf51-56c6-488a-88e2-90afcbb3d880\" (UID: \"129baf51-56c6-488a-88e2-90afcbb3d880\") " Feb 18 09:19:58 crc kubenswrapper[4556]: I0218 09:19:58.621776 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/129baf51-56c6-488a-88e2-90afcbb3d880-kube-api-access-6zccq" (OuterVolumeSpecName: "kube-api-access-6zccq") pod "129baf51-56c6-488a-88e2-90afcbb3d880" (UID: "129baf51-56c6-488a-88e2-90afcbb3d880"). InnerVolumeSpecName "kube-api-access-6zccq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:19:58 crc kubenswrapper[4556]: I0218 09:19:58.669727 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/129baf51-56c6-488a-88e2-90afcbb3d880-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "129baf51-56c6-488a-88e2-90afcbb3d880" (UID: "129baf51-56c6-488a-88e2-90afcbb3d880"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:19:58 crc kubenswrapper[4556]: I0218 09:19:58.673648 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/129baf51-56c6-488a-88e2-90afcbb3d880-config" (OuterVolumeSpecName: "config") pod "129baf51-56c6-488a-88e2-90afcbb3d880" (UID: "129baf51-56c6-488a-88e2-90afcbb3d880"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:19:58 crc kubenswrapper[4556]: I0218 09:19:58.674437 4556 generic.go:334] "Generic (PLEG): container finished" podID="129baf51-56c6-488a-88e2-90afcbb3d880" containerID="6cbad85d5780b1b4e977194f87ef2f9451600f0524aecf9002beb6201ffcc69f" exitCode=0 Feb 18 09:19:58 crc kubenswrapper[4556]: I0218 09:19:58.674597 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="67a3f575-2f76-4447-b777-25b0b93ad11a" containerName="nova-scheduler-scheduler" containerID="cri-o://136ed6176137995b75948550d7dc95861646c7cc08f0e807d940a4c6a25153c7" gracePeriod=30 Feb 18 09:19:58 crc kubenswrapper[4556]: I0218 09:19:58.674960 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-849fd69845-bdbc8" event={"ID":"129baf51-56c6-488a-88e2-90afcbb3d880","Type":"ContainerDied","Data":"6cbad85d5780b1b4e977194f87ef2f9451600f0524aecf9002beb6201ffcc69f"} Feb 18 09:19:58 crc kubenswrapper[4556]: I0218 09:19:58.675018 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-849fd69845-bdbc8" event={"ID":"129baf51-56c6-488a-88e2-90afcbb3d880","Type":"ContainerDied","Data":"3a83c3f218b20ae33af533dcd9f74f5c2e0878811b25c91c0b1e9c44aa7886e1"} Feb 18 09:19:58 crc kubenswrapper[4556]: I0218 09:19:58.675020 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/129baf51-56c6-488a-88e2-90afcbb3d880-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "129baf51-56c6-488a-88e2-90afcbb3d880" (UID: "129baf51-56c6-488a-88e2-90afcbb3d880"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:19:58 crc kubenswrapper[4556]: I0218 09:19:58.675087 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-849fd69845-bdbc8" Feb 18 09:19:58 crc kubenswrapper[4556]: I0218 09:19:58.675037 4556 scope.go:117] "RemoveContainer" containerID="6cbad85d5780b1b4e977194f87ef2f9451600f0524aecf9002beb6201ffcc69f" Feb 18 09:19:58 crc kubenswrapper[4556]: I0218 09:19:58.675468 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d64f15b2-c072-4bf1-b60b-3e73d6df2414" containerName="nova-api-log" containerID="cri-o://02727f04dcdb857126a77711e83fdfc2bf3d35b672c828dcea31d7eb9ab89b9e" gracePeriod=30 Feb 18 09:19:58 crc kubenswrapper[4556]: I0218 09:19:58.675605 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d64f15b2-c072-4bf1-b60b-3e73d6df2414" containerName="nova-api-api" containerID="cri-o://c26e129cdfd10ac7f662bce7f0cd0c203089095198aae4710eadd93dc6092a61" gracePeriod=30 Feb 18 09:19:58 crc kubenswrapper[4556]: E0218 09:19:58.678858 4556 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="136ed6176137995b75948550d7dc95861646c7cc08f0e807d940a4c6a25153c7" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 18 09:19:58 crc kubenswrapper[4556]: I0218 09:19:58.684337 4556 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d64f15b2-c072-4bf1-b60b-3e73d6df2414" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.188:8774/\": EOF" Feb 18 09:19:58 crc kubenswrapper[4556]: I0218 09:19:58.684428 4556 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d64f15b2-c072-4bf1-b60b-3e73d6df2414" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.188:8774/\": EOF" Feb 18 09:19:58 crc kubenswrapper[4556]: I0218 09:19:58.689467 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/129baf51-56c6-488a-88e2-90afcbb3d880-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "129baf51-56c6-488a-88e2-90afcbb3d880" (UID: "129baf51-56c6-488a-88e2-90afcbb3d880"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:19:58 crc kubenswrapper[4556]: E0218 09:19:58.693913 4556 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="136ed6176137995b75948550d7dc95861646c7cc08f0e807d940a4c6a25153c7" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 18 09:19:58 crc kubenswrapper[4556]: E0218 09:19:58.704488 4556 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="136ed6176137995b75948550d7dc95861646c7cc08f0e807d940a4c6a25153c7" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 18 09:19:58 crc kubenswrapper[4556]: E0218 09:19:58.704539 4556 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="67a3f575-2f76-4447-b777-25b0b93ad11a" containerName="nova-scheduler-scheduler" Feb 18 09:19:58 crc kubenswrapper[4556]: I0218 09:19:58.711571 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/129baf51-56c6-488a-88e2-90afcbb3d880-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "129baf51-56c6-488a-88e2-90afcbb3d880" (UID: "129baf51-56c6-488a-88e2-90afcbb3d880"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:19:58 crc kubenswrapper[4556]: I0218 09:19:58.713382 4556 scope.go:117] "RemoveContainer" containerID="20742bc1bf20986ce09b83f67db4ef2c5890e69a52e453bf65c6c8a6b4b3a88b" Feb 18 09:19:58 crc kubenswrapper[4556]: I0218 09:19:58.716563 4556 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/129baf51-56c6-488a-88e2-90afcbb3d880-config\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:58 crc kubenswrapper[4556]: I0218 09:19:58.716587 4556 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/129baf51-56c6-488a-88e2-90afcbb3d880-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:58 crc kubenswrapper[4556]: I0218 09:19:58.716598 4556 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/129baf51-56c6-488a-88e2-90afcbb3d880-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:58 crc kubenswrapper[4556]: I0218 09:19:58.716609 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6zccq\" (UniqueName: \"kubernetes.io/projected/129baf51-56c6-488a-88e2-90afcbb3d880-kube-api-access-6zccq\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:58 crc kubenswrapper[4556]: I0218 09:19:58.716619 4556 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/129baf51-56c6-488a-88e2-90afcbb3d880-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:58 crc kubenswrapper[4556]: I0218 09:19:58.716627 4556 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/129baf51-56c6-488a-88e2-90afcbb3d880-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:58 crc kubenswrapper[4556]: I0218 09:19:58.732669 4556 scope.go:117] "RemoveContainer" containerID="6cbad85d5780b1b4e977194f87ef2f9451600f0524aecf9002beb6201ffcc69f" Feb 18 09:19:58 crc kubenswrapper[4556]: E0218 09:19:58.733022 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6cbad85d5780b1b4e977194f87ef2f9451600f0524aecf9002beb6201ffcc69f\": container with ID starting with 6cbad85d5780b1b4e977194f87ef2f9451600f0524aecf9002beb6201ffcc69f not found: ID does not exist" containerID="6cbad85d5780b1b4e977194f87ef2f9451600f0524aecf9002beb6201ffcc69f" Feb 18 09:19:58 crc kubenswrapper[4556]: I0218 09:19:58.733067 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6cbad85d5780b1b4e977194f87ef2f9451600f0524aecf9002beb6201ffcc69f"} err="failed to get container status \"6cbad85d5780b1b4e977194f87ef2f9451600f0524aecf9002beb6201ffcc69f\": rpc error: code = NotFound desc = could not find container \"6cbad85d5780b1b4e977194f87ef2f9451600f0524aecf9002beb6201ffcc69f\": container with ID starting with 6cbad85d5780b1b4e977194f87ef2f9451600f0524aecf9002beb6201ffcc69f not found: ID does not exist" Feb 18 09:19:58 crc kubenswrapper[4556]: I0218 09:19:58.733098 4556 scope.go:117] "RemoveContainer" containerID="20742bc1bf20986ce09b83f67db4ef2c5890e69a52e453bf65c6c8a6b4b3a88b" Feb 18 09:19:58 crc kubenswrapper[4556]: E0218 09:19:58.733386 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20742bc1bf20986ce09b83f67db4ef2c5890e69a52e453bf65c6c8a6b4b3a88b\": container with ID starting with 20742bc1bf20986ce09b83f67db4ef2c5890e69a52e453bf65c6c8a6b4b3a88b not found: ID does not exist" containerID="20742bc1bf20986ce09b83f67db4ef2c5890e69a52e453bf65c6c8a6b4b3a88b" Feb 18 09:19:58 crc kubenswrapper[4556]: I0218 09:19:58.733418 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20742bc1bf20986ce09b83f67db4ef2c5890e69a52e453bf65c6c8a6b4b3a88b"} err="failed to get container status \"20742bc1bf20986ce09b83f67db4ef2c5890e69a52e453bf65c6c8a6b4b3a88b\": rpc error: code = NotFound desc = could not find container \"20742bc1bf20986ce09b83f67db4ef2c5890e69a52e453bf65c6c8a6b4b3a88b\": container with ID starting with 20742bc1bf20986ce09b83f67db4ef2c5890e69a52e453bf65c6c8a6b4b3a88b not found: ID does not exist" Feb 18 09:19:59 crc kubenswrapper[4556]: I0218 09:19:59.008200 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-849fd69845-bdbc8"] Feb 18 09:19:59 crc kubenswrapper[4556]: I0218 09:19:59.014737 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-849fd69845-bdbc8"] Feb 18 09:19:59 crc kubenswrapper[4556]: I0218 09:19:59.293349 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="129baf51-56c6-488a-88e2-90afcbb3d880" path="/var/lib/kubelet/pods/129baf51-56c6-488a-88e2-90afcbb3d880/volumes" Feb 18 09:19:59 crc kubenswrapper[4556]: I0218 09:19:59.525280 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 18 09:19:59 crc kubenswrapper[4556]: I0218 09:19:59.643594 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67a3f575-2f76-4447-b777-25b0b93ad11a-combined-ca-bundle\") pod \"67a3f575-2f76-4447-b777-25b0b93ad11a\" (UID: \"67a3f575-2f76-4447-b777-25b0b93ad11a\") " Feb 18 09:19:59 crc kubenswrapper[4556]: I0218 09:19:59.643651 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ct5gz\" (UniqueName: \"kubernetes.io/projected/67a3f575-2f76-4447-b777-25b0b93ad11a-kube-api-access-ct5gz\") pod \"67a3f575-2f76-4447-b777-25b0b93ad11a\" (UID: \"67a3f575-2f76-4447-b777-25b0b93ad11a\") " Feb 18 09:19:59 crc kubenswrapper[4556]: I0218 09:19:59.643752 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67a3f575-2f76-4447-b777-25b0b93ad11a-config-data\") pod \"67a3f575-2f76-4447-b777-25b0b93ad11a\" (UID: \"67a3f575-2f76-4447-b777-25b0b93ad11a\") " Feb 18 09:19:59 crc kubenswrapper[4556]: I0218 09:19:59.649807 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67a3f575-2f76-4447-b777-25b0b93ad11a-kube-api-access-ct5gz" (OuterVolumeSpecName: "kube-api-access-ct5gz") pod "67a3f575-2f76-4447-b777-25b0b93ad11a" (UID: "67a3f575-2f76-4447-b777-25b0b93ad11a"). InnerVolumeSpecName "kube-api-access-ct5gz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:19:59 crc kubenswrapper[4556]: I0218 09:19:59.668893 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67a3f575-2f76-4447-b777-25b0b93ad11a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "67a3f575-2f76-4447-b777-25b0b93ad11a" (UID: "67a3f575-2f76-4447-b777-25b0b93ad11a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:19:59 crc kubenswrapper[4556]: I0218 09:19:59.669317 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67a3f575-2f76-4447-b777-25b0b93ad11a-config-data" (OuterVolumeSpecName: "config-data") pod "67a3f575-2f76-4447-b777-25b0b93ad11a" (UID: "67a3f575-2f76-4447-b777-25b0b93ad11a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:19:59 crc kubenswrapper[4556]: I0218 09:19:59.723054 4556 generic.go:334] "Generic (PLEG): container finished" podID="d64f15b2-c072-4bf1-b60b-3e73d6df2414" containerID="02727f04dcdb857126a77711e83fdfc2bf3d35b672c828dcea31d7eb9ab89b9e" exitCode=143 Feb 18 09:19:59 crc kubenswrapper[4556]: I0218 09:19:59.723235 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d64f15b2-c072-4bf1-b60b-3e73d6df2414","Type":"ContainerDied","Data":"02727f04dcdb857126a77711e83fdfc2bf3d35b672c828dcea31d7eb9ab89b9e"} Feb 18 09:19:59 crc kubenswrapper[4556]: I0218 09:19:59.728016 4556 generic.go:334] "Generic (PLEG): container finished" podID="67a3f575-2f76-4447-b777-25b0b93ad11a" containerID="136ed6176137995b75948550d7dc95861646c7cc08f0e807d940a4c6a25153c7" exitCode=0 Feb 18 09:19:59 crc kubenswrapper[4556]: I0218 09:19:59.728055 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"67a3f575-2f76-4447-b777-25b0b93ad11a","Type":"ContainerDied","Data":"136ed6176137995b75948550d7dc95861646c7cc08f0e807d940a4c6a25153c7"} Feb 18 09:19:59 crc kubenswrapper[4556]: I0218 09:19:59.728413 4556 scope.go:117] "RemoveContainer" containerID="136ed6176137995b75948550d7dc95861646c7cc08f0e807d940a4c6a25153c7" Feb 18 09:19:59 crc kubenswrapper[4556]: I0218 09:19:59.728080 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"67a3f575-2f76-4447-b777-25b0b93ad11a","Type":"ContainerDied","Data":"bfc9b8c15721abf316f82254b9f9e41ade7dfc282a53cc807f4daa2c8ca7cae5"} Feb 18 09:19:59 crc kubenswrapper[4556]: I0218 09:19:59.728505 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 18 09:19:59 crc kubenswrapper[4556]: I0218 09:19:59.745862 4556 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67a3f575-2f76-4447-b777-25b0b93ad11a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:59 crc kubenswrapper[4556]: I0218 09:19:59.745950 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ct5gz\" (UniqueName: \"kubernetes.io/projected/67a3f575-2f76-4447-b777-25b0b93ad11a-kube-api-access-ct5gz\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:59 crc kubenswrapper[4556]: I0218 09:19:59.746023 4556 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67a3f575-2f76-4447-b777-25b0b93ad11a-config-data\") on node \"crc\" DevicePath \"\"" Feb 18 09:19:59 crc kubenswrapper[4556]: I0218 09:19:59.752672 4556 scope.go:117] "RemoveContainer" containerID="136ed6176137995b75948550d7dc95861646c7cc08f0e807d940a4c6a25153c7" Feb 18 09:19:59 crc kubenswrapper[4556]: E0218 09:19:59.753185 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"136ed6176137995b75948550d7dc95861646c7cc08f0e807d940a4c6a25153c7\": container with ID starting with 136ed6176137995b75948550d7dc95861646c7cc08f0e807d940a4c6a25153c7 not found: ID does not exist" containerID="136ed6176137995b75948550d7dc95861646c7cc08f0e807d940a4c6a25153c7" Feb 18 09:19:59 crc kubenswrapper[4556]: I0218 09:19:59.753216 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"136ed6176137995b75948550d7dc95861646c7cc08f0e807d940a4c6a25153c7"} err="failed to get container status \"136ed6176137995b75948550d7dc95861646c7cc08f0e807d940a4c6a25153c7\": rpc error: code = NotFound desc = could not find container \"136ed6176137995b75948550d7dc95861646c7cc08f0e807d940a4c6a25153c7\": container with ID starting with 136ed6176137995b75948550d7dc95861646c7cc08f0e807d940a4c6a25153c7 not found: ID does not exist" Feb 18 09:19:59 crc kubenswrapper[4556]: I0218 09:19:59.776867 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 18 09:19:59 crc kubenswrapper[4556]: I0218 09:19:59.784352 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Feb 18 09:19:59 crc kubenswrapper[4556]: I0218 09:19:59.796298 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Feb 18 09:19:59 crc kubenswrapper[4556]: E0218 09:19:59.796916 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="129baf51-56c6-488a-88e2-90afcbb3d880" containerName="dnsmasq-dns" Feb 18 09:19:59 crc kubenswrapper[4556]: I0218 09:19:59.796995 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="129baf51-56c6-488a-88e2-90afcbb3d880" containerName="dnsmasq-dns" Feb 18 09:19:59 crc kubenswrapper[4556]: E0218 09:19:59.797066 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="129baf51-56c6-488a-88e2-90afcbb3d880" containerName="init" Feb 18 09:19:59 crc kubenswrapper[4556]: I0218 09:19:59.797115 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="129baf51-56c6-488a-88e2-90afcbb3d880" containerName="init" Feb 18 09:19:59 crc kubenswrapper[4556]: E0218 09:19:59.797197 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24dad0a1-29ed-49a7-be65-d024a02fcff4" containerName="nova-manage" Feb 18 09:19:59 crc kubenswrapper[4556]: I0218 09:19:59.797244 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="24dad0a1-29ed-49a7-be65-d024a02fcff4" containerName="nova-manage" Feb 18 09:19:59 crc kubenswrapper[4556]: E0218 09:19:59.797302 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67a3f575-2f76-4447-b777-25b0b93ad11a" containerName="nova-scheduler-scheduler" Feb 18 09:19:59 crc kubenswrapper[4556]: I0218 09:19:59.797359 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="67a3f575-2f76-4447-b777-25b0b93ad11a" containerName="nova-scheduler-scheduler" Feb 18 09:19:59 crc kubenswrapper[4556]: I0218 09:19:59.797605 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="129baf51-56c6-488a-88e2-90afcbb3d880" containerName="dnsmasq-dns" Feb 18 09:19:59 crc kubenswrapper[4556]: I0218 09:19:59.797672 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="24dad0a1-29ed-49a7-be65-d024a02fcff4" containerName="nova-manage" Feb 18 09:19:59 crc kubenswrapper[4556]: I0218 09:19:59.797727 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="67a3f575-2f76-4447-b777-25b0b93ad11a" containerName="nova-scheduler-scheduler" Feb 18 09:19:59 crc kubenswrapper[4556]: I0218 09:19:59.805048 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 18 09:19:59 crc kubenswrapper[4556]: I0218 09:19:59.808507 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Feb 18 09:19:59 crc kubenswrapper[4556]: I0218 09:19:59.815135 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 18 09:19:59 crc kubenswrapper[4556]: I0218 09:19:59.847522 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgvfb\" (UniqueName: \"kubernetes.io/projected/5d34e777-73c9-4dd3-8b9a-1e6fc8f8aefa-kube-api-access-pgvfb\") pod \"nova-scheduler-0\" (UID: \"5d34e777-73c9-4dd3-8b9a-1e6fc8f8aefa\") " pod="openstack/nova-scheduler-0" Feb 18 09:19:59 crc kubenswrapper[4556]: I0218 09:19:59.847862 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d34e777-73c9-4dd3-8b9a-1e6fc8f8aefa-config-data\") pod \"nova-scheduler-0\" (UID: \"5d34e777-73c9-4dd3-8b9a-1e6fc8f8aefa\") " pod="openstack/nova-scheduler-0" Feb 18 09:19:59 crc kubenswrapper[4556]: I0218 09:19:59.848345 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d34e777-73c9-4dd3-8b9a-1e6fc8f8aefa-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"5d34e777-73c9-4dd3-8b9a-1e6fc8f8aefa\") " pod="openstack/nova-scheduler-0" Feb 18 09:19:59 crc kubenswrapper[4556]: I0218 09:19:59.950445 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgvfb\" (UniqueName: \"kubernetes.io/projected/5d34e777-73c9-4dd3-8b9a-1e6fc8f8aefa-kube-api-access-pgvfb\") pod \"nova-scheduler-0\" (UID: \"5d34e777-73c9-4dd3-8b9a-1e6fc8f8aefa\") " pod="openstack/nova-scheduler-0" Feb 18 09:19:59 crc kubenswrapper[4556]: I0218 09:19:59.950733 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d34e777-73c9-4dd3-8b9a-1e6fc8f8aefa-config-data\") pod \"nova-scheduler-0\" (UID: \"5d34e777-73c9-4dd3-8b9a-1e6fc8f8aefa\") " pod="openstack/nova-scheduler-0" Feb 18 09:19:59 crc kubenswrapper[4556]: I0218 09:19:59.950829 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d34e777-73c9-4dd3-8b9a-1e6fc8f8aefa-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"5d34e777-73c9-4dd3-8b9a-1e6fc8f8aefa\") " pod="openstack/nova-scheduler-0" Feb 18 09:19:59 crc kubenswrapper[4556]: I0218 09:19:59.954468 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d34e777-73c9-4dd3-8b9a-1e6fc8f8aefa-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"5d34e777-73c9-4dd3-8b9a-1e6fc8f8aefa\") " pod="openstack/nova-scheduler-0" Feb 18 09:19:59 crc kubenswrapper[4556]: I0218 09:19:59.954755 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d34e777-73c9-4dd3-8b9a-1e6fc8f8aefa-config-data\") pod \"nova-scheduler-0\" (UID: \"5d34e777-73c9-4dd3-8b9a-1e6fc8f8aefa\") " pod="openstack/nova-scheduler-0" Feb 18 09:19:59 crc kubenswrapper[4556]: I0218 09:19:59.964994 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgvfb\" (UniqueName: \"kubernetes.io/projected/5d34e777-73c9-4dd3-8b9a-1e6fc8f8aefa-kube-api-access-pgvfb\") pod \"nova-scheduler-0\" (UID: \"5d34e777-73c9-4dd3-8b9a-1e6fc8f8aefa\") " pod="openstack/nova-scheduler-0" Feb 18 09:20:00 crc kubenswrapper[4556]: I0218 09:20:00.121377 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 18 09:20:00 crc kubenswrapper[4556]: I0218 09:20:00.521789 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 18 09:20:00 crc kubenswrapper[4556]: W0218 09:20:00.528785 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5d34e777_73c9_4dd3_8b9a_1e6fc8f8aefa.slice/crio-111bc0dcb3d47b87ca4e5a40e5ddbbdce67639d8fd5e1088f438944ec5f72af9 WatchSource:0}: Error finding container 111bc0dcb3d47b87ca4e5a40e5ddbbdce67639d8fd5e1088f438944ec5f72af9: Status 404 returned error can't find the container with id 111bc0dcb3d47b87ca4e5a40e5ddbbdce67639d8fd5e1088f438944ec5f72af9 Feb 18 09:20:00 crc kubenswrapper[4556]: I0218 09:20:00.741623 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"5d34e777-73c9-4dd3-8b9a-1e6fc8f8aefa","Type":"ContainerStarted","Data":"e233eebe3d83a854854bdd8bd048694b15180c1bbabf93145a0c20a68d8c5e35"} Feb 18 09:20:00 crc kubenswrapper[4556]: I0218 09:20:00.741868 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"5d34e777-73c9-4dd3-8b9a-1e6fc8f8aefa","Type":"ContainerStarted","Data":"111bc0dcb3d47b87ca4e5a40e5ddbbdce67639d8fd5e1088f438944ec5f72af9"} Feb 18 09:20:00 crc kubenswrapper[4556]: I0218 09:20:00.766562 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=1.7665342179999999 podStartE2EDuration="1.766534218s" podCreationTimestamp="2026-02-18 09:19:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:20:00.755661563 +0000 UTC m=+957.772622543" watchObservedRunningTime="2026-02-18 09:20:00.766534218 +0000 UTC m=+957.783495199" Feb 18 09:20:01 crc kubenswrapper[4556]: I0218 09:20:01.292080 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67a3f575-2f76-4447-b777-25b0b93ad11a" path="/var/lib/kubelet/pods/67a3f575-2f76-4447-b777-25b0b93ad11a/volumes" Feb 18 09:20:02 crc kubenswrapper[4556]: I0218 09:20:02.050466 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Feb 18 09:20:04 crc kubenswrapper[4556]: I0218 09:20:04.413669 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 18 09:20:04 crc kubenswrapper[4556]: I0218 09:20:04.455809 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d64f15b2-c072-4bf1-b60b-3e73d6df2414-combined-ca-bundle\") pod \"d64f15b2-c072-4bf1-b60b-3e73d6df2414\" (UID: \"d64f15b2-c072-4bf1-b60b-3e73d6df2414\") " Feb 18 09:20:04 crc kubenswrapper[4556]: I0218 09:20:04.455869 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d64f15b2-c072-4bf1-b60b-3e73d6df2414-config-data\") pod \"d64f15b2-c072-4bf1-b60b-3e73d6df2414\" (UID: \"d64f15b2-c072-4bf1-b60b-3e73d6df2414\") " Feb 18 09:20:04 crc kubenswrapper[4556]: I0218 09:20:04.455893 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d64f15b2-c072-4bf1-b60b-3e73d6df2414-logs\") pod \"d64f15b2-c072-4bf1-b60b-3e73d6df2414\" (UID: \"d64f15b2-c072-4bf1-b60b-3e73d6df2414\") " Feb 18 09:20:04 crc kubenswrapper[4556]: I0218 09:20:04.456043 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qcslx\" (UniqueName: \"kubernetes.io/projected/d64f15b2-c072-4bf1-b60b-3e73d6df2414-kube-api-access-qcslx\") pod \"d64f15b2-c072-4bf1-b60b-3e73d6df2414\" (UID: \"d64f15b2-c072-4bf1-b60b-3e73d6df2414\") " Feb 18 09:20:04 crc kubenswrapper[4556]: I0218 09:20:04.456389 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d64f15b2-c072-4bf1-b60b-3e73d6df2414-logs" (OuterVolumeSpecName: "logs") pod "d64f15b2-c072-4bf1-b60b-3e73d6df2414" (UID: "d64f15b2-c072-4bf1-b60b-3e73d6df2414"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:20:04 crc kubenswrapper[4556]: I0218 09:20:04.456869 4556 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d64f15b2-c072-4bf1-b60b-3e73d6df2414-logs\") on node \"crc\" DevicePath \"\"" Feb 18 09:20:04 crc kubenswrapper[4556]: I0218 09:20:04.463195 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d64f15b2-c072-4bf1-b60b-3e73d6df2414-kube-api-access-qcslx" (OuterVolumeSpecName: "kube-api-access-qcslx") pod "d64f15b2-c072-4bf1-b60b-3e73d6df2414" (UID: "d64f15b2-c072-4bf1-b60b-3e73d6df2414"). InnerVolumeSpecName "kube-api-access-qcslx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:20:04 crc kubenswrapper[4556]: I0218 09:20:04.484350 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d64f15b2-c072-4bf1-b60b-3e73d6df2414-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d64f15b2-c072-4bf1-b60b-3e73d6df2414" (UID: "d64f15b2-c072-4bf1-b60b-3e73d6df2414"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:20:04 crc kubenswrapper[4556]: I0218 09:20:04.485487 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d64f15b2-c072-4bf1-b60b-3e73d6df2414-config-data" (OuterVolumeSpecName: "config-data") pod "d64f15b2-c072-4bf1-b60b-3e73d6df2414" (UID: "d64f15b2-c072-4bf1-b60b-3e73d6df2414"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:20:04 crc kubenswrapper[4556]: I0218 09:20:04.559119 4556 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d64f15b2-c072-4bf1-b60b-3e73d6df2414-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:20:04 crc kubenswrapper[4556]: I0218 09:20:04.559167 4556 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d64f15b2-c072-4bf1-b60b-3e73d6df2414-config-data\") on node \"crc\" DevicePath \"\"" Feb 18 09:20:04 crc kubenswrapper[4556]: I0218 09:20:04.559178 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qcslx\" (UniqueName: \"kubernetes.io/projected/d64f15b2-c072-4bf1-b60b-3e73d6df2414-kube-api-access-qcslx\") on node \"crc\" DevicePath \"\"" Feb 18 09:20:04 crc kubenswrapper[4556]: I0218 09:20:04.782496 4556 generic.go:334] "Generic (PLEG): container finished" podID="d64f15b2-c072-4bf1-b60b-3e73d6df2414" containerID="c26e129cdfd10ac7f662bce7f0cd0c203089095198aae4710eadd93dc6092a61" exitCode=0 Feb 18 09:20:04 crc kubenswrapper[4556]: I0218 09:20:04.782563 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d64f15b2-c072-4bf1-b60b-3e73d6df2414","Type":"ContainerDied","Data":"c26e129cdfd10ac7f662bce7f0cd0c203089095198aae4710eadd93dc6092a61"} Feb 18 09:20:04 crc kubenswrapper[4556]: I0218 09:20:04.782609 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d64f15b2-c072-4bf1-b60b-3e73d6df2414","Type":"ContainerDied","Data":"c29843d3a2915e6608ef40777d7dda94e47f77b27730fb500caa53d064c5a9a3"} Feb 18 09:20:04 crc kubenswrapper[4556]: I0218 09:20:04.782632 4556 scope.go:117] "RemoveContainer" containerID="c26e129cdfd10ac7f662bce7f0cd0c203089095198aae4710eadd93dc6092a61" Feb 18 09:20:04 crc kubenswrapper[4556]: I0218 09:20:04.782630 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 18 09:20:04 crc kubenswrapper[4556]: I0218 09:20:04.805452 4556 scope.go:117] "RemoveContainer" containerID="02727f04dcdb857126a77711e83fdfc2bf3d35b672c828dcea31d7eb9ab89b9e" Feb 18 09:20:04 crc kubenswrapper[4556]: I0218 09:20:04.814389 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 18 09:20:04 crc kubenswrapper[4556]: I0218 09:20:04.824931 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 18 09:20:04 crc kubenswrapper[4556]: I0218 09:20:04.825409 4556 scope.go:117] "RemoveContainer" containerID="c26e129cdfd10ac7f662bce7f0cd0c203089095198aae4710eadd93dc6092a61" Feb 18 09:20:04 crc kubenswrapper[4556]: E0218 09:20:04.825726 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c26e129cdfd10ac7f662bce7f0cd0c203089095198aae4710eadd93dc6092a61\": container with ID starting with c26e129cdfd10ac7f662bce7f0cd0c203089095198aae4710eadd93dc6092a61 not found: ID does not exist" containerID="c26e129cdfd10ac7f662bce7f0cd0c203089095198aae4710eadd93dc6092a61" Feb 18 09:20:04 crc kubenswrapper[4556]: I0218 09:20:04.825760 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c26e129cdfd10ac7f662bce7f0cd0c203089095198aae4710eadd93dc6092a61"} err="failed to get container status \"c26e129cdfd10ac7f662bce7f0cd0c203089095198aae4710eadd93dc6092a61\": rpc error: code = NotFound desc = could not find container \"c26e129cdfd10ac7f662bce7f0cd0c203089095198aae4710eadd93dc6092a61\": container with ID starting with c26e129cdfd10ac7f662bce7f0cd0c203089095198aae4710eadd93dc6092a61 not found: ID does not exist" Feb 18 09:20:04 crc kubenswrapper[4556]: I0218 09:20:04.825782 4556 scope.go:117] "RemoveContainer" containerID="02727f04dcdb857126a77711e83fdfc2bf3d35b672c828dcea31d7eb9ab89b9e" Feb 18 09:20:04 crc kubenswrapper[4556]: E0218 09:20:04.826541 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02727f04dcdb857126a77711e83fdfc2bf3d35b672c828dcea31d7eb9ab89b9e\": container with ID starting with 02727f04dcdb857126a77711e83fdfc2bf3d35b672c828dcea31d7eb9ab89b9e not found: ID does not exist" containerID="02727f04dcdb857126a77711e83fdfc2bf3d35b672c828dcea31d7eb9ab89b9e" Feb 18 09:20:04 crc kubenswrapper[4556]: I0218 09:20:04.826569 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02727f04dcdb857126a77711e83fdfc2bf3d35b672c828dcea31d7eb9ab89b9e"} err="failed to get container status \"02727f04dcdb857126a77711e83fdfc2bf3d35b672c828dcea31d7eb9ab89b9e\": rpc error: code = NotFound desc = could not find container \"02727f04dcdb857126a77711e83fdfc2bf3d35b672c828dcea31d7eb9ab89b9e\": container with ID starting with 02727f04dcdb857126a77711e83fdfc2bf3d35b672c828dcea31d7eb9ab89b9e not found: ID does not exist" Feb 18 09:20:04 crc kubenswrapper[4556]: I0218 09:20:04.832755 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 18 09:20:04 crc kubenswrapper[4556]: E0218 09:20:04.833095 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d64f15b2-c072-4bf1-b60b-3e73d6df2414" containerName="nova-api-api" Feb 18 09:20:04 crc kubenswrapper[4556]: I0218 09:20:04.833113 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="d64f15b2-c072-4bf1-b60b-3e73d6df2414" containerName="nova-api-api" Feb 18 09:20:04 crc kubenswrapper[4556]: E0218 09:20:04.833126 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d64f15b2-c072-4bf1-b60b-3e73d6df2414" containerName="nova-api-log" Feb 18 09:20:04 crc kubenswrapper[4556]: I0218 09:20:04.833134 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="d64f15b2-c072-4bf1-b60b-3e73d6df2414" containerName="nova-api-log" Feb 18 09:20:04 crc kubenswrapper[4556]: I0218 09:20:04.833356 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="d64f15b2-c072-4bf1-b60b-3e73d6df2414" containerName="nova-api-log" Feb 18 09:20:04 crc kubenswrapper[4556]: I0218 09:20:04.833375 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="d64f15b2-c072-4bf1-b60b-3e73d6df2414" containerName="nova-api-api" Feb 18 09:20:04 crc kubenswrapper[4556]: I0218 09:20:04.834703 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 18 09:20:04 crc kubenswrapper[4556]: I0218 09:20:04.837528 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 18 09:20:04 crc kubenswrapper[4556]: I0218 09:20:04.840418 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 18 09:20:04 crc kubenswrapper[4556]: I0218 09:20:04.865087 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c7a2344-f7ad-40c2-84e9-60abcf73571a-config-data\") pod \"nova-api-0\" (UID: \"4c7a2344-f7ad-40c2-84e9-60abcf73571a\") " pod="openstack/nova-api-0" Feb 18 09:20:04 crc kubenswrapper[4556]: I0218 09:20:04.865303 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4c7a2344-f7ad-40c2-84e9-60abcf73571a-logs\") pod \"nova-api-0\" (UID: \"4c7a2344-f7ad-40c2-84e9-60abcf73571a\") " pod="openstack/nova-api-0" Feb 18 09:20:04 crc kubenswrapper[4556]: I0218 09:20:04.865512 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c7a2344-f7ad-40c2-84e9-60abcf73571a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4c7a2344-f7ad-40c2-84e9-60abcf73571a\") " pod="openstack/nova-api-0" Feb 18 09:20:04 crc kubenswrapper[4556]: I0218 09:20:04.865617 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4h5t\" (UniqueName: \"kubernetes.io/projected/4c7a2344-f7ad-40c2-84e9-60abcf73571a-kube-api-access-c4h5t\") pod \"nova-api-0\" (UID: \"4c7a2344-f7ad-40c2-84e9-60abcf73571a\") " pod="openstack/nova-api-0" Feb 18 09:20:04 crc kubenswrapper[4556]: I0218 09:20:04.967450 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c7a2344-f7ad-40c2-84e9-60abcf73571a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4c7a2344-f7ad-40c2-84e9-60abcf73571a\") " pod="openstack/nova-api-0" Feb 18 09:20:04 crc kubenswrapper[4556]: I0218 09:20:04.967699 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4h5t\" (UniqueName: \"kubernetes.io/projected/4c7a2344-f7ad-40c2-84e9-60abcf73571a-kube-api-access-c4h5t\") pod \"nova-api-0\" (UID: \"4c7a2344-f7ad-40c2-84e9-60abcf73571a\") " pod="openstack/nova-api-0" Feb 18 09:20:04 crc kubenswrapper[4556]: I0218 09:20:04.967986 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c7a2344-f7ad-40c2-84e9-60abcf73571a-config-data\") pod \"nova-api-0\" (UID: \"4c7a2344-f7ad-40c2-84e9-60abcf73571a\") " pod="openstack/nova-api-0" Feb 18 09:20:04 crc kubenswrapper[4556]: I0218 09:20:04.968251 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4c7a2344-f7ad-40c2-84e9-60abcf73571a-logs\") pod \"nova-api-0\" (UID: \"4c7a2344-f7ad-40c2-84e9-60abcf73571a\") " pod="openstack/nova-api-0" Feb 18 09:20:04 crc kubenswrapper[4556]: I0218 09:20:04.968870 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4c7a2344-f7ad-40c2-84e9-60abcf73571a-logs\") pod \"nova-api-0\" (UID: \"4c7a2344-f7ad-40c2-84e9-60abcf73571a\") " pod="openstack/nova-api-0" Feb 18 09:20:04 crc kubenswrapper[4556]: I0218 09:20:04.972086 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c7a2344-f7ad-40c2-84e9-60abcf73571a-config-data\") pod \"nova-api-0\" (UID: \"4c7a2344-f7ad-40c2-84e9-60abcf73571a\") " pod="openstack/nova-api-0" Feb 18 09:20:04 crc kubenswrapper[4556]: I0218 09:20:04.972828 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c7a2344-f7ad-40c2-84e9-60abcf73571a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4c7a2344-f7ad-40c2-84e9-60abcf73571a\") " pod="openstack/nova-api-0" Feb 18 09:20:04 crc kubenswrapper[4556]: I0218 09:20:04.984899 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4h5t\" (UniqueName: \"kubernetes.io/projected/4c7a2344-f7ad-40c2-84e9-60abcf73571a-kube-api-access-c4h5t\") pod \"nova-api-0\" (UID: \"4c7a2344-f7ad-40c2-84e9-60abcf73571a\") " pod="openstack/nova-api-0" Feb 18 09:20:05 crc kubenswrapper[4556]: I0218 09:20:05.122042 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Feb 18 09:20:05 crc kubenswrapper[4556]: I0218 09:20:05.159303 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 18 09:20:05 crc kubenswrapper[4556]: I0218 09:20:05.292509 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d64f15b2-c072-4bf1-b60b-3e73d6df2414" path="/var/lib/kubelet/pods/d64f15b2-c072-4bf1-b60b-3e73d6df2414/volumes" Feb 18 09:20:05 crc kubenswrapper[4556]: I0218 09:20:05.565680 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 18 09:20:05 crc kubenswrapper[4556]: I0218 09:20:05.791810 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4c7a2344-f7ad-40c2-84e9-60abcf73571a","Type":"ContainerStarted","Data":"6b0cc9a7e039bccb8d4df425c5577d32ffa7d1ec8bf2b43efd97ff0ee1d5bb50"} Feb 18 09:20:05 crc kubenswrapper[4556]: I0218 09:20:05.792199 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4c7a2344-f7ad-40c2-84e9-60abcf73571a","Type":"ContainerStarted","Data":"96332f1c6215a8fc2e3fab8bf3b6de057e91e68295678a607be37be840224e09"} Feb 18 09:20:06 crc kubenswrapper[4556]: I0218 09:20:06.802709 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4c7a2344-f7ad-40c2-84e9-60abcf73571a","Type":"ContainerStarted","Data":"ba073e3d7ddd3fbd98e6ef10ee94d22dd5e5331510f771de93405ce4667d327e"} Feb 18 09:20:06 crc kubenswrapper[4556]: I0218 09:20:06.820280 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.820258804 podStartE2EDuration="2.820258804s" podCreationTimestamp="2026-02-18 09:20:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:20:06.818209646 +0000 UTC m=+963.835170626" watchObservedRunningTime="2026-02-18 09:20:06.820258804 +0000 UTC m=+963.837219783" Feb 18 09:20:08 crc kubenswrapper[4556]: I0218 09:20:08.695090 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Feb 18 09:20:10 crc kubenswrapper[4556]: I0218 09:20:10.122223 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Feb 18 09:20:10 crc kubenswrapper[4556]: I0218 09:20:10.150022 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Feb 18 09:20:10 crc kubenswrapper[4556]: I0218 09:20:10.858060 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Feb 18 09:20:11 crc kubenswrapper[4556]: I0218 09:20:11.657287 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 18 09:20:11 crc kubenswrapper[4556]: I0218 09:20:11.657696 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="95f33d13-a69d-4add-88a3-21126d52a0ed" containerName="kube-state-metrics" containerID="cri-o://4318c39a39db1314fa695847829c132f7eaf428f7ea492aabc0dd8c161ec63ff" gracePeriod=30 Feb 18 09:20:12 crc kubenswrapper[4556]: I0218 09:20:12.176001 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 18 09:20:12 crc kubenswrapper[4556]: I0218 09:20:12.225417 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lrmbg\" (UniqueName: \"kubernetes.io/projected/95f33d13-a69d-4add-88a3-21126d52a0ed-kube-api-access-lrmbg\") pod \"95f33d13-a69d-4add-88a3-21126d52a0ed\" (UID: \"95f33d13-a69d-4add-88a3-21126d52a0ed\") " Feb 18 09:20:12 crc kubenswrapper[4556]: I0218 09:20:12.232920 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95f33d13-a69d-4add-88a3-21126d52a0ed-kube-api-access-lrmbg" (OuterVolumeSpecName: "kube-api-access-lrmbg") pod "95f33d13-a69d-4add-88a3-21126d52a0ed" (UID: "95f33d13-a69d-4add-88a3-21126d52a0ed"). InnerVolumeSpecName "kube-api-access-lrmbg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:20:12 crc kubenswrapper[4556]: I0218 09:20:12.329341 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lrmbg\" (UniqueName: \"kubernetes.io/projected/95f33d13-a69d-4add-88a3-21126d52a0ed-kube-api-access-lrmbg\") on node \"crc\" DevicePath \"\"" Feb 18 09:20:12 crc kubenswrapper[4556]: I0218 09:20:12.863708 4556 generic.go:334] "Generic (PLEG): container finished" podID="95f33d13-a69d-4add-88a3-21126d52a0ed" containerID="4318c39a39db1314fa695847829c132f7eaf428f7ea492aabc0dd8c161ec63ff" exitCode=2 Feb 18 09:20:12 crc kubenswrapper[4556]: I0218 09:20:12.863781 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"95f33d13-a69d-4add-88a3-21126d52a0ed","Type":"ContainerDied","Data":"4318c39a39db1314fa695847829c132f7eaf428f7ea492aabc0dd8c161ec63ff"} Feb 18 09:20:12 crc kubenswrapper[4556]: I0218 09:20:12.863787 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 18 09:20:12 crc kubenswrapper[4556]: I0218 09:20:12.863851 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"95f33d13-a69d-4add-88a3-21126d52a0ed","Type":"ContainerDied","Data":"d4b199f9ebbe07f1c5a9f8a157f6d56adb74becd56ff76cb650f578ac3dcd5f6"} Feb 18 09:20:12 crc kubenswrapper[4556]: I0218 09:20:12.863878 4556 scope.go:117] "RemoveContainer" containerID="4318c39a39db1314fa695847829c132f7eaf428f7ea492aabc0dd8c161ec63ff" Feb 18 09:20:12 crc kubenswrapper[4556]: I0218 09:20:12.896916 4556 scope.go:117] "RemoveContainer" containerID="4318c39a39db1314fa695847829c132f7eaf428f7ea492aabc0dd8c161ec63ff" Feb 18 09:20:12 crc kubenswrapper[4556]: I0218 09:20:12.897081 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 18 09:20:12 crc kubenswrapper[4556]: E0218 09:20:12.897616 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4318c39a39db1314fa695847829c132f7eaf428f7ea492aabc0dd8c161ec63ff\": container with ID starting with 4318c39a39db1314fa695847829c132f7eaf428f7ea492aabc0dd8c161ec63ff not found: ID does not exist" containerID="4318c39a39db1314fa695847829c132f7eaf428f7ea492aabc0dd8c161ec63ff" Feb 18 09:20:12 crc kubenswrapper[4556]: I0218 09:20:12.897654 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4318c39a39db1314fa695847829c132f7eaf428f7ea492aabc0dd8c161ec63ff"} err="failed to get container status \"4318c39a39db1314fa695847829c132f7eaf428f7ea492aabc0dd8c161ec63ff\": rpc error: code = NotFound desc = could not find container \"4318c39a39db1314fa695847829c132f7eaf428f7ea492aabc0dd8c161ec63ff\": container with ID starting with 4318c39a39db1314fa695847829c132f7eaf428f7ea492aabc0dd8c161ec63ff not found: ID does not exist" Feb 18 09:20:12 crc kubenswrapper[4556]: I0218 09:20:12.905231 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 18 09:20:12 crc kubenswrapper[4556]: I0218 09:20:12.914538 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Feb 18 09:20:12 crc kubenswrapper[4556]: E0218 09:20:12.914959 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95f33d13-a69d-4add-88a3-21126d52a0ed" containerName="kube-state-metrics" Feb 18 09:20:12 crc kubenswrapper[4556]: I0218 09:20:12.914978 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="95f33d13-a69d-4add-88a3-21126d52a0ed" containerName="kube-state-metrics" Feb 18 09:20:12 crc kubenswrapper[4556]: I0218 09:20:12.915165 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="95f33d13-a69d-4add-88a3-21126d52a0ed" containerName="kube-state-metrics" Feb 18 09:20:12 crc kubenswrapper[4556]: I0218 09:20:12.915863 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 18 09:20:12 crc kubenswrapper[4556]: I0218 09:20:12.918637 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Feb 18 09:20:12 crc kubenswrapper[4556]: I0218 09:20:12.918761 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Feb 18 09:20:12 crc kubenswrapper[4556]: I0218 09:20:12.923437 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 18 09:20:12 crc kubenswrapper[4556]: I0218 09:20:12.940844 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4bshg\" (UniqueName: \"kubernetes.io/projected/e54e3af3-e2c2-4e5b-885d-0071d05acbb4-kube-api-access-4bshg\") pod \"kube-state-metrics-0\" (UID: \"e54e3af3-e2c2-4e5b-885d-0071d05acbb4\") " pod="openstack/kube-state-metrics-0" Feb 18 09:20:12 crc kubenswrapper[4556]: I0218 09:20:12.941250 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e54e3af3-e2c2-4e5b-885d-0071d05acbb4-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"e54e3af3-e2c2-4e5b-885d-0071d05acbb4\") " pod="openstack/kube-state-metrics-0" Feb 18 09:20:12 crc kubenswrapper[4556]: I0218 09:20:12.941419 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/e54e3af3-e2c2-4e5b-885d-0071d05acbb4-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"e54e3af3-e2c2-4e5b-885d-0071d05acbb4\") " pod="openstack/kube-state-metrics-0" Feb 18 09:20:12 crc kubenswrapper[4556]: I0218 09:20:12.941611 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/e54e3af3-e2c2-4e5b-885d-0071d05acbb4-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"e54e3af3-e2c2-4e5b-885d-0071d05acbb4\") " pod="openstack/kube-state-metrics-0" Feb 18 09:20:13 crc kubenswrapper[4556]: I0218 09:20:13.043309 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e54e3af3-e2c2-4e5b-885d-0071d05acbb4-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"e54e3af3-e2c2-4e5b-885d-0071d05acbb4\") " pod="openstack/kube-state-metrics-0" Feb 18 09:20:13 crc kubenswrapper[4556]: I0218 09:20:13.043626 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/e54e3af3-e2c2-4e5b-885d-0071d05acbb4-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"e54e3af3-e2c2-4e5b-885d-0071d05acbb4\") " pod="openstack/kube-state-metrics-0" Feb 18 09:20:13 crc kubenswrapper[4556]: I0218 09:20:13.043727 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/e54e3af3-e2c2-4e5b-885d-0071d05acbb4-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"e54e3af3-e2c2-4e5b-885d-0071d05acbb4\") " pod="openstack/kube-state-metrics-0" Feb 18 09:20:13 crc kubenswrapper[4556]: I0218 09:20:13.043814 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4bshg\" (UniqueName: \"kubernetes.io/projected/e54e3af3-e2c2-4e5b-885d-0071d05acbb4-kube-api-access-4bshg\") pod \"kube-state-metrics-0\" (UID: \"e54e3af3-e2c2-4e5b-885d-0071d05acbb4\") " pod="openstack/kube-state-metrics-0" Feb 18 09:20:13 crc kubenswrapper[4556]: I0218 09:20:13.047797 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e54e3af3-e2c2-4e5b-885d-0071d05acbb4-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"e54e3af3-e2c2-4e5b-885d-0071d05acbb4\") " pod="openstack/kube-state-metrics-0" Feb 18 09:20:13 crc kubenswrapper[4556]: I0218 09:20:13.048210 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/e54e3af3-e2c2-4e5b-885d-0071d05acbb4-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"e54e3af3-e2c2-4e5b-885d-0071d05acbb4\") " pod="openstack/kube-state-metrics-0" Feb 18 09:20:13 crc kubenswrapper[4556]: I0218 09:20:13.048472 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/e54e3af3-e2c2-4e5b-885d-0071d05acbb4-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"e54e3af3-e2c2-4e5b-885d-0071d05acbb4\") " pod="openstack/kube-state-metrics-0" Feb 18 09:20:13 crc kubenswrapper[4556]: I0218 09:20:13.061938 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4bshg\" (UniqueName: \"kubernetes.io/projected/e54e3af3-e2c2-4e5b-885d-0071d05acbb4-kube-api-access-4bshg\") pod \"kube-state-metrics-0\" (UID: \"e54e3af3-e2c2-4e5b-885d-0071d05acbb4\") " pod="openstack/kube-state-metrics-0" Feb 18 09:20:13 crc kubenswrapper[4556]: I0218 09:20:13.224852 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 18 09:20:13 crc kubenswrapper[4556]: I0218 09:20:13.225233 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f16caa9f-5685-4f11-a112-65d1ed02241b" containerName="ceilometer-central-agent" containerID="cri-o://9bcf089e0dd023a0eb2a47ebef6f90f371b331baf6e3080a9a09414e4ae043b7" gracePeriod=30 Feb 18 09:20:13 crc kubenswrapper[4556]: I0218 09:20:13.225258 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f16caa9f-5685-4f11-a112-65d1ed02241b" containerName="proxy-httpd" containerID="cri-o://cb1f58413a0abdbb6acb88e4efccd9fa33ff1fcb407f6954739fcde9cbfd29bb" gracePeriod=30 Feb 18 09:20:13 crc kubenswrapper[4556]: I0218 09:20:13.225233 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f16caa9f-5685-4f11-a112-65d1ed02241b" containerName="sg-core" containerID="cri-o://90309d247b8912b399ac4dd0bcf7c92ed86b4ad0f5768d5cc4813794ab056e5a" gracePeriod=30 Feb 18 09:20:13 crc kubenswrapper[4556]: I0218 09:20:13.225361 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f16caa9f-5685-4f11-a112-65d1ed02241b" containerName="ceilometer-notification-agent" containerID="cri-o://6d63c4d4d746825902bcb72d7e5602ab09335c7b337d7c0363dba47c1d39b045" gracePeriod=30 Feb 18 09:20:13 crc kubenswrapper[4556]: I0218 09:20:13.233782 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 18 09:20:13 crc kubenswrapper[4556]: I0218 09:20:13.307607 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95f33d13-a69d-4add-88a3-21126d52a0ed" path="/var/lib/kubelet/pods/95f33d13-a69d-4add-88a3-21126d52a0ed/volumes" Feb 18 09:20:13 crc kubenswrapper[4556]: I0218 09:20:13.661596 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 18 09:20:13 crc kubenswrapper[4556]: I0218 09:20:13.872690 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"e54e3af3-e2c2-4e5b-885d-0071d05acbb4","Type":"ContainerStarted","Data":"84dcbd110677b23fbe5a4536579992113b9e49f7f1818820c12c7adb1629b01e"} Feb 18 09:20:13 crc kubenswrapper[4556]: I0218 09:20:13.875602 4556 generic.go:334] "Generic (PLEG): container finished" podID="f16caa9f-5685-4f11-a112-65d1ed02241b" containerID="cb1f58413a0abdbb6acb88e4efccd9fa33ff1fcb407f6954739fcde9cbfd29bb" exitCode=0 Feb 18 09:20:13 crc kubenswrapper[4556]: I0218 09:20:13.875627 4556 generic.go:334] "Generic (PLEG): container finished" podID="f16caa9f-5685-4f11-a112-65d1ed02241b" containerID="90309d247b8912b399ac4dd0bcf7c92ed86b4ad0f5768d5cc4813794ab056e5a" exitCode=2 Feb 18 09:20:13 crc kubenswrapper[4556]: I0218 09:20:13.875635 4556 generic.go:334] "Generic (PLEG): container finished" podID="f16caa9f-5685-4f11-a112-65d1ed02241b" containerID="9bcf089e0dd023a0eb2a47ebef6f90f371b331baf6e3080a9a09414e4ae043b7" exitCode=0 Feb 18 09:20:13 crc kubenswrapper[4556]: I0218 09:20:13.875650 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f16caa9f-5685-4f11-a112-65d1ed02241b","Type":"ContainerDied","Data":"cb1f58413a0abdbb6acb88e4efccd9fa33ff1fcb407f6954739fcde9cbfd29bb"} Feb 18 09:20:13 crc kubenswrapper[4556]: I0218 09:20:13.875668 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f16caa9f-5685-4f11-a112-65d1ed02241b","Type":"ContainerDied","Data":"90309d247b8912b399ac4dd0bcf7c92ed86b4ad0f5768d5cc4813794ab056e5a"} Feb 18 09:20:13 crc kubenswrapper[4556]: I0218 09:20:13.875678 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f16caa9f-5685-4f11-a112-65d1ed02241b","Type":"ContainerDied","Data":"9bcf089e0dd023a0eb2a47ebef6f90f371b331baf6e3080a9a09414e4ae043b7"} Feb 18 09:20:14 crc kubenswrapper[4556]: I0218 09:20:14.884808 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"e54e3af3-e2c2-4e5b-885d-0071d05acbb4","Type":"ContainerStarted","Data":"0d766086882683de01b81a803402454e61e68ab895b48f4458c41ee64e907add"} Feb 18 09:20:14 crc kubenswrapper[4556]: I0218 09:20:14.885196 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Feb 18 09:20:14 crc kubenswrapper[4556]: I0218 09:20:14.901266 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.606157718 podStartE2EDuration="2.901249527s" podCreationTimestamp="2026-02-18 09:20:12 +0000 UTC" firstStartedPulling="2026-02-18 09:20:13.668030381 +0000 UTC m=+970.684991361" lastFinishedPulling="2026-02-18 09:20:13.963122189 +0000 UTC m=+970.980083170" observedRunningTime="2026-02-18 09:20:14.89789082 +0000 UTC m=+971.914851800" watchObservedRunningTime="2026-02-18 09:20:14.901249527 +0000 UTC m=+971.918210507" Feb 18 09:20:15 crc kubenswrapper[4556]: I0218 09:20:15.160406 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 18 09:20:15 crc kubenswrapper[4556]: I0218 09:20:15.160464 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 18 09:20:15 crc kubenswrapper[4556]: I0218 09:20:15.791531 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 18 09:20:15 crc kubenswrapper[4556]: I0218 09:20:15.890422 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f16caa9f-5685-4f11-a112-65d1ed02241b-config-data\") pod \"f16caa9f-5685-4f11-a112-65d1ed02241b\" (UID: \"f16caa9f-5685-4f11-a112-65d1ed02241b\") " Feb 18 09:20:15 crc kubenswrapper[4556]: I0218 09:20:15.890702 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f16caa9f-5685-4f11-a112-65d1ed02241b-scripts\") pod \"f16caa9f-5685-4f11-a112-65d1ed02241b\" (UID: \"f16caa9f-5685-4f11-a112-65d1ed02241b\") " Feb 18 09:20:15 crc kubenswrapper[4556]: I0218 09:20:15.890730 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f16caa9f-5685-4f11-a112-65d1ed02241b-sg-core-conf-yaml\") pod \"f16caa9f-5685-4f11-a112-65d1ed02241b\" (UID: \"f16caa9f-5685-4f11-a112-65d1ed02241b\") " Feb 18 09:20:15 crc kubenswrapper[4556]: I0218 09:20:15.890765 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f16caa9f-5685-4f11-a112-65d1ed02241b-combined-ca-bundle\") pod \"f16caa9f-5685-4f11-a112-65d1ed02241b\" (UID: \"f16caa9f-5685-4f11-a112-65d1ed02241b\") " Feb 18 09:20:15 crc kubenswrapper[4556]: I0218 09:20:15.890796 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f16caa9f-5685-4f11-a112-65d1ed02241b-run-httpd\") pod \"f16caa9f-5685-4f11-a112-65d1ed02241b\" (UID: \"f16caa9f-5685-4f11-a112-65d1ed02241b\") " Feb 18 09:20:15 crc kubenswrapper[4556]: I0218 09:20:15.890812 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f16caa9f-5685-4f11-a112-65d1ed02241b-log-httpd\") pod \"f16caa9f-5685-4f11-a112-65d1ed02241b\" (UID: \"f16caa9f-5685-4f11-a112-65d1ed02241b\") " Feb 18 09:20:15 crc kubenswrapper[4556]: I0218 09:20:15.890827 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9hjm9\" (UniqueName: \"kubernetes.io/projected/f16caa9f-5685-4f11-a112-65d1ed02241b-kube-api-access-9hjm9\") pod \"f16caa9f-5685-4f11-a112-65d1ed02241b\" (UID: \"f16caa9f-5685-4f11-a112-65d1ed02241b\") " Feb 18 09:20:15 crc kubenswrapper[4556]: I0218 09:20:15.891681 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f16caa9f-5685-4f11-a112-65d1ed02241b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f16caa9f-5685-4f11-a112-65d1ed02241b" (UID: "f16caa9f-5685-4f11-a112-65d1ed02241b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:20:15 crc kubenswrapper[4556]: I0218 09:20:15.891970 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f16caa9f-5685-4f11-a112-65d1ed02241b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f16caa9f-5685-4f11-a112-65d1ed02241b" (UID: "f16caa9f-5685-4f11-a112-65d1ed02241b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:20:15 crc kubenswrapper[4556]: I0218 09:20:15.896782 4556 generic.go:334] "Generic (PLEG): container finished" podID="f16caa9f-5685-4f11-a112-65d1ed02241b" containerID="6d63c4d4d746825902bcb72d7e5602ab09335c7b337d7c0363dba47c1d39b045" exitCode=0 Feb 18 09:20:15 crc kubenswrapper[4556]: I0218 09:20:15.897229 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 18 09:20:15 crc kubenswrapper[4556]: I0218 09:20:15.897690 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f16caa9f-5685-4f11-a112-65d1ed02241b","Type":"ContainerDied","Data":"6d63c4d4d746825902bcb72d7e5602ab09335c7b337d7c0363dba47c1d39b045"} Feb 18 09:20:15 crc kubenswrapper[4556]: I0218 09:20:15.897806 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f16caa9f-5685-4f11-a112-65d1ed02241b","Type":"ContainerDied","Data":"a7ffd15e0245698e2eacf432969bb0cbb3efa327a6e2ed55ec4bdb14ce07ae93"} Feb 18 09:20:15 crc kubenswrapper[4556]: I0218 09:20:15.897887 4556 scope.go:117] "RemoveContainer" containerID="cb1f58413a0abdbb6acb88e4efccd9fa33ff1fcb407f6954739fcde9cbfd29bb" Feb 18 09:20:15 crc kubenswrapper[4556]: I0218 09:20:15.899230 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f16caa9f-5685-4f11-a112-65d1ed02241b-scripts" (OuterVolumeSpecName: "scripts") pod "f16caa9f-5685-4f11-a112-65d1ed02241b" (UID: "f16caa9f-5685-4f11-a112-65d1ed02241b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:20:15 crc kubenswrapper[4556]: I0218 09:20:15.900509 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f16caa9f-5685-4f11-a112-65d1ed02241b-kube-api-access-9hjm9" (OuterVolumeSpecName: "kube-api-access-9hjm9") pod "f16caa9f-5685-4f11-a112-65d1ed02241b" (UID: "f16caa9f-5685-4f11-a112-65d1ed02241b"). InnerVolumeSpecName "kube-api-access-9hjm9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:20:15 crc kubenswrapper[4556]: I0218 09:20:15.915336 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f16caa9f-5685-4f11-a112-65d1ed02241b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f16caa9f-5685-4f11-a112-65d1ed02241b" (UID: "f16caa9f-5685-4f11-a112-65d1ed02241b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:20:15 crc kubenswrapper[4556]: I0218 09:20:15.946860 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f16caa9f-5685-4f11-a112-65d1ed02241b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f16caa9f-5685-4f11-a112-65d1ed02241b" (UID: "f16caa9f-5685-4f11-a112-65d1ed02241b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:20:15 crc kubenswrapper[4556]: I0218 09:20:15.970525 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f16caa9f-5685-4f11-a112-65d1ed02241b-config-data" (OuterVolumeSpecName: "config-data") pod "f16caa9f-5685-4f11-a112-65d1ed02241b" (UID: "f16caa9f-5685-4f11-a112-65d1ed02241b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:20:15 crc kubenswrapper[4556]: I0218 09:20:15.993439 4556 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f16caa9f-5685-4f11-a112-65d1ed02241b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:20:15 crc kubenswrapper[4556]: I0218 09:20:15.993478 4556 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f16caa9f-5685-4f11-a112-65d1ed02241b-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 18 09:20:15 crc kubenswrapper[4556]: I0218 09:20:15.993487 4556 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f16caa9f-5685-4f11-a112-65d1ed02241b-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 18 09:20:15 crc kubenswrapper[4556]: I0218 09:20:15.993497 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9hjm9\" (UniqueName: \"kubernetes.io/projected/f16caa9f-5685-4f11-a112-65d1ed02241b-kube-api-access-9hjm9\") on node \"crc\" DevicePath \"\"" Feb 18 09:20:15 crc kubenswrapper[4556]: I0218 09:20:15.993511 4556 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f16caa9f-5685-4f11-a112-65d1ed02241b-config-data\") on node \"crc\" DevicePath \"\"" Feb 18 09:20:15 crc kubenswrapper[4556]: I0218 09:20:15.993520 4556 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f16caa9f-5685-4f11-a112-65d1ed02241b-scripts\") on node \"crc\" DevicePath \"\"" Feb 18 09:20:15 crc kubenswrapper[4556]: I0218 09:20:15.993528 4556 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f16caa9f-5685-4f11-a112-65d1ed02241b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 18 09:20:15 crc kubenswrapper[4556]: I0218 09:20:15.998972 4556 scope.go:117] "RemoveContainer" containerID="90309d247b8912b399ac4dd0bcf7c92ed86b4ad0f5768d5cc4813794ab056e5a" Feb 18 09:20:16 crc kubenswrapper[4556]: I0218 09:20:16.015797 4556 scope.go:117] "RemoveContainer" containerID="6d63c4d4d746825902bcb72d7e5602ab09335c7b337d7c0363dba47c1d39b045" Feb 18 09:20:16 crc kubenswrapper[4556]: I0218 09:20:16.033170 4556 scope.go:117] "RemoveContainer" containerID="9bcf089e0dd023a0eb2a47ebef6f90f371b331baf6e3080a9a09414e4ae043b7" Feb 18 09:20:16 crc kubenswrapper[4556]: I0218 09:20:16.048314 4556 scope.go:117] "RemoveContainer" containerID="cb1f58413a0abdbb6acb88e4efccd9fa33ff1fcb407f6954739fcde9cbfd29bb" Feb 18 09:20:16 crc kubenswrapper[4556]: E0218 09:20:16.048693 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb1f58413a0abdbb6acb88e4efccd9fa33ff1fcb407f6954739fcde9cbfd29bb\": container with ID starting with cb1f58413a0abdbb6acb88e4efccd9fa33ff1fcb407f6954739fcde9cbfd29bb not found: ID does not exist" containerID="cb1f58413a0abdbb6acb88e4efccd9fa33ff1fcb407f6954739fcde9cbfd29bb" Feb 18 09:20:16 crc kubenswrapper[4556]: I0218 09:20:16.048738 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb1f58413a0abdbb6acb88e4efccd9fa33ff1fcb407f6954739fcde9cbfd29bb"} err="failed to get container status \"cb1f58413a0abdbb6acb88e4efccd9fa33ff1fcb407f6954739fcde9cbfd29bb\": rpc error: code = NotFound desc = could not find container \"cb1f58413a0abdbb6acb88e4efccd9fa33ff1fcb407f6954739fcde9cbfd29bb\": container with ID starting with cb1f58413a0abdbb6acb88e4efccd9fa33ff1fcb407f6954739fcde9cbfd29bb not found: ID does not exist" Feb 18 09:20:16 crc kubenswrapper[4556]: I0218 09:20:16.048764 4556 scope.go:117] "RemoveContainer" containerID="90309d247b8912b399ac4dd0bcf7c92ed86b4ad0f5768d5cc4813794ab056e5a" Feb 18 09:20:16 crc kubenswrapper[4556]: E0218 09:20:16.049048 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90309d247b8912b399ac4dd0bcf7c92ed86b4ad0f5768d5cc4813794ab056e5a\": container with ID starting with 90309d247b8912b399ac4dd0bcf7c92ed86b4ad0f5768d5cc4813794ab056e5a not found: ID does not exist" containerID="90309d247b8912b399ac4dd0bcf7c92ed86b4ad0f5768d5cc4813794ab056e5a" Feb 18 09:20:16 crc kubenswrapper[4556]: I0218 09:20:16.049084 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90309d247b8912b399ac4dd0bcf7c92ed86b4ad0f5768d5cc4813794ab056e5a"} err="failed to get container status \"90309d247b8912b399ac4dd0bcf7c92ed86b4ad0f5768d5cc4813794ab056e5a\": rpc error: code = NotFound desc = could not find container \"90309d247b8912b399ac4dd0bcf7c92ed86b4ad0f5768d5cc4813794ab056e5a\": container with ID starting with 90309d247b8912b399ac4dd0bcf7c92ed86b4ad0f5768d5cc4813794ab056e5a not found: ID does not exist" Feb 18 09:20:16 crc kubenswrapper[4556]: I0218 09:20:16.049112 4556 scope.go:117] "RemoveContainer" containerID="6d63c4d4d746825902bcb72d7e5602ab09335c7b337d7c0363dba47c1d39b045" Feb 18 09:20:16 crc kubenswrapper[4556]: E0218 09:20:16.049477 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d63c4d4d746825902bcb72d7e5602ab09335c7b337d7c0363dba47c1d39b045\": container with ID starting with 6d63c4d4d746825902bcb72d7e5602ab09335c7b337d7c0363dba47c1d39b045 not found: ID does not exist" containerID="6d63c4d4d746825902bcb72d7e5602ab09335c7b337d7c0363dba47c1d39b045" Feb 18 09:20:16 crc kubenswrapper[4556]: I0218 09:20:16.049509 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d63c4d4d746825902bcb72d7e5602ab09335c7b337d7c0363dba47c1d39b045"} err="failed to get container status \"6d63c4d4d746825902bcb72d7e5602ab09335c7b337d7c0363dba47c1d39b045\": rpc error: code = NotFound desc = could not find container \"6d63c4d4d746825902bcb72d7e5602ab09335c7b337d7c0363dba47c1d39b045\": container with ID starting with 6d63c4d4d746825902bcb72d7e5602ab09335c7b337d7c0363dba47c1d39b045 not found: ID does not exist" Feb 18 09:20:16 crc kubenswrapper[4556]: I0218 09:20:16.049532 4556 scope.go:117] "RemoveContainer" containerID="9bcf089e0dd023a0eb2a47ebef6f90f371b331baf6e3080a9a09414e4ae043b7" Feb 18 09:20:16 crc kubenswrapper[4556]: E0218 09:20:16.049806 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9bcf089e0dd023a0eb2a47ebef6f90f371b331baf6e3080a9a09414e4ae043b7\": container with ID starting with 9bcf089e0dd023a0eb2a47ebef6f90f371b331baf6e3080a9a09414e4ae043b7 not found: ID does not exist" containerID="9bcf089e0dd023a0eb2a47ebef6f90f371b331baf6e3080a9a09414e4ae043b7" Feb 18 09:20:16 crc kubenswrapper[4556]: I0218 09:20:16.049836 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9bcf089e0dd023a0eb2a47ebef6f90f371b331baf6e3080a9a09414e4ae043b7"} err="failed to get container status \"9bcf089e0dd023a0eb2a47ebef6f90f371b331baf6e3080a9a09414e4ae043b7\": rpc error: code = NotFound desc = could not find container \"9bcf089e0dd023a0eb2a47ebef6f90f371b331baf6e3080a9a09414e4ae043b7\": container with ID starting with 9bcf089e0dd023a0eb2a47ebef6f90f371b331baf6e3080a9a09414e4ae043b7 not found: ID does not exist" Feb 18 09:20:16 crc kubenswrapper[4556]: I0218 09:20:16.232428 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 18 09:20:16 crc kubenswrapper[4556]: I0218 09:20:16.239169 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 18 09:20:16 crc kubenswrapper[4556]: I0218 09:20:16.243330 4556 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="4c7a2344-f7ad-40c2-84e9-60abcf73571a" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.196:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 18 09:20:16 crc kubenswrapper[4556]: I0218 09:20:16.243376 4556 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="4c7a2344-f7ad-40c2-84e9-60abcf73571a" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.196:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 18 09:20:16 crc kubenswrapper[4556]: I0218 09:20:16.261685 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 18 09:20:16 crc kubenswrapper[4556]: E0218 09:20:16.262069 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f16caa9f-5685-4f11-a112-65d1ed02241b" containerName="ceilometer-central-agent" Feb 18 09:20:16 crc kubenswrapper[4556]: I0218 09:20:16.262090 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="f16caa9f-5685-4f11-a112-65d1ed02241b" containerName="ceilometer-central-agent" Feb 18 09:20:16 crc kubenswrapper[4556]: E0218 09:20:16.262098 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f16caa9f-5685-4f11-a112-65d1ed02241b" containerName="ceilometer-notification-agent" Feb 18 09:20:16 crc kubenswrapper[4556]: I0218 09:20:16.262105 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="f16caa9f-5685-4f11-a112-65d1ed02241b" containerName="ceilometer-notification-agent" Feb 18 09:20:16 crc kubenswrapper[4556]: E0218 09:20:16.262112 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f16caa9f-5685-4f11-a112-65d1ed02241b" containerName="sg-core" Feb 18 09:20:16 crc kubenswrapper[4556]: I0218 09:20:16.262118 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="f16caa9f-5685-4f11-a112-65d1ed02241b" containerName="sg-core" Feb 18 09:20:16 crc kubenswrapper[4556]: E0218 09:20:16.262146 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f16caa9f-5685-4f11-a112-65d1ed02241b" containerName="proxy-httpd" Feb 18 09:20:16 crc kubenswrapper[4556]: I0218 09:20:16.262174 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="f16caa9f-5685-4f11-a112-65d1ed02241b" containerName="proxy-httpd" Feb 18 09:20:16 crc kubenswrapper[4556]: I0218 09:20:16.262374 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="f16caa9f-5685-4f11-a112-65d1ed02241b" containerName="proxy-httpd" Feb 18 09:20:16 crc kubenswrapper[4556]: I0218 09:20:16.262392 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="f16caa9f-5685-4f11-a112-65d1ed02241b" containerName="sg-core" Feb 18 09:20:16 crc kubenswrapper[4556]: I0218 09:20:16.262399 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="f16caa9f-5685-4f11-a112-65d1ed02241b" containerName="ceilometer-central-agent" Feb 18 09:20:16 crc kubenswrapper[4556]: I0218 09:20:16.262419 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="f16caa9f-5685-4f11-a112-65d1ed02241b" containerName="ceilometer-notification-agent" Feb 18 09:20:16 crc kubenswrapper[4556]: I0218 09:20:16.263849 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 18 09:20:16 crc kubenswrapper[4556]: I0218 09:20:16.270013 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 18 09:20:16 crc kubenswrapper[4556]: I0218 09:20:16.270243 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Feb 18 09:20:16 crc kubenswrapper[4556]: I0218 09:20:16.270421 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 18 09:20:16 crc kubenswrapper[4556]: I0218 09:20:16.277551 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 18 09:20:16 crc kubenswrapper[4556]: I0218 09:20:16.398645 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/855cf21a-1bf7-49e5-b27c-47a8fe960e42-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"855cf21a-1bf7-49e5-b27c-47a8fe960e42\") " pod="openstack/ceilometer-0" Feb 18 09:20:16 crc kubenswrapper[4556]: I0218 09:20:16.398797 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/855cf21a-1bf7-49e5-b27c-47a8fe960e42-scripts\") pod \"ceilometer-0\" (UID: \"855cf21a-1bf7-49e5-b27c-47a8fe960e42\") " pod="openstack/ceilometer-0" Feb 18 09:20:16 crc kubenswrapper[4556]: I0218 09:20:16.398911 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/855cf21a-1bf7-49e5-b27c-47a8fe960e42-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"855cf21a-1bf7-49e5-b27c-47a8fe960e42\") " pod="openstack/ceilometer-0" Feb 18 09:20:16 crc kubenswrapper[4556]: I0218 09:20:16.398979 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/855cf21a-1bf7-49e5-b27c-47a8fe960e42-run-httpd\") pod \"ceilometer-0\" (UID: \"855cf21a-1bf7-49e5-b27c-47a8fe960e42\") " pod="openstack/ceilometer-0" Feb 18 09:20:16 crc kubenswrapper[4556]: I0218 09:20:16.399037 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/855cf21a-1bf7-49e5-b27c-47a8fe960e42-config-data\") pod \"ceilometer-0\" (UID: \"855cf21a-1bf7-49e5-b27c-47a8fe960e42\") " pod="openstack/ceilometer-0" Feb 18 09:20:16 crc kubenswrapper[4556]: I0218 09:20:16.409576 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/855cf21a-1bf7-49e5-b27c-47a8fe960e42-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"855cf21a-1bf7-49e5-b27c-47a8fe960e42\") " pod="openstack/ceilometer-0" Feb 18 09:20:16 crc kubenswrapper[4556]: I0218 09:20:16.409666 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/855cf21a-1bf7-49e5-b27c-47a8fe960e42-log-httpd\") pod \"ceilometer-0\" (UID: \"855cf21a-1bf7-49e5-b27c-47a8fe960e42\") " pod="openstack/ceilometer-0" Feb 18 09:20:16 crc kubenswrapper[4556]: I0218 09:20:16.409774 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jb5mx\" (UniqueName: \"kubernetes.io/projected/855cf21a-1bf7-49e5-b27c-47a8fe960e42-kube-api-access-jb5mx\") pod \"ceilometer-0\" (UID: \"855cf21a-1bf7-49e5-b27c-47a8fe960e42\") " pod="openstack/ceilometer-0" Feb 18 09:20:16 crc kubenswrapper[4556]: I0218 09:20:16.511820 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/855cf21a-1bf7-49e5-b27c-47a8fe960e42-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"855cf21a-1bf7-49e5-b27c-47a8fe960e42\") " pod="openstack/ceilometer-0" Feb 18 09:20:16 crc kubenswrapper[4556]: I0218 09:20:16.511926 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/855cf21a-1bf7-49e5-b27c-47a8fe960e42-scripts\") pod \"ceilometer-0\" (UID: \"855cf21a-1bf7-49e5-b27c-47a8fe960e42\") " pod="openstack/ceilometer-0" Feb 18 09:20:16 crc kubenswrapper[4556]: I0218 09:20:16.512013 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/855cf21a-1bf7-49e5-b27c-47a8fe960e42-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"855cf21a-1bf7-49e5-b27c-47a8fe960e42\") " pod="openstack/ceilometer-0" Feb 18 09:20:16 crc kubenswrapper[4556]: I0218 09:20:16.512047 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/855cf21a-1bf7-49e5-b27c-47a8fe960e42-run-httpd\") pod \"ceilometer-0\" (UID: \"855cf21a-1bf7-49e5-b27c-47a8fe960e42\") " pod="openstack/ceilometer-0" Feb 18 09:20:16 crc kubenswrapper[4556]: I0218 09:20:16.512084 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/855cf21a-1bf7-49e5-b27c-47a8fe960e42-config-data\") pod \"ceilometer-0\" (UID: \"855cf21a-1bf7-49e5-b27c-47a8fe960e42\") " pod="openstack/ceilometer-0" Feb 18 09:20:16 crc kubenswrapper[4556]: I0218 09:20:16.512135 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/855cf21a-1bf7-49e5-b27c-47a8fe960e42-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"855cf21a-1bf7-49e5-b27c-47a8fe960e42\") " pod="openstack/ceilometer-0" Feb 18 09:20:16 crc kubenswrapper[4556]: I0218 09:20:16.512174 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/855cf21a-1bf7-49e5-b27c-47a8fe960e42-log-httpd\") pod \"ceilometer-0\" (UID: \"855cf21a-1bf7-49e5-b27c-47a8fe960e42\") " pod="openstack/ceilometer-0" Feb 18 09:20:16 crc kubenswrapper[4556]: I0218 09:20:16.512227 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jb5mx\" (UniqueName: \"kubernetes.io/projected/855cf21a-1bf7-49e5-b27c-47a8fe960e42-kube-api-access-jb5mx\") pod \"ceilometer-0\" (UID: \"855cf21a-1bf7-49e5-b27c-47a8fe960e42\") " pod="openstack/ceilometer-0" Feb 18 09:20:16 crc kubenswrapper[4556]: I0218 09:20:16.512712 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/855cf21a-1bf7-49e5-b27c-47a8fe960e42-run-httpd\") pod \"ceilometer-0\" (UID: \"855cf21a-1bf7-49e5-b27c-47a8fe960e42\") " pod="openstack/ceilometer-0" Feb 18 09:20:16 crc kubenswrapper[4556]: I0218 09:20:16.512801 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/855cf21a-1bf7-49e5-b27c-47a8fe960e42-log-httpd\") pod \"ceilometer-0\" (UID: \"855cf21a-1bf7-49e5-b27c-47a8fe960e42\") " pod="openstack/ceilometer-0" Feb 18 09:20:16 crc kubenswrapper[4556]: I0218 09:20:16.515491 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/855cf21a-1bf7-49e5-b27c-47a8fe960e42-scripts\") pod \"ceilometer-0\" (UID: \"855cf21a-1bf7-49e5-b27c-47a8fe960e42\") " pod="openstack/ceilometer-0" Feb 18 09:20:16 crc kubenswrapper[4556]: I0218 09:20:16.515585 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/855cf21a-1bf7-49e5-b27c-47a8fe960e42-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"855cf21a-1bf7-49e5-b27c-47a8fe960e42\") " pod="openstack/ceilometer-0" Feb 18 09:20:16 crc kubenswrapper[4556]: I0218 09:20:16.516419 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/855cf21a-1bf7-49e5-b27c-47a8fe960e42-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"855cf21a-1bf7-49e5-b27c-47a8fe960e42\") " pod="openstack/ceilometer-0" Feb 18 09:20:16 crc kubenswrapper[4556]: I0218 09:20:16.517123 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/855cf21a-1bf7-49e5-b27c-47a8fe960e42-config-data\") pod \"ceilometer-0\" (UID: \"855cf21a-1bf7-49e5-b27c-47a8fe960e42\") " pod="openstack/ceilometer-0" Feb 18 09:20:16 crc kubenswrapper[4556]: I0218 09:20:16.517501 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/855cf21a-1bf7-49e5-b27c-47a8fe960e42-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"855cf21a-1bf7-49e5-b27c-47a8fe960e42\") " pod="openstack/ceilometer-0" Feb 18 09:20:16 crc kubenswrapper[4556]: I0218 09:20:16.527100 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jb5mx\" (UniqueName: \"kubernetes.io/projected/855cf21a-1bf7-49e5-b27c-47a8fe960e42-kube-api-access-jb5mx\") pod \"ceilometer-0\" (UID: \"855cf21a-1bf7-49e5-b27c-47a8fe960e42\") " pod="openstack/ceilometer-0" Feb 18 09:20:16 crc kubenswrapper[4556]: I0218 09:20:16.580285 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 18 09:20:17 crc kubenswrapper[4556]: I0218 09:20:17.006795 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 18 09:20:17 crc kubenswrapper[4556]: I0218 09:20:17.310912 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f16caa9f-5685-4f11-a112-65d1ed02241b" path="/var/lib/kubelet/pods/f16caa9f-5685-4f11-a112-65d1ed02241b/volumes" Feb 18 09:20:17 crc kubenswrapper[4556]: I0218 09:20:17.918365 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"855cf21a-1bf7-49e5-b27c-47a8fe960e42","Type":"ContainerStarted","Data":"d6dd7effa6f1baf748f47faedfdd57c3c9527b8f05be61c68495cd18a10f263f"} Feb 18 09:20:17 crc kubenswrapper[4556]: I0218 09:20:17.918739 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"855cf21a-1bf7-49e5-b27c-47a8fe960e42","Type":"ContainerStarted","Data":"938c2e818c458c5ceb9c1e3558089112f5524a6763219b3a89d978af5746b2d2"} Feb 18 09:20:18 crc kubenswrapper[4556]: I0218 09:20:18.928355 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"855cf21a-1bf7-49e5-b27c-47a8fe960e42","Type":"ContainerStarted","Data":"0ec74627d62a117c272bed8fdd57b93e3f5a88c7e79f7fcd985279ee964840f1"} Feb 18 09:20:19 crc kubenswrapper[4556]: I0218 09:20:19.938904 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"855cf21a-1bf7-49e5-b27c-47a8fe960e42","Type":"ContainerStarted","Data":"7bb494813010282cbf0390a221a66dd0231c3ef2a0bf1812709276fa3689354e"} Feb 18 09:20:20 crc kubenswrapper[4556]: I0218 09:20:20.949195 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"855cf21a-1bf7-49e5-b27c-47a8fe960e42","Type":"ContainerStarted","Data":"6a816d7032e88656777541c6d25ea7a7a7c708781de32ccd8cdda1f02e90faaa"} Feb 18 09:20:20 crc kubenswrapper[4556]: I0218 09:20:20.950012 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 18 09:20:20 crc kubenswrapper[4556]: I0218 09:20:20.970298 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.549961811 podStartE2EDuration="4.970276906s" podCreationTimestamp="2026-02-18 09:20:16 +0000 UTC" firstStartedPulling="2026-02-18 09:20:17.012353253 +0000 UTC m=+974.029314234" lastFinishedPulling="2026-02-18 09:20:20.432668349 +0000 UTC m=+977.449629329" observedRunningTime="2026-02-18 09:20:20.965811059 +0000 UTC m=+977.982772039" watchObservedRunningTime="2026-02-18 09:20:20.970276906 +0000 UTC m=+977.987237886" Feb 18 09:20:22 crc kubenswrapper[4556]: I0218 09:20:22.972303 4556 generic.go:334] "Generic (PLEG): container finished" podID="2330f995-37c7-4b3e-a465-4e682349346f" containerID="f6a2777c15f011057e9eb148ba4d170bda4c52701b24e2f8ccd0d1db588afa4f" exitCode=137 Feb 18 09:20:22 crc kubenswrapper[4556]: I0218 09:20:22.972383 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2330f995-37c7-4b3e-a465-4e682349346f","Type":"ContainerDied","Data":"f6a2777c15f011057e9eb148ba4d170bda4c52701b24e2f8ccd0d1db588afa4f"} Feb 18 09:20:22 crc kubenswrapper[4556]: I0218 09:20:22.972802 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2330f995-37c7-4b3e-a465-4e682349346f","Type":"ContainerDied","Data":"408bda8acc5dc2e8cf10b40b04b6699c74680ecfb14fa057bd58ab263a48b395"} Feb 18 09:20:22 crc kubenswrapper[4556]: I0218 09:20:22.972829 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="408bda8acc5dc2e8cf10b40b04b6699c74680ecfb14fa057bd58ab263a48b395" Feb 18 09:20:22 crc kubenswrapper[4556]: I0218 09:20:22.975397 4556 generic.go:334] "Generic (PLEG): container finished" podID="4ea5a5ac-537e-40d2-a5dc-4cc2314bf097" containerID="5b0edfc3e404d52fc9c324525b661c589d816bf646b1ce5470138f85ad4b763a" exitCode=137 Feb 18 09:20:22 crc kubenswrapper[4556]: I0218 09:20:22.975453 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"4ea5a5ac-537e-40d2-a5dc-4cc2314bf097","Type":"ContainerDied","Data":"5b0edfc3e404d52fc9c324525b661c589d816bf646b1ce5470138f85ad4b763a"} Feb 18 09:20:22 crc kubenswrapper[4556]: I0218 09:20:22.975490 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"4ea5a5ac-537e-40d2-a5dc-4cc2314bf097","Type":"ContainerDied","Data":"cb28819905c553b8f23d96fcc570350621ef2ca814d42ff1c3f0ef0768480d07"} Feb 18 09:20:22 crc kubenswrapper[4556]: I0218 09:20:22.975503 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb28819905c553b8f23d96fcc570350621ef2ca814d42ff1c3f0ef0768480d07" Feb 18 09:20:22 crc kubenswrapper[4556]: I0218 09:20:22.995789 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 18 09:20:23 crc kubenswrapper[4556]: I0218 09:20:22.999943 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 18 09:20:23 crc kubenswrapper[4556]: I0218 09:20:23.073327 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ea5a5ac-537e-40d2-a5dc-4cc2314bf097-combined-ca-bundle\") pod \"4ea5a5ac-537e-40d2-a5dc-4cc2314bf097\" (UID: \"4ea5a5ac-537e-40d2-a5dc-4cc2314bf097\") " Feb 18 09:20:23 crc kubenswrapper[4556]: I0218 09:20:23.073474 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2330f995-37c7-4b3e-a465-4e682349346f-logs\") pod \"2330f995-37c7-4b3e-a465-4e682349346f\" (UID: \"2330f995-37c7-4b3e-a465-4e682349346f\") " Feb 18 09:20:23 crc kubenswrapper[4556]: I0218 09:20:23.073549 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2330f995-37c7-4b3e-a465-4e682349346f-combined-ca-bundle\") pod \"2330f995-37c7-4b3e-a465-4e682349346f\" (UID: \"2330f995-37c7-4b3e-a465-4e682349346f\") " Feb 18 09:20:23 crc kubenswrapper[4556]: I0218 09:20:23.073595 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ea5a5ac-537e-40d2-a5dc-4cc2314bf097-config-data\") pod \"4ea5a5ac-537e-40d2-a5dc-4cc2314bf097\" (UID: \"4ea5a5ac-537e-40d2-a5dc-4cc2314bf097\") " Feb 18 09:20:23 crc kubenswrapper[4556]: I0218 09:20:23.073621 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b6d6r\" (UniqueName: \"kubernetes.io/projected/2330f995-37c7-4b3e-a465-4e682349346f-kube-api-access-b6d6r\") pod \"2330f995-37c7-4b3e-a465-4e682349346f\" (UID: \"2330f995-37c7-4b3e-a465-4e682349346f\") " Feb 18 09:20:23 crc kubenswrapper[4556]: I0218 09:20:23.073664 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2330f995-37c7-4b3e-a465-4e682349346f-config-data\") pod \"2330f995-37c7-4b3e-a465-4e682349346f\" (UID: \"2330f995-37c7-4b3e-a465-4e682349346f\") " Feb 18 09:20:23 crc kubenswrapper[4556]: I0218 09:20:23.073781 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x8jpp\" (UniqueName: \"kubernetes.io/projected/4ea5a5ac-537e-40d2-a5dc-4cc2314bf097-kube-api-access-x8jpp\") pod \"4ea5a5ac-537e-40d2-a5dc-4cc2314bf097\" (UID: \"4ea5a5ac-537e-40d2-a5dc-4cc2314bf097\") " Feb 18 09:20:23 crc kubenswrapper[4556]: I0218 09:20:23.074591 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2330f995-37c7-4b3e-a465-4e682349346f-logs" (OuterVolumeSpecName: "logs") pod "2330f995-37c7-4b3e-a465-4e682349346f" (UID: "2330f995-37c7-4b3e-a465-4e682349346f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:20:23 crc kubenswrapper[4556]: I0218 09:20:23.091374 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2330f995-37c7-4b3e-a465-4e682349346f-kube-api-access-b6d6r" (OuterVolumeSpecName: "kube-api-access-b6d6r") pod "2330f995-37c7-4b3e-a465-4e682349346f" (UID: "2330f995-37c7-4b3e-a465-4e682349346f"). InnerVolumeSpecName "kube-api-access-b6d6r". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:20:23 crc kubenswrapper[4556]: I0218 09:20:23.094020 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ea5a5ac-537e-40d2-a5dc-4cc2314bf097-kube-api-access-x8jpp" (OuterVolumeSpecName: "kube-api-access-x8jpp") pod "4ea5a5ac-537e-40d2-a5dc-4cc2314bf097" (UID: "4ea5a5ac-537e-40d2-a5dc-4cc2314bf097"). InnerVolumeSpecName "kube-api-access-x8jpp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:20:23 crc kubenswrapper[4556]: I0218 09:20:23.101637 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2330f995-37c7-4b3e-a465-4e682349346f-config-data" (OuterVolumeSpecName: "config-data") pod "2330f995-37c7-4b3e-a465-4e682349346f" (UID: "2330f995-37c7-4b3e-a465-4e682349346f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:20:23 crc kubenswrapper[4556]: I0218 09:20:23.102906 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ea5a5ac-537e-40d2-a5dc-4cc2314bf097-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4ea5a5ac-537e-40d2-a5dc-4cc2314bf097" (UID: "4ea5a5ac-537e-40d2-a5dc-4cc2314bf097"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:20:23 crc kubenswrapper[4556]: I0218 09:20:23.103442 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2330f995-37c7-4b3e-a465-4e682349346f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2330f995-37c7-4b3e-a465-4e682349346f" (UID: "2330f995-37c7-4b3e-a465-4e682349346f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:20:23 crc kubenswrapper[4556]: I0218 09:20:23.104224 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ea5a5ac-537e-40d2-a5dc-4cc2314bf097-config-data" (OuterVolumeSpecName: "config-data") pod "4ea5a5ac-537e-40d2-a5dc-4cc2314bf097" (UID: "4ea5a5ac-537e-40d2-a5dc-4cc2314bf097"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:20:23 crc kubenswrapper[4556]: I0218 09:20:23.175920 4556 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2330f995-37c7-4b3e-a465-4e682349346f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:20:23 crc kubenswrapper[4556]: I0218 09:20:23.175962 4556 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ea5a5ac-537e-40d2-a5dc-4cc2314bf097-config-data\") on node \"crc\" DevicePath \"\"" Feb 18 09:20:23 crc kubenswrapper[4556]: I0218 09:20:23.175974 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b6d6r\" (UniqueName: \"kubernetes.io/projected/2330f995-37c7-4b3e-a465-4e682349346f-kube-api-access-b6d6r\") on node \"crc\" DevicePath \"\"" Feb 18 09:20:23 crc kubenswrapper[4556]: I0218 09:20:23.175989 4556 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2330f995-37c7-4b3e-a465-4e682349346f-config-data\") on node \"crc\" DevicePath \"\"" Feb 18 09:20:23 crc kubenswrapper[4556]: I0218 09:20:23.176015 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x8jpp\" (UniqueName: \"kubernetes.io/projected/4ea5a5ac-537e-40d2-a5dc-4cc2314bf097-kube-api-access-x8jpp\") on node \"crc\" DevicePath \"\"" Feb 18 09:20:23 crc kubenswrapper[4556]: I0218 09:20:23.176025 4556 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ea5a5ac-537e-40d2-a5dc-4cc2314bf097-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:20:23 crc kubenswrapper[4556]: I0218 09:20:23.176034 4556 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2330f995-37c7-4b3e-a465-4e682349346f-logs\") on node \"crc\" DevicePath \"\"" Feb 18 09:20:23 crc kubenswrapper[4556]: I0218 09:20:23.250020 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Feb 18 09:20:23 crc kubenswrapper[4556]: I0218 09:20:23.991927 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 18 09:20:23 crc kubenswrapper[4556]: I0218 09:20:23.992224 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 18 09:20:24 crc kubenswrapper[4556]: I0218 09:20:24.018282 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 18 09:20:24 crc kubenswrapper[4556]: I0218 09:20:24.026061 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Feb 18 09:20:24 crc kubenswrapper[4556]: I0218 09:20:24.035858 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 18 09:20:24 crc kubenswrapper[4556]: I0218 09:20:24.046699 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 18 09:20:24 crc kubenswrapper[4556]: I0218 09:20:24.057371 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 18 09:20:24 crc kubenswrapper[4556]: E0218 09:20:24.057836 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ea5a5ac-537e-40d2-a5dc-4cc2314bf097" containerName="nova-cell1-novncproxy-novncproxy" Feb 18 09:20:24 crc kubenswrapper[4556]: I0218 09:20:24.057850 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ea5a5ac-537e-40d2-a5dc-4cc2314bf097" containerName="nova-cell1-novncproxy-novncproxy" Feb 18 09:20:24 crc kubenswrapper[4556]: E0218 09:20:24.057871 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2330f995-37c7-4b3e-a465-4e682349346f" containerName="nova-metadata-log" Feb 18 09:20:24 crc kubenswrapper[4556]: I0218 09:20:24.057877 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="2330f995-37c7-4b3e-a465-4e682349346f" containerName="nova-metadata-log" Feb 18 09:20:24 crc kubenswrapper[4556]: E0218 09:20:24.057889 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2330f995-37c7-4b3e-a465-4e682349346f" containerName="nova-metadata-metadata" Feb 18 09:20:24 crc kubenswrapper[4556]: I0218 09:20:24.057894 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="2330f995-37c7-4b3e-a465-4e682349346f" containerName="nova-metadata-metadata" Feb 18 09:20:24 crc kubenswrapper[4556]: I0218 09:20:24.058138 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="2330f995-37c7-4b3e-a465-4e682349346f" containerName="nova-metadata-log" Feb 18 09:20:24 crc kubenswrapper[4556]: I0218 09:20:24.058168 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ea5a5ac-537e-40d2-a5dc-4cc2314bf097" containerName="nova-cell1-novncproxy-novncproxy" Feb 18 09:20:24 crc kubenswrapper[4556]: I0218 09:20:24.058179 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="2330f995-37c7-4b3e-a465-4e682349346f" containerName="nova-metadata-metadata" Feb 18 09:20:24 crc kubenswrapper[4556]: I0218 09:20:24.059175 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 18 09:20:24 crc kubenswrapper[4556]: I0218 09:20:24.062860 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 18 09:20:24 crc kubenswrapper[4556]: I0218 09:20:24.063103 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Feb 18 09:20:24 crc kubenswrapper[4556]: I0218 09:20:24.069697 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 18 09:20:24 crc kubenswrapper[4556]: I0218 09:20:24.073304 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 18 09:20:24 crc kubenswrapper[4556]: I0218 09:20:24.074596 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 18 09:20:24 crc kubenswrapper[4556]: I0218 09:20:24.077015 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Feb 18 09:20:24 crc kubenswrapper[4556]: I0218 09:20:24.078236 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 18 09:20:24 crc kubenswrapper[4556]: I0218 09:20:24.078941 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Feb 18 09:20:24 crc kubenswrapper[4556]: I0218 09:20:24.079117 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Feb 18 09:20:24 crc kubenswrapper[4556]: I0218 09:20:24.093756 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f3e12da-26fe-4d2a-ae62-9a3e90c5b752-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1f3e12da-26fe-4d2a-ae62-9a3e90c5b752\") " pod="openstack/nova-metadata-0" Feb 18 09:20:24 crc kubenswrapper[4556]: I0218 09:20:24.093798 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f3e12da-26fe-4d2a-ae62-9a3e90c5b752-config-data\") pod \"nova-metadata-0\" (UID: \"1f3e12da-26fe-4d2a-ae62-9a3e90c5b752\") " pod="openstack/nova-metadata-0" Feb 18 09:20:24 crc kubenswrapper[4556]: I0218 09:20:24.094435 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f3e12da-26fe-4d2a-ae62-9a3e90c5b752-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1f3e12da-26fe-4d2a-ae62-9a3e90c5b752\") " pod="openstack/nova-metadata-0" Feb 18 09:20:24 crc kubenswrapper[4556]: I0218 09:20:24.094480 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jczm\" (UniqueName: \"kubernetes.io/projected/1f3e12da-26fe-4d2a-ae62-9a3e90c5b752-kube-api-access-5jczm\") pod \"nova-metadata-0\" (UID: \"1f3e12da-26fe-4d2a-ae62-9a3e90c5b752\") " pod="openstack/nova-metadata-0" Feb 18 09:20:24 crc kubenswrapper[4556]: I0218 09:20:24.094504 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f3e12da-26fe-4d2a-ae62-9a3e90c5b752-logs\") pod \"nova-metadata-0\" (UID: \"1f3e12da-26fe-4d2a-ae62-9a3e90c5b752\") " pod="openstack/nova-metadata-0" Feb 18 09:20:24 crc kubenswrapper[4556]: I0218 09:20:24.197116 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1c20375-9fcb-4886-a91c-3dd36651532a-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c1c20375-9fcb-4886-a91c-3dd36651532a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 18 09:20:24 crc kubenswrapper[4556]: I0218 09:20:24.197235 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1c20375-9fcb-4886-a91c-3dd36651532a-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"c1c20375-9fcb-4886-a91c-3dd36651532a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 18 09:20:24 crc kubenswrapper[4556]: I0218 09:20:24.197298 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52wpl\" (UniqueName: \"kubernetes.io/projected/c1c20375-9fcb-4886-a91c-3dd36651532a-kube-api-access-52wpl\") pod \"nova-cell1-novncproxy-0\" (UID: \"c1c20375-9fcb-4886-a91c-3dd36651532a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 18 09:20:24 crc kubenswrapper[4556]: I0218 09:20:24.197385 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1c20375-9fcb-4886-a91c-3dd36651532a-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c1c20375-9fcb-4886-a91c-3dd36651532a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 18 09:20:24 crc kubenswrapper[4556]: I0218 09:20:24.197417 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1c20375-9fcb-4886-a91c-3dd36651532a-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"c1c20375-9fcb-4886-a91c-3dd36651532a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 18 09:20:24 crc kubenswrapper[4556]: I0218 09:20:24.197518 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f3e12da-26fe-4d2a-ae62-9a3e90c5b752-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1f3e12da-26fe-4d2a-ae62-9a3e90c5b752\") " pod="openstack/nova-metadata-0" Feb 18 09:20:24 crc kubenswrapper[4556]: I0218 09:20:24.197570 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f3e12da-26fe-4d2a-ae62-9a3e90c5b752-config-data\") pod \"nova-metadata-0\" (UID: \"1f3e12da-26fe-4d2a-ae62-9a3e90c5b752\") " pod="openstack/nova-metadata-0" Feb 18 09:20:24 crc kubenswrapper[4556]: I0218 09:20:24.197629 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f3e12da-26fe-4d2a-ae62-9a3e90c5b752-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1f3e12da-26fe-4d2a-ae62-9a3e90c5b752\") " pod="openstack/nova-metadata-0" Feb 18 09:20:24 crc kubenswrapper[4556]: I0218 09:20:24.197665 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jczm\" (UniqueName: \"kubernetes.io/projected/1f3e12da-26fe-4d2a-ae62-9a3e90c5b752-kube-api-access-5jczm\") pod \"nova-metadata-0\" (UID: \"1f3e12da-26fe-4d2a-ae62-9a3e90c5b752\") " pod="openstack/nova-metadata-0" Feb 18 09:20:24 crc kubenswrapper[4556]: I0218 09:20:24.197682 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f3e12da-26fe-4d2a-ae62-9a3e90c5b752-logs\") pod \"nova-metadata-0\" (UID: \"1f3e12da-26fe-4d2a-ae62-9a3e90c5b752\") " pod="openstack/nova-metadata-0" Feb 18 09:20:24 crc kubenswrapper[4556]: I0218 09:20:24.198113 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f3e12da-26fe-4d2a-ae62-9a3e90c5b752-logs\") pod \"nova-metadata-0\" (UID: \"1f3e12da-26fe-4d2a-ae62-9a3e90c5b752\") " pod="openstack/nova-metadata-0" Feb 18 09:20:24 crc kubenswrapper[4556]: I0218 09:20:24.204028 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f3e12da-26fe-4d2a-ae62-9a3e90c5b752-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1f3e12da-26fe-4d2a-ae62-9a3e90c5b752\") " pod="openstack/nova-metadata-0" Feb 18 09:20:24 crc kubenswrapper[4556]: I0218 09:20:24.204090 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f3e12da-26fe-4d2a-ae62-9a3e90c5b752-config-data\") pod \"nova-metadata-0\" (UID: \"1f3e12da-26fe-4d2a-ae62-9a3e90c5b752\") " pod="openstack/nova-metadata-0" Feb 18 09:20:24 crc kubenswrapper[4556]: I0218 09:20:24.209823 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f3e12da-26fe-4d2a-ae62-9a3e90c5b752-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1f3e12da-26fe-4d2a-ae62-9a3e90c5b752\") " pod="openstack/nova-metadata-0" Feb 18 09:20:24 crc kubenswrapper[4556]: I0218 09:20:24.215952 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jczm\" (UniqueName: \"kubernetes.io/projected/1f3e12da-26fe-4d2a-ae62-9a3e90c5b752-kube-api-access-5jczm\") pod \"nova-metadata-0\" (UID: \"1f3e12da-26fe-4d2a-ae62-9a3e90c5b752\") " pod="openstack/nova-metadata-0" Feb 18 09:20:24 crc kubenswrapper[4556]: I0218 09:20:24.298962 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52wpl\" (UniqueName: \"kubernetes.io/projected/c1c20375-9fcb-4886-a91c-3dd36651532a-kube-api-access-52wpl\") pod \"nova-cell1-novncproxy-0\" (UID: \"c1c20375-9fcb-4886-a91c-3dd36651532a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 18 09:20:24 crc kubenswrapper[4556]: I0218 09:20:24.299046 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1c20375-9fcb-4886-a91c-3dd36651532a-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c1c20375-9fcb-4886-a91c-3dd36651532a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 18 09:20:24 crc kubenswrapper[4556]: I0218 09:20:24.299071 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1c20375-9fcb-4886-a91c-3dd36651532a-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"c1c20375-9fcb-4886-a91c-3dd36651532a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 18 09:20:24 crc kubenswrapper[4556]: I0218 09:20:24.299224 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1c20375-9fcb-4886-a91c-3dd36651532a-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c1c20375-9fcb-4886-a91c-3dd36651532a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 18 09:20:24 crc kubenswrapper[4556]: I0218 09:20:24.299279 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1c20375-9fcb-4886-a91c-3dd36651532a-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"c1c20375-9fcb-4886-a91c-3dd36651532a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 18 09:20:24 crc kubenswrapper[4556]: I0218 09:20:24.302439 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1c20375-9fcb-4886-a91c-3dd36651532a-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"c1c20375-9fcb-4886-a91c-3dd36651532a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 18 09:20:24 crc kubenswrapper[4556]: I0218 09:20:24.303132 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1c20375-9fcb-4886-a91c-3dd36651532a-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c1c20375-9fcb-4886-a91c-3dd36651532a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 18 09:20:24 crc kubenswrapper[4556]: I0218 09:20:24.310671 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1c20375-9fcb-4886-a91c-3dd36651532a-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c1c20375-9fcb-4886-a91c-3dd36651532a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 18 09:20:24 crc kubenswrapper[4556]: I0218 09:20:24.311144 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1c20375-9fcb-4886-a91c-3dd36651532a-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"c1c20375-9fcb-4886-a91c-3dd36651532a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 18 09:20:24 crc kubenswrapper[4556]: I0218 09:20:24.313602 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52wpl\" (UniqueName: \"kubernetes.io/projected/c1c20375-9fcb-4886-a91c-3dd36651532a-kube-api-access-52wpl\") pod \"nova-cell1-novncproxy-0\" (UID: \"c1c20375-9fcb-4886-a91c-3dd36651532a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 18 09:20:24 crc kubenswrapper[4556]: I0218 09:20:24.386235 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 18 09:20:24 crc kubenswrapper[4556]: I0218 09:20:24.394437 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 18 09:20:24 crc kubenswrapper[4556]: I0218 09:20:24.812625 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 18 09:20:24 crc kubenswrapper[4556]: W0218 09:20:24.814777 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1f3e12da_26fe_4d2a_ae62_9a3e90c5b752.slice/crio-b699a8125f463362973c934736b8ceeaefeb9673fd94c80f14929f68464a2e7d WatchSource:0}: Error finding container b699a8125f463362973c934736b8ceeaefeb9673fd94c80f14929f68464a2e7d: Status 404 returned error can't find the container with id b699a8125f463362973c934736b8ceeaefeb9673fd94c80f14929f68464a2e7d Feb 18 09:20:24 crc kubenswrapper[4556]: I0218 09:20:24.891131 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 18 09:20:24 crc kubenswrapper[4556]: W0218 09:20:24.913124 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc1c20375_9fcb_4886_a91c_3dd36651532a.slice/crio-f0add84d84649b3e0e2a91eb3a3d20d8f7bfc4991f19309b6d8eab4fff3e2e6e WatchSource:0}: Error finding container f0add84d84649b3e0e2a91eb3a3d20d8f7bfc4991f19309b6d8eab4fff3e2e6e: Status 404 returned error can't find the container with id f0add84d84649b3e0e2a91eb3a3d20d8f7bfc4991f19309b6d8eab4fff3e2e6e Feb 18 09:20:25 crc kubenswrapper[4556]: I0218 09:20:25.001722 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1f3e12da-26fe-4d2a-ae62-9a3e90c5b752","Type":"ContainerStarted","Data":"b699a8125f463362973c934736b8ceeaefeb9673fd94c80f14929f68464a2e7d"} Feb 18 09:20:25 crc kubenswrapper[4556]: I0218 09:20:25.003609 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"c1c20375-9fcb-4886-a91c-3dd36651532a","Type":"ContainerStarted","Data":"f0add84d84649b3e0e2a91eb3a3d20d8f7bfc4991f19309b6d8eab4fff3e2e6e"} Feb 18 09:20:25 crc kubenswrapper[4556]: I0218 09:20:25.164062 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 18 09:20:25 crc kubenswrapper[4556]: I0218 09:20:25.164705 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 18 09:20:25 crc kubenswrapper[4556]: I0218 09:20:25.164890 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 18 09:20:25 crc kubenswrapper[4556]: I0218 09:20:25.176463 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 18 09:20:25 crc kubenswrapper[4556]: I0218 09:20:25.294902 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2330f995-37c7-4b3e-a465-4e682349346f" path="/var/lib/kubelet/pods/2330f995-37c7-4b3e-a465-4e682349346f/volumes" Feb 18 09:20:25 crc kubenswrapper[4556]: I0218 09:20:25.295530 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ea5a5ac-537e-40d2-a5dc-4cc2314bf097" path="/var/lib/kubelet/pods/4ea5a5ac-537e-40d2-a5dc-4cc2314bf097/volumes" Feb 18 09:20:26 crc kubenswrapper[4556]: I0218 09:20:26.014358 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1f3e12da-26fe-4d2a-ae62-9a3e90c5b752","Type":"ContainerStarted","Data":"ea88e490cd9bdf056778cbda1d520395ed86855d37a5c3dec73dbb3079f8c795"} Feb 18 09:20:26 crc kubenswrapper[4556]: I0218 09:20:26.014677 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1f3e12da-26fe-4d2a-ae62-9a3e90c5b752","Type":"ContainerStarted","Data":"a903312a63e39580644536734e7abd02a75c7922bd9cd0e7e5635d7c7d0e5fca"} Feb 18 09:20:26 crc kubenswrapper[4556]: I0218 09:20:26.018303 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"c1c20375-9fcb-4886-a91c-3dd36651532a","Type":"ContainerStarted","Data":"1ed50cee9f397c7fcb049d4d4dc137dea63b37ecabe6509f232d2c781814706d"} Feb 18 09:20:26 crc kubenswrapper[4556]: I0218 09:20:26.018366 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 18 09:20:26 crc kubenswrapper[4556]: I0218 09:20:26.021413 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 18 09:20:26 crc kubenswrapper[4556]: I0218 09:20:26.068365 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.068348736 podStartE2EDuration="2.068348736s" podCreationTimestamp="2026-02-18 09:20:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:20:26.040145642 +0000 UTC m=+983.057106621" watchObservedRunningTime="2026-02-18 09:20:26.068348736 +0000 UTC m=+983.085309716" Feb 18 09:20:26 crc kubenswrapper[4556]: I0218 09:20:26.075536 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.075514186 podStartE2EDuration="2.075514186s" podCreationTimestamp="2026-02-18 09:20:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:20:26.071252565 +0000 UTC m=+983.088213545" watchObservedRunningTime="2026-02-18 09:20:26.075514186 +0000 UTC m=+983.092475166" Feb 18 09:20:26 crc kubenswrapper[4556]: I0218 09:20:26.223571 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-74b99f9475-7n7kb"] Feb 18 09:20:26 crc kubenswrapper[4556]: I0218 09:20:26.227568 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74b99f9475-7n7kb" Feb 18 09:20:26 crc kubenswrapper[4556]: I0218 09:20:26.252034 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74b99f9475-7n7kb"] Feb 18 09:20:26 crc kubenswrapper[4556]: I0218 09:20:26.332556 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a7122b24-4132-4967-a132-631e9ec2a11a-ovsdbserver-nb\") pod \"dnsmasq-dns-74b99f9475-7n7kb\" (UID: \"a7122b24-4132-4967-a132-631e9ec2a11a\") " pod="openstack/dnsmasq-dns-74b99f9475-7n7kb" Feb 18 09:20:26 crc kubenswrapper[4556]: I0218 09:20:26.332625 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2flzl\" (UniqueName: \"kubernetes.io/projected/a7122b24-4132-4967-a132-631e9ec2a11a-kube-api-access-2flzl\") pod \"dnsmasq-dns-74b99f9475-7n7kb\" (UID: \"a7122b24-4132-4967-a132-631e9ec2a11a\") " pod="openstack/dnsmasq-dns-74b99f9475-7n7kb" Feb 18 09:20:26 crc kubenswrapper[4556]: I0218 09:20:26.332703 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a7122b24-4132-4967-a132-631e9ec2a11a-dns-swift-storage-0\") pod \"dnsmasq-dns-74b99f9475-7n7kb\" (UID: \"a7122b24-4132-4967-a132-631e9ec2a11a\") " pod="openstack/dnsmasq-dns-74b99f9475-7n7kb" Feb 18 09:20:26 crc kubenswrapper[4556]: I0218 09:20:26.332763 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a7122b24-4132-4967-a132-631e9ec2a11a-ovsdbserver-sb\") pod \"dnsmasq-dns-74b99f9475-7n7kb\" (UID: \"a7122b24-4132-4967-a132-631e9ec2a11a\") " pod="openstack/dnsmasq-dns-74b99f9475-7n7kb" Feb 18 09:20:26 crc kubenswrapper[4556]: I0218 09:20:26.333084 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a7122b24-4132-4967-a132-631e9ec2a11a-dns-svc\") pod \"dnsmasq-dns-74b99f9475-7n7kb\" (UID: \"a7122b24-4132-4967-a132-631e9ec2a11a\") " pod="openstack/dnsmasq-dns-74b99f9475-7n7kb" Feb 18 09:20:26 crc kubenswrapper[4556]: I0218 09:20:26.333128 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7122b24-4132-4967-a132-631e9ec2a11a-config\") pod \"dnsmasq-dns-74b99f9475-7n7kb\" (UID: \"a7122b24-4132-4967-a132-631e9ec2a11a\") " pod="openstack/dnsmasq-dns-74b99f9475-7n7kb" Feb 18 09:20:26 crc kubenswrapper[4556]: I0218 09:20:26.435898 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a7122b24-4132-4967-a132-631e9ec2a11a-dns-svc\") pod \"dnsmasq-dns-74b99f9475-7n7kb\" (UID: \"a7122b24-4132-4967-a132-631e9ec2a11a\") " pod="openstack/dnsmasq-dns-74b99f9475-7n7kb" Feb 18 09:20:26 crc kubenswrapper[4556]: I0218 09:20:26.435950 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7122b24-4132-4967-a132-631e9ec2a11a-config\") pod \"dnsmasq-dns-74b99f9475-7n7kb\" (UID: \"a7122b24-4132-4967-a132-631e9ec2a11a\") " pod="openstack/dnsmasq-dns-74b99f9475-7n7kb" Feb 18 09:20:26 crc kubenswrapper[4556]: I0218 09:20:26.436213 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a7122b24-4132-4967-a132-631e9ec2a11a-ovsdbserver-nb\") pod \"dnsmasq-dns-74b99f9475-7n7kb\" (UID: \"a7122b24-4132-4967-a132-631e9ec2a11a\") " pod="openstack/dnsmasq-dns-74b99f9475-7n7kb" Feb 18 09:20:26 crc kubenswrapper[4556]: I0218 09:20:26.436405 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2flzl\" (UniqueName: \"kubernetes.io/projected/a7122b24-4132-4967-a132-631e9ec2a11a-kube-api-access-2flzl\") pod \"dnsmasq-dns-74b99f9475-7n7kb\" (UID: \"a7122b24-4132-4967-a132-631e9ec2a11a\") " pod="openstack/dnsmasq-dns-74b99f9475-7n7kb" Feb 18 09:20:26 crc kubenswrapper[4556]: I0218 09:20:26.436813 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a7122b24-4132-4967-a132-631e9ec2a11a-dns-svc\") pod \"dnsmasq-dns-74b99f9475-7n7kb\" (UID: \"a7122b24-4132-4967-a132-631e9ec2a11a\") " pod="openstack/dnsmasq-dns-74b99f9475-7n7kb" Feb 18 09:20:26 crc kubenswrapper[4556]: I0218 09:20:26.437212 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a7122b24-4132-4967-a132-631e9ec2a11a-ovsdbserver-nb\") pod \"dnsmasq-dns-74b99f9475-7n7kb\" (UID: \"a7122b24-4132-4967-a132-631e9ec2a11a\") " pod="openstack/dnsmasq-dns-74b99f9475-7n7kb" Feb 18 09:20:26 crc kubenswrapper[4556]: I0218 09:20:26.437257 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7122b24-4132-4967-a132-631e9ec2a11a-config\") pod \"dnsmasq-dns-74b99f9475-7n7kb\" (UID: \"a7122b24-4132-4967-a132-631e9ec2a11a\") " pod="openstack/dnsmasq-dns-74b99f9475-7n7kb" Feb 18 09:20:26 crc kubenswrapper[4556]: I0218 09:20:26.437413 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a7122b24-4132-4967-a132-631e9ec2a11a-dns-swift-storage-0\") pod \"dnsmasq-dns-74b99f9475-7n7kb\" (UID: \"a7122b24-4132-4967-a132-631e9ec2a11a\") " pod="openstack/dnsmasq-dns-74b99f9475-7n7kb" Feb 18 09:20:26 crc kubenswrapper[4556]: I0218 09:20:26.437531 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a7122b24-4132-4967-a132-631e9ec2a11a-ovsdbserver-sb\") pod \"dnsmasq-dns-74b99f9475-7n7kb\" (UID: \"a7122b24-4132-4967-a132-631e9ec2a11a\") " pod="openstack/dnsmasq-dns-74b99f9475-7n7kb" Feb 18 09:20:26 crc kubenswrapper[4556]: I0218 09:20:26.438026 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a7122b24-4132-4967-a132-631e9ec2a11a-dns-swift-storage-0\") pod \"dnsmasq-dns-74b99f9475-7n7kb\" (UID: \"a7122b24-4132-4967-a132-631e9ec2a11a\") " pod="openstack/dnsmasq-dns-74b99f9475-7n7kb" Feb 18 09:20:26 crc kubenswrapper[4556]: I0218 09:20:26.438294 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a7122b24-4132-4967-a132-631e9ec2a11a-ovsdbserver-sb\") pod \"dnsmasq-dns-74b99f9475-7n7kb\" (UID: \"a7122b24-4132-4967-a132-631e9ec2a11a\") " pod="openstack/dnsmasq-dns-74b99f9475-7n7kb" Feb 18 09:20:26 crc kubenswrapper[4556]: I0218 09:20:26.458625 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2flzl\" (UniqueName: \"kubernetes.io/projected/a7122b24-4132-4967-a132-631e9ec2a11a-kube-api-access-2flzl\") pod \"dnsmasq-dns-74b99f9475-7n7kb\" (UID: \"a7122b24-4132-4967-a132-631e9ec2a11a\") " pod="openstack/dnsmasq-dns-74b99f9475-7n7kb" Feb 18 09:20:26 crc kubenswrapper[4556]: I0218 09:20:26.555799 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74b99f9475-7n7kb" Feb 18 09:20:26 crc kubenswrapper[4556]: I0218 09:20:26.990981 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74b99f9475-7n7kb"] Feb 18 09:20:26 crc kubenswrapper[4556]: W0218 09:20:26.997383 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda7122b24_4132_4967_a132_631e9ec2a11a.slice/crio-eec59aeb42f6a1cfdcc9f8ebe6b85c1d01c207de75c7cbc46c24134ad8d66349 WatchSource:0}: Error finding container eec59aeb42f6a1cfdcc9f8ebe6b85c1d01c207de75c7cbc46c24134ad8d66349: Status 404 returned error can't find the container with id eec59aeb42f6a1cfdcc9f8ebe6b85c1d01c207de75c7cbc46c24134ad8d66349 Feb 18 09:20:27 crc kubenswrapper[4556]: I0218 09:20:27.035762 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74b99f9475-7n7kb" event={"ID":"a7122b24-4132-4967-a132-631e9ec2a11a","Type":"ContainerStarted","Data":"eec59aeb42f6a1cfdcc9f8ebe6b85c1d01c207de75c7cbc46c24134ad8d66349"} Feb 18 09:20:28 crc kubenswrapper[4556]: I0218 09:20:28.047183 4556 generic.go:334] "Generic (PLEG): container finished" podID="a7122b24-4132-4967-a132-631e9ec2a11a" containerID="b7eeec31510c3a6176773eb9808a367f07b7ee25d6e1167b0475f5e8add8aaba" exitCode=0 Feb 18 09:20:28 crc kubenswrapper[4556]: I0218 09:20:28.047287 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74b99f9475-7n7kb" event={"ID":"a7122b24-4132-4967-a132-631e9ec2a11a","Type":"ContainerDied","Data":"b7eeec31510c3a6176773eb9808a367f07b7ee25d6e1167b0475f5e8add8aaba"} Feb 18 09:20:28 crc kubenswrapper[4556]: I0218 09:20:28.203653 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 18 09:20:28 crc kubenswrapper[4556]: I0218 09:20:28.204172 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="855cf21a-1bf7-49e5-b27c-47a8fe960e42" containerName="ceilometer-central-agent" containerID="cri-o://d6dd7effa6f1baf748f47faedfdd57c3c9527b8f05be61c68495cd18a10f263f" gracePeriod=30 Feb 18 09:20:28 crc kubenswrapper[4556]: I0218 09:20:28.204261 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="855cf21a-1bf7-49e5-b27c-47a8fe960e42" containerName="proxy-httpd" containerID="cri-o://6a816d7032e88656777541c6d25ea7a7a7c708781de32ccd8cdda1f02e90faaa" gracePeriod=30 Feb 18 09:20:28 crc kubenswrapper[4556]: I0218 09:20:28.204260 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="855cf21a-1bf7-49e5-b27c-47a8fe960e42" containerName="ceilometer-notification-agent" containerID="cri-o://0ec74627d62a117c272bed8fdd57b93e3f5a88c7e79f7fcd985279ee964840f1" gracePeriod=30 Feb 18 09:20:28 crc kubenswrapper[4556]: I0218 09:20:28.204499 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="855cf21a-1bf7-49e5-b27c-47a8fe960e42" containerName="sg-core" containerID="cri-o://7bb494813010282cbf0390a221a66dd0231c3ef2a0bf1812709276fa3689354e" gracePeriod=30 Feb 18 09:20:28 crc kubenswrapper[4556]: I0218 09:20:28.479312 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 18 09:20:29 crc kubenswrapper[4556]: I0218 09:20:29.056873 4556 generic.go:334] "Generic (PLEG): container finished" podID="855cf21a-1bf7-49e5-b27c-47a8fe960e42" containerID="6a816d7032e88656777541c6d25ea7a7a7c708781de32ccd8cdda1f02e90faaa" exitCode=0 Feb 18 09:20:29 crc kubenswrapper[4556]: I0218 09:20:29.056902 4556 generic.go:334] "Generic (PLEG): container finished" podID="855cf21a-1bf7-49e5-b27c-47a8fe960e42" containerID="7bb494813010282cbf0390a221a66dd0231c3ef2a0bf1812709276fa3689354e" exitCode=2 Feb 18 09:20:29 crc kubenswrapper[4556]: I0218 09:20:29.056910 4556 generic.go:334] "Generic (PLEG): container finished" podID="855cf21a-1bf7-49e5-b27c-47a8fe960e42" containerID="0ec74627d62a117c272bed8fdd57b93e3f5a88c7e79f7fcd985279ee964840f1" exitCode=0 Feb 18 09:20:29 crc kubenswrapper[4556]: I0218 09:20:29.056916 4556 generic.go:334] "Generic (PLEG): container finished" podID="855cf21a-1bf7-49e5-b27c-47a8fe960e42" containerID="d6dd7effa6f1baf748f47faedfdd57c3c9527b8f05be61c68495cd18a10f263f" exitCode=0 Feb 18 09:20:29 crc kubenswrapper[4556]: I0218 09:20:29.056957 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"855cf21a-1bf7-49e5-b27c-47a8fe960e42","Type":"ContainerDied","Data":"6a816d7032e88656777541c6d25ea7a7a7c708781de32ccd8cdda1f02e90faaa"} Feb 18 09:20:29 crc kubenswrapper[4556]: I0218 09:20:29.056996 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"855cf21a-1bf7-49e5-b27c-47a8fe960e42","Type":"ContainerDied","Data":"7bb494813010282cbf0390a221a66dd0231c3ef2a0bf1812709276fa3689354e"} Feb 18 09:20:29 crc kubenswrapper[4556]: I0218 09:20:29.057025 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"855cf21a-1bf7-49e5-b27c-47a8fe960e42","Type":"ContainerDied","Data":"0ec74627d62a117c272bed8fdd57b93e3f5a88c7e79f7fcd985279ee964840f1"} Feb 18 09:20:29 crc kubenswrapper[4556]: I0218 09:20:29.057036 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"855cf21a-1bf7-49e5-b27c-47a8fe960e42","Type":"ContainerDied","Data":"d6dd7effa6f1baf748f47faedfdd57c3c9527b8f05be61c68495cd18a10f263f"} Feb 18 09:20:29 crc kubenswrapper[4556]: I0218 09:20:29.059350 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="4c7a2344-f7ad-40c2-84e9-60abcf73571a" containerName="nova-api-log" containerID="cri-o://6b0cc9a7e039bccb8d4df425c5577d32ffa7d1ec8bf2b43efd97ff0ee1d5bb50" gracePeriod=30 Feb 18 09:20:29 crc kubenswrapper[4556]: I0218 09:20:29.061307 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74b99f9475-7n7kb" event={"ID":"a7122b24-4132-4967-a132-631e9ec2a11a","Type":"ContainerStarted","Data":"8bf3c5eb66360572cb8dde27f0e8c31413e0e5c2c108fc1a73999f8e43241056"} Feb 18 09:20:29 crc kubenswrapper[4556]: I0218 09:20:29.061345 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-74b99f9475-7n7kb" Feb 18 09:20:29 crc kubenswrapper[4556]: I0218 09:20:29.061655 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="4c7a2344-f7ad-40c2-84e9-60abcf73571a" containerName="nova-api-api" containerID="cri-o://ba073e3d7ddd3fbd98e6ef10ee94d22dd5e5331510f771de93405ce4667d327e" gracePeriod=30 Feb 18 09:20:29 crc kubenswrapper[4556]: I0218 09:20:29.079621 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-74b99f9475-7n7kb" podStartSLOduration=3.079610068 podStartE2EDuration="3.079610068s" podCreationTimestamp="2026-02-18 09:20:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:20:29.076889755 +0000 UTC m=+986.093850736" watchObservedRunningTime="2026-02-18 09:20:29.079610068 +0000 UTC m=+986.096571048" Feb 18 09:20:29 crc kubenswrapper[4556]: I0218 09:20:29.337573 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 18 09:20:29 crc kubenswrapper[4556]: I0218 09:20:29.386521 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 18 09:20:29 crc kubenswrapper[4556]: I0218 09:20:29.387786 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 18 09:20:29 crc kubenswrapper[4556]: I0218 09:20:29.394992 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Feb 18 09:20:29 crc kubenswrapper[4556]: I0218 09:20:29.518625 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/855cf21a-1bf7-49e5-b27c-47a8fe960e42-ceilometer-tls-certs\") pod \"855cf21a-1bf7-49e5-b27c-47a8fe960e42\" (UID: \"855cf21a-1bf7-49e5-b27c-47a8fe960e42\") " Feb 18 09:20:29 crc kubenswrapper[4556]: I0218 09:20:29.518813 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/855cf21a-1bf7-49e5-b27c-47a8fe960e42-run-httpd\") pod \"855cf21a-1bf7-49e5-b27c-47a8fe960e42\" (UID: \"855cf21a-1bf7-49e5-b27c-47a8fe960e42\") " Feb 18 09:20:29 crc kubenswrapper[4556]: I0218 09:20:29.518913 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jb5mx\" (UniqueName: \"kubernetes.io/projected/855cf21a-1bf7-49e5-b27c-47a8fe960e42-kube-api-access-jb5mx\") pod \"855cf21a-1bf7-49e5-b27c-47a8fe960e42\" (UID: \"855cf21a-1bf7-49e5-b27c-47a8fe960e42\") " Feb 18 09:20:29 crc kubenswrapper[4556]: I0218 09:20:29.518947 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/855cf21a-1bf7-49e5-b27c-47a8fe960e42-log-httpd\") pod \"855cf21a-1bf7-49e5-b27c-47a8fe960e42\" (UID: \"855cf21a-1bf7-49e5-b27c-47a8fe960e42\") " Feb 18 09:20:29 crc kubenswrapper[4556]: I0218 09:20:29.519052 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/855cf21a-1bf7-49e5-b27c-47a8fe960e42-sg-core-conf-yaml\") pod \"855cf21a-1bf7-49e5-b27c-47a8fe960e42\" (UID: \"855cf21a-1bf7-49e5-b27c-47a8fe960e42\") " Feb 18 09:20:29 crc kubenswrapper[4556]: I0218 09:20:29.519091 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/855cf21a-1bf7-49e5-b27c-47a8fe960e42-config-data\") pod \"855cf21a-1bf7-49e5-b27c-47a8fe960e42\" (UID: \"855cf21a-1bf7-49e5-b27c-47a8fe960e42\") " Feb 18 09:20:29 crc kubenswrapper[4556]: I0218 09:20:29.519122 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/855cf21a-1bf7-49e5-b27c-47a8fe960e42-combined-ca-bundle\") pod \"855cf21a-1bf7-49e5-b27c-47a8fe960e42\" (UID: \"855cf21a-1bf7-49e5-b27c-47a8fe960e42\") " Feb 18 09:20:29 crc kubenswrapper[4556]: I0218 09:20:29.519082 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/855cf21a-1bf7-49e5-b27c-47a8fe960e42-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "855cf21a-1bf7-49e5-b27c-47a8fe960e42" (UID: "855cf21a-1bf7-49e5-b27c-47a8fe960e42"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:20:29 crc kubenswrapper[4556]: I0218 09:20:29.519214 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/855cf21a-1bf7-49e5-b27c-47a8fe960e42-scripts\") pod \"855cf21a-1bf7-49e5-b27c-47a8fe960e42\" (UID: \"855cf21a-1bf7-49e5-b27c-47a8fe960e42\") " Feb 18 09:20:29 crc kubenswrapper[4556]: I0218 09:20:29.519892 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/855cf21a-1bf7-49e5-b27c-47a8fe960e42-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "855cf21a-1bf7-49e5-b27c-47a8fe960e42" (UID: "855cf21a-1bf7-49e5-b27c-47a8fe960e42"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:20:29 crc kubenswrapper[4556]: I0218 09:20:29.522269 4556 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/855cf21a-1bf7-49e5-b27c-47a8fe960e42-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 18 09:20:29 crc kubenswrapper[4556]: I0218 09:20:29.522299 4556 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/855cf21a-1bf7-49e5-b27c-47a8fe960e42-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 18 09:20:29 crc kubenswrapper[4556]: I0218 09:20:29.526702 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/855cf21a-1bf7-49e5-b27c-47a8fe960e42-scripts" (OuterVolumeSpecName: "scripts") pod "855cf21a-1bf7-49e5-b27c-47a8fe960e42" (UID: "855cf21a-1bf7-49e5-b27c-47a8fe960e42"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:20:29 crc kubenswrapper[4556]: I0218 09:20:29.526831 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/855cf21a-1bf7-49e5-b27c-47a8fe960e42-kube-api-access-jb5mx" (OuterVolumeSpecName: "kube-api-access-jb5mx") pod "855cf21a-1bf7-49e5-b27c-47a8fe960e42" (UID: "855cf21a-1bf7-49e5-b27c-47a8fe960e42"). InnerVolumeSpecName "kube-api-access-jb5mx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:20:29 crc kubenswrapper[4556]: I0218 09:20:29.555300 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/855cf21a-1bf7-49e5-b27c-47a8fe960e42-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "855cf21a-1bf7-49e5-b27c-47a8fe960e42" (UID: "855cf21a-1bf7-49e5-b27c-47a8fe960e42"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:20:29 crc kubenswrapper[4556]: I0218 09:20:29.582542 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/855cf21a-1bf7-49e5-b27c-47a8fe960e42-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "855cf21a-1bf7-49e5-b27c-47a8fe960e42" (UID: "855cf21a-1bf7-49e5-b27c-47a8fe960e42"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:20:29 crc kubenswrapper[4556]: I0218 09:20:29.597278 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/855cf21a-1bf7-49e5-b27c-47a8fe960e42-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "855cf21a-1bf7-49e5-b27c-47a8fe960e42" (UID: "855cf21a-1bf7-49e5-b27c-47a8fe960e42"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:20:29 crc kubenswrapper[4556]: I0218 09:20:29.613433 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/855cf21a-1bf7-49e5-b27c-47a8fe960e42-config-data" (OuterVolumeSpecName: "config-data") pod "855cf21a-1bf7-49e5-b27c-47a8fe960e42" (UID: "855cf21a-1bf7-49e5-b27c-47a8fe960e42"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:20:29 crc kubenswrapper[4556]: I0218 09:20:29.625199 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jb5mx\" (UniqueName: \"kubernetes.io/projected/855cf21a-1bf7-49e5-b27c-47a8fe960e42-kube-api-access-jb5mx\") on node \"crc\" DevicePath \"\"" Feb 18 09:20:29 crc kubenswrapper[4556]: I0218 09:20:29.625233 4556 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/855cf21a-1bf7-49e5-b27c-47a8fe960e42-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 18 09:20:29 crc kubenswrapper[4556]: I0218 09:20:29.625247 4556 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/855cf21a-1bf7-49e5-b27c-47a8fe960e42-config-data\") on node \"crc\" DevicePath \"\"" Feb 18 09:20:29 crc kubenswrapper[4556]: I0218 09:20:29.625261 4556 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/855cf21a-1bf7-49e5-b27c-47a8fe960e42-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:20:29 crc kubenswrapper[4556]: I0218 09:20:29.625271 4556 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/855cf21a-1bf7-49e5-b27c-47a8fe960e42-scripts\") on node \"crc\" DevicePath \"\"" Feb 18 09:20:29 crc kubenswrapper[4556]: I0218 09:20:29.625282 4556 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/855cf21a-1bf7-49e5-b27c-47a8fe960e42-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 18 09:20:30 crc kubenswrapper[4556]: I0218 09:20:30.069283 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"855cf21a-1bf7-49e5-b27c-47a8fe960e42","Type":"ContainerDied","Data":"938c2e818c458c5ceb9c1e3558089112f5524a6763219b3a89d978af5746b2d2"} Feb 18 09:20:30 crc kubenswrapper[4556]: I0218 09:20:30.069345 4556 scope.go:117] "RemoveContainer" containerID="6a816d7032e88656777541c6d25ea7a7a7c708781de32ccd8cdda1f02e90faaa" Feb 18 09:20:30 crc kubenswrapper[4556]: I0218 09:20:30.069473 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 18 09:20:30 crc kubenswrapper[4556]: I0218 09:20:30.072576 4556 generic.go:334] "Generic (PLEG): container finished" podID="4c7a2344-f7ad-40c2-84e9-60abcf73571a" containerID="6b0cc9a7e039bccb8d4df425c5577d32ffa7d1ec8bf2b43efd97ff0ee1d5bb50" exitCode=143 Feb 18 09:20:30 crc kubenswrapper[4556]: I0218 09:20:30.072715 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4c7a2344-f7ad-40c2-84e9-60abcf73571a","Type":"ContainerDied","Data":"6b0cc9a7e039bccb8d4df425c5577d32ffa7d1ec8bf2b43efd97ff0ee1d5bb50"} Feb 18 09:20:30 crc kubenswrapper[4556]: I0218 09:20:30.130590 4556 scope.go:117] "RemoveContainer" containerID="7bb494813010282cbf0390a221a66dd0231c3ef2a0bf1812709276fa3689354e" Feb 18 09:20:30 crc kubenswrapper[4556]: I0218 09:20:30.143996 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 18 09:20:30 crc kubenswrapper[4556]: I0218 09:20:30.161437 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 18 09:20:30 crc kubenswrapper[4556]: I0218 09:20:30.167323 4556 scope.go:117] "RemoveContainer" containerID="0ec74627d62a117c272bed8fdd57b93e3f5a88c7e79f7fcd985279ee964840f1" Feb 18 09:20:30 crc kubenswrapper[4556]: I0218 09:20:30.169523 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 18 09:20:30 crc kubenswrapper[4556]: E0218 09:20:30.169900 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="855cf21a-1bf7-49e5-b27c-47a8fe960e42" containerName="proxy-httpd" Feb 18 09:20:30 crc kubenswrapper[4556]: I0218 09:20:30.169921 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="855cf21a-1bf7-49e5-b27c-47a8fe960e42" containerName="proxy-httpd" Feb 18 09:20:30 crc kubenswrapper[4556]: E0218 09:20:30.169939 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="855cf21a-1bf7-49e5-b27c-47a8fe960e42" containerName="ceilometer-central-agent" Feb 18 09:20:30 crc kubenswrapper[4556]: I0218 09:20:30.169945 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="855cf21a-1bf7-49e5-b27c-47a8fe960e42" containerName="ceilometer-central-agent" Feb 18 09:20:30 crc kubenswrapper[4556]: E0218 09:20:30.169958 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="855cf21a-1bf7-49e5-b27c-47a8fe960e42" containerName="sg-core" Feb 18 09:20:30 crc kubenswrapper[4556]: I0218 09:20:30.169964 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="855cf21a-1bf7-49e5-b27c-47a8fe960e42" containerName="sg-core" Feb 18 09:20:30 crc kubenswrapper[4556]: E0218 09:20:30.169981 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="855cf21a-1bf7-49e5-b27c-47a8fe960e42" containerName="ceilometer-notification-agent" Feb 18 09:20:30 crc kubenswrapper[4556]: I0218 09:20:30.169987 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="855cf21a-1bf7-49e5-b27c-47a8fe960e42" containerName="ceilometer-notification-agent" Feb 18 09:20:30 crc kubenswrapper[4556]: I0218 09:20:30.170199 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="855cf21a-1bf7-49e5-b27c-47a8fe960e42" containerName="ceilometer-notification-agent" Feb 18 09:20:30 crc kubenswrapper[4556]: I0218 09:20:30.170215 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="855cf21a-1bf7-49e5-b27c-47a8fe960e42" containerName="sg-core" Feb 18 09:20:30 crc kubenswrapper[4556]: I0218 09:20:30.170226 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="855cf21a-1bf7-49e5-b27c-47a8fe960e42" containerName="proxy-httpd" Feb 18 09:20:30 crc kubenswrapper[4556]: I0218 09:20:30.170240 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="855cf21a-1bf7-49e5-b27c-47a8fe960e42" containerName="ceilometer-central-agent" Feb 18 09:20:30 crc kubenswrapper[4556]: I0218 09:20:30.171828 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 18 09:20:30 crc kubenswrapper[4556]: I0218 09:20:30.178502 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Feb 18 09:20:30 crc kubenswrapper[4556]: I0218 09:20:30.178537 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 18 09:20:30 crc kubenswrapper[4556]: I0218 09:20:30.178804 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 18 09:20:30 crc kubenswrapper[4556]: I0218 09:20:30.186353 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 18 09:20:30 crc kubenswrapper[4556]: I0218 09:20:30.191461 4556 scope.go:117] "RemoveContainer" containerID="d6dd7effa6f1baf748f47faedfdd57c3c9527b8f05be61c68495cd18a10f263f" Feb 18 09:20:30 crc kubenswrapper[4556]: I0218 09:20:30.239350 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d105f935-cc14-4293-8101-17adaef31666-scripts\") pod \"ceilometer-0\" (UID: \"d105f935-cc14-4293-8101-17adaef31666\") " pod="openstack/ceilometer-0" Feb 18 09:20:30 crc kubenswrapper[4556]: I0218 09:20:30.239390 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d105f935-cc14-4293-8101-17adaef31666-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d105f935-cc14-4293-8101-17adaef31666\") " pod="openstack/ceilometer-0" Feb 18 09:20:30 crc kubenswrapper[4556]: I0218 09:20:30.239463 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d105f935-cc14-4293-8101-17adaef31666-config-data\") pod \"ceilometer-0\" (UID: \"d105f935-cc14-4293-8101-17adaef31666\") " pod="openstack/ceilometer-0" Feb 18 09:20:30 crc kubenswrapper[4556]: I0218 09:20:30.239509 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdnk7\" (UniqueName: \"kubernetes.io/projected/d105f935-cc14-4293-8101-17adaef31666-kube-api-access-cdnk7\") pod \"ceilometer-0\" (UID: \"d105f935-cc14-4293-8101-17adaef31666\") " pod="openstack/ceilometer-0" Feb 18 09:20:30 crc kubenswrapper[4556]: I0218 09:20:30.239562 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d105f935-cc14-4293-8101-17adaef31666-log-httpd\") pod \"ceilometer-0\" (UID: \"d105f935-cc14-4293-8101-17adaef31666\") " pod="openstack/ceilometer-0" Feb 18 09:20:30 crc kubenswrapper[4556]: I0218 09:20:30.239651 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d105f935-cc14-4293-8101-17adaef31666-run-httpd\") pod \"ceilometer-0\" (UID: \"d105f935-cc14-4293-8101-17adaef31666\") " pod="openstack/ceilometer-0" Feb 18 09:20:30 crc kubenswrapper[4556]: I0218 09:20:30.239847 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d105f935-cc14-4293-8101-17adaef31666-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d105f935-cc14-4293-8101-17adaef31666\") " pod="openstack/ceilometer-0" Feb 18 09:20:30 crc kubenswrapper[4556]: I0218 09:20:30.239987 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d105f935-cc14-4293-8101-17adaef31666-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d105f935-cc14-4293-8101-17adaef31666\") " pod="openstack/ceilometer-0" Feb 18 09:20:30 crc kubenswrapper[4556]: I0218 09:20:30.341437 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d105f935-cc14-4293-8101-17adaef31666-run-httpd\") pod \"ceilometer-0\" (UID: \"d105f935-cc14-4293-8101-17adaef31666\") " pod="openstack/ceilometer-0" Feb 18 09:20:30 crc kubenswrapper[4556]: I0218 09:20:30.341496 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d105f935-cc14-4293-8101-17adaef31666-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d105f935-cc14-4293-8101-17adaef31666\") " pod="openstack/ceilometer-0" Feb 18 09:20:30 crc kubenswrapper[4556]: I0218 09:20:30.341530 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d105f935-cc14-4293-8101-17adaef31666-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d105f935-cc14-4293-8101-17adaef31666\") " pod="openstack/ceilometer-0" Feb 18 09:20:30 crc kubenswrapper[4556]: I0218 09:20:30.341579 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d105f935-cc14-4293-8101-17adaef31666-scripts\") pod \"ceilometer-0\" (UID: \"d105f935-cc14-4293-8101-17adaef31666\") " pod="openstack/ceilometer-0" Feb 18 09:20:30 crc kubenswrapper[4556]: I0218 09:20:30.341599 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d105f935-cc14-4293-8101-17adaef31666-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d105f935-cc14-4293-8101-17adaef31666\") " pod="openstack/ceilometer-0" Feb 18 09:20:30 crc kubenswrapper[4556]: I0218 09:20:30.341655 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d105f935-cc14-4293-8101-17adaef31666-config-data\") pod \"ceilometer-0\" (UID: \"d105f935-cc14-4293-8101-17adaef31666\") " pod="openstack/ceilometer-0" Feb 18 09:20:30 crc kubenswrapper[4556]: I0218 09:20:30.341700 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdnk7\" (UniqueName: \"kubernetes.io/projected/d105f935-cc14-4293-8101-17adaef31666-kube-api-access-cdnk7\") pod \"ceilometer-0\" (UID: \"d105f935-cc14-4293-8101-17adaef31666\") " pod="openstack/ceilometer-0" Feb 18 09:20:30 crc kubenswrapper[4556]: I0218 09:20:30.341741 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d105f935-cc14-4293-8101-17adaef31666-log-httpd\") pod \"ceilometer-0\" (UID: \"d105f935-cc14-4293-8101-17adaef31666\") " pod="openstack/ceilometer-0" Feb 18 09:20:30 crc kubenswrapper[4556]: I0218 09:20:30.342456 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d105f935-cc14-4293-8101-17adaef31666-run-httpd\") pod \"ceilometer-0\" (UID: \"d105f935-cc14-4293-8101-17adaef31666\") " pod="openstack/ceilometer-0" Feb 18 09:20:30 crc kubenswrapper[4556]: I0218 09:20:30.342558 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d105f935-cc14-4293-8101-17adaef31666-log-httpd\") pod \"ceilometer-0\" (UID: \"d105f935-cc14-4293-8101-17adaef31666\") " pod="openstack/ceilometer-0" Feb 18 09:20:30 crc kubenswrapper[4556]: I0218 09:20:30.345374 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d105f935-cc14-4293-8101-17adaef31666-scripts\") pod \"ceilometer-0\" (UID: \"d105f935-cc14-4293-8101-17adaef31666\") " pod="openstack/ceilometer-0" Feb 18 09:20:30 crc kubenswrapper[4556]: I0218 09:20:30.345585 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d105f935-cc14-4293-8101-17adaef31666-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d105f935-cc14-4293-8101-17adaef31666\") " pod="openstack/ceilometer-0" Feb 18 09:20:30 crc kubenswrapper[4556]: I0218 09:20:30.346806 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d105f935-cc14-4293-8101-17adaef31666-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d105f935-cc14-4293-8101-17adaef31666\") " pod="openstack/ceilometer-0" Feb 18 09:20:30 crc kubenswrapper[4556]: I0218 09:20:30.347118 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d105f935-cc14-4293-8101-17adaef31666-config-data\") pod \"ceilometer-0\" (UID: \"d105f935-cc14-4293-8101-17adaef31666\") " pod="openstack/ceilometer-0" Feb 18 09:20:30 crc kubenswrapper[4556]: I0218 09:20:30.348650 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d105f935-cc14-4293-8101-17adaef31666-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d105f935-cc14-4293-8101-17adaef31666\") " pod="openstack/ceilometer-0" Feb 18 09:20:30 crc kubenswrapper[4556]: I0218 09:20:30.358535 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdnk7\" (UniqueName: \"kubernetes.io/projected/d105f935-cc14-4293-8101-17adaef31666-kube-api-access-cdnk7\") pod \"ceilometer-0\" (UID: \"d105f935-cc14-4293-8101-17adaef31666\") " pod="openstack/ceilometer-0" Feb 18 09:20:30 crc kubenswrapper[4556]: I0218 09:20:30.486891 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 18 09:20:30 crc kubenswrapper[4556]: I0218 09:20:30.865707 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 18 09:20:30 crc kubenswrapper[4556]: W0218 09:20:30.869877 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd105f935_cc14_4293_8101_17adaef31666.slice/crio-57f91dce89832973d068291f24fd2c6dd42bdc13bd4f548d3d895cd9fc07aa80 WatchSource:0}: Error finding container 57f91dce89832973d068291f24fd2c6dd42bdc13bd4f548d3d895cd9fc07aa80: Status 404 returned error can't find the container with id 57f91dce89832973d068291f24fd2c6dd42bdc13bd4f548d3d895cd9fc07aa80 Feb 18 09:20:31 crc kubenswrapper[4556]: I0218 09:20:31.082986 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d105f935-cc14-4293-8101-17adaef31666","Type":"ContainerStarted","Data":"57f91dce89832973d068291f24fd2c6dd42bdc13bd4f548d3d895cd9fc07aa80"} Feb 18 09:20:31 crc kubenswrapper[4556]: I0218 09:20:31.290980 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="855cf21a-1bf7-49e5-b27c-47a8fe960e42" path="/var/lib/kubelet/pods/855cf21a-1bf7-49e5-b27c-47a8fe960e42/volumes" Feb 18 09:20:32 crc kubenswrapper[4556]: I0218 09:20:32.091038 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d105f935-cc14-4293-8101-17adaef31666","Type":"ContainerStarted","Data":"96152665a0c8b88808d6b72c4eecd8e6c59abac9b90ae3511dd92633792ce0b6"} Feb 18 09:20:32 crc kubenswrapper[4556]: I0218 09:20:32.650630 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 18 09:20:32 crc kubenswrapper[4556]: I0218 09:20:32.718284 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c4h5t\" (UniqueName: \"kubernetes.io/projected/4c7a2344-f7ad-40c2-84e9-60abcf73571a-kube-api-access-c4h5t\") pod \"4c7a2344-f7ad-40c2-84e9-60abcf73571a\" (UID: \"4c7a2344-f7ad-40c2-84e9-60abcf73571a\") " Feb 18 09:20:32 crc kubenswrapper[4556]: I0218 09:20:32.718793 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c7a2344-f7ad-40c2-84e9-60abcf73571a-combined-ca-bundle\") pod \"4c7a2344-f7ad-40c2-84e9-60abcf73571a\" (UID: \"4c7a2344-f7ad-40c2-84e9-60abcf73571a\") " Feb 18 09:20:32 crc kubenswrapper[4556]: I0218 09:20:32.719177 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4c7a2344-f7ad-40c2-84e9-60abcf73571a-logs\") pod \"4c7a2344-f7ad-40c2-84e9-60abcf73571a\" (UID: \"4c7a2344-f7ad-40c2-84e9-60abcf73571a\") " Feb 18 09:20:32 crc kubenswrapper[4556]: I0218 09:20:32.719277 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c7a2344-f7ad-40c2-84e9-60abcf73571a-config-data\") pod \"4c7a2344-f7ad-40c2-84e9-60abcf73571a\" (UID: \"4c7a2344-f7ad-40c2-84e9-60abcf73571a\") " Feb 18 09:20:32 crc kubenswrapper[4556]: I0218 09:20:32.719556 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c7a2344-f7ad-40c2-84e9-60abcf73571a-logs" (OuterVolumeSpecName: "logs") pod "4c7a2344-f7ad-40c2-84e9-60abcf73571a" (UID: "4c7a2344-f7ad-40c2-84e9-60abcf73571a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:20:32 crc kubenswrapper[4556]: I0218 09:20:32.720492 4556 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4c7a2344-f7ad-40c2-84e9-60abcf73571a-logs\") on node \"crc\" DevicePath \"\"" Feb 18 09:20:32 crc kubenswrapper[4556]: I0218 09:20:32.723278 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c7a2344-f7ad-40c2-84e9-60abcf73571a-kube-api-access-c4h5t" (OuterVolumeSpecName: "kube-api-access-c4h5t") pod "4c7a2344-f7ad-40c2-84e9-60abcf73571a" (UID: "4c7a2344-f7ad-40c2-84e9-60abcf73571a"). InnerVolumeSpecName "kube-api-access-c4h5t". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:20:32 crc kubenswrapper[4556]: I0218 09:20:32.742562 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c7a2344-f7ad-40c2-84e9-60abcf73571a-config-data" (OuterVolumeSpecName: "config-data") pod "4c7a2344-f7ad-40c2-84e9-60abcf73571a" (UID: "4c7a2344-f7ad-40c2-84e9-60abcf73571a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:20:32 crc kubenswrapper[4556]: I0218 09:20:32.746472 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c7a2344-f7ad-40c2-84e9-60abcf73571a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4c7a2344-f7ad-40c2-84e9-60abcf73571a" (UID: "4c7a2344-f7ad-40c2-84e9-60abcf73571a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:20:32 crc kubenswrapper[4556]: I0218 09:20:32.821648 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c4h5t\" (UniqueName: \"kubernetes.io/projected/4c7a2344-f7ad-40c2-84e9-60abcf73571a-kube-api-access-c4h5t\") on node \"crc\" DevicePath \"\"" Feb 18 09:20:32 crc kubenswrapper[4556]: I0218 09:20:32.821679 4556 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c7a2344-f7ad-40c2-84e9-60abcf73571a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:20:32 crc kubenswrapper[4556]: I0218 09:20:32.821692 4556 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c7a2344-f7ad-40c2-84e9-60abcf73571a-config-data\") on node \"crc\" DevicePath \"\"" Feb 18 09:20:33 crc kubenswrapper[4556]: I0218 09:20:33.103902 4556 generic.go:334] "Generic (PLEG): container finished" podID="4c7a2344-f7ad-40c2-84e9-60abcf73571a" containerID="ba073e3d7ddd3fbd98e6ef10ee94d22dd5e5331510f771de93405ce4667d327e" exitCode=0 Feb 18 09:20:33 crc kubenswrapper[4556]: I0218 09:20:33.104004 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 18 09:20:33 crc kubenswrapper[4556]: I0218 09:20:33.104012 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4c7a2344-f7ad-40c2-84e9-60abcf73571a","Type":"ContainerDied","Data":"ba073e3d7ddd3fbd98e6ef10ee94d22dd5e5331510f771de93405ce4667d327e"} Feb 18 09:20:33 crc kubenswrapper[4556]: I0218 09:20:33.104953 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4c7a2344-f7ad-40c2-84e9-60abcf73571a","Type":"ContainerDied","Data":"96332f1c6215a8fc2e3fab8bf3b6de057e91e68295678a607be37be840224e09"} Feb 18 09:20:33 crc kubenswrapper[4556]: I0218 09:20:33.105001 4556 scope.go:117] "RemoveContainer" containerID="ba073e3d7ddd3fbd98e6ef10ee94d22dd5e5331510f771de93405ce4667d327e" Feb 18 09:20:33 crc kubenswrapper[4556]: I0218 09:20:33.116008 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d105f935-cc14-4293-8101-17adaef31666","Type":"ContainerStarted","Data":"566da5d72ede0706857f16c1c48bd1de38a296763e852ec9996fc54cc4b057b1"} Feb 18 09:20:33 crc kubenswrapper[4556]: I0218 09:20:33.171264 4556 scope.go:117] "RemoveContainer" containerID="6b0cc9a7e039bccb8d4df425c5577d32ffa7d1ec8bf2b43efd97ff0ee1d5bb50" Feb 18 09:20:33 crc kubenswrapper[4556]: I0218 09:20:33.180988 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 18 09:20:33 crc kubenswrapper[4556]: I0218 09:20:33.201781 4556 scope.go:117] "RemoveContainer" containerID="ba073e3d7ddd3fbd98e6ef10ee94d22dd5e5331510f771de93405ce4667d327e" Feb 18 09:20:33 crc kubenswrapper[4556]: I0218 09:20:33.202245 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 18 09:20:33 crc kubenswrapper[4556]: E0218 09:20:33.202511 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba073e3d7ddd3fbd98e6ef10ee94d22dd5e5331510f771de93405ce4667d327e\": container with ID starting with ba073e3d7ddd3fbd98e6ef10ee94d22dd5e5331510f771de93405ce4667d327e not found: ID does not exist" containerID="ba073e3d7ddd3fbd98e6ef10ee94d22dd5e5331510f771de93405ce4667d327e" Feb 18 09:20:33 crc kubenswrapper[4556]: I0218 09:20:33.202549 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba073e3d7ddd3fbd98e6ef10ee94d22dd5e5331510f771de93405ce4667d327e"} err="failed to get container status \"ba073e3d7ddd3fbd98e6ef10ee94d22dd5e5331510f771de93405ce4667d327e\": rpc error: code = NotFound desc = could not find container \"ba073e3d7ddd3fbd98e6ef10ee94d22dd5e5331510f771de93405ce4667d327e\": container with ID starting with ba073e3d7ddd3fbd98e6ef10ee94d22dd5e5331510f771de93405ce4667d327e not found: ID does not exist" Feb 18 09:20:33 crc kubenswrapper[4556]: I0218 09:20:33.202571 4556 scope.go:117] "RemoveContainer" containerID="6b0cc9a7e039bccb8d4df425c5577d32ffa7d1ec8bf2b43efd97ff0ee1d5bb50" Feb 18 09:20:33 crc kubenswrapper[4556]: E0218 09:20:33.203388 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b0cc9a7e039bccb8d4df425c5577d32ffa7d1ec8bf2b43efd97ff0ee1d5bb50\": container with ID starting with 6b0cc9a7e039bccb8d4df425c5577d32ffa7d1ec8bf2b43efd97ff0ee1d5bb50 not found: ID does not exist" containerID="6b0cc9a7e039bccb8d4df425c5577d32ffa7d1ec8bf2b43efd97ff0ee1d5bb50" Feb 18 09:20:33 crc kubenswrapper[4556]: I0218 09:20:33.203410 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b0cc9a7e039bccb8d4df425c5577d32ffa7d1ec8bf2b43efd97ff0ee1d5bb50"} err="failed to get container status \"6b0cc9a7e039bccb8d4df425c5577d32ffa7d1ec8bf2b43efd97ff0ee1d5bb50\": rpc error: code = NotFound desc = could not find container \"6b0cc9a7e039bccb8d4df425c5577d32ffa7d1ec8bf2b43efd97ff0ee1d5bb50\": container with ID starting with 6b0cc9a7e039bccb8d4df425c5577d32ffa7d1ec8bf2b43efd97ff0ee1d5bb50 not found: ID does not exist" Feb 18 09:20:33 crc kubenswrapper[4556]: I0218 09:20:33.211293 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 18 09:20:33 crc kubenswrapper[4556]: E0218 09:20:33.211743 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c7a2344-f7ad-40c2-84e9-60abcf73571a" containerName="nova-api-log" Feb 18 09:20:33 crc kubenswrapper[4556]: I0218 09:20:33.211755 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c7a2344-f7ad-40c2-84e9-60abcf73571a" containerName="nova-api-log" Feb 18 09:20:33 crc kubenswrapper[4556]: E0218 09:20:33.211770 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c7a2344-f7ad-40c2-84e9-60abcf73571a" containerName="nova-api-api" Feb 18 09:20:33 crc kubenswrapper[4556]: I0218 09:20:33.211781 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c7a2344-f7ad-40c2-84e9-60abcf73571a" containerName="nova-api-api" Feb 18 09:20:33 crc kubenswrapper[4556]: I0218 09:20:33.211978 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c7a2344-f7ad-40c2-84e9-60abcf73571a" containerName="nova-api-api" Feb 18 09:20:33 crc kubenswrapper[4556]: I0218 09:20:33.211986 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c7a2344-f7ad-40c2-84e9-60abcf73571a" containerName="nova-api-log" Feb 18 09:20:33 crc kubenswrapper[4556]: I0218 09:20:33.213049 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 18 09:20:33 crc kubenswrapper[4556]: I0218 09:20:33.215201 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 18 09:20:33 crc kubenswrapper[4556]: I0218 09:20:33.215844 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Feb 18 09:20:33 crc kubenswrapper[4556]: I0218 09:20:33.215882 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Feb 18 09:20:33 crc kubenswrapper[4556]: I0218 09:20:33.220723 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 18 09:20:33 crc kubenswrapper[4556]: I0218 09:20:33.226374 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87c7e959-f5b1-4d11-84d2-908268955e99-logs\") pod \"nova-api-0\" (UID: \"87c7e959-f5b1-4d11-84d2-908268955e99\") " pod="openstack/nova-api-0" Feb 18 09:20:33 crc kubenswrapper[4556]: I0218 09:20:33.226534 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/87c7e959-f5b1-4d11-84d2-908268955e99-public-tls-certs\") pod \"nova-api-0\" (UID: \"87c7e959-f5b1-4d11-84d2-908268955e99\") " pod="openstack/nova-api-0" Feb 18 09:20:33 crc kubenswrapper[4556]: I0218 09:20:33.226788 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6xdk\" (UniqueName: \"kubernetes.io/projected/87c7e959-f5b1-4d11-84d2-908268955e99-kube-api-access-s6xdk\") pod \"nova-api-0\" (UID: \"87c7e959-f5b1-4d11-84d2-908268955e99\") " pod="openstack/nova-api-0" Feb 18 09:20:33 crc kubenswrapper[4556]: I0218 09:20:33.226898 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87c7e959-f5b1-4d11-84d2-908268955e99-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"87c7e959-f5b1-4d11-84d2-908268955e99\") " pod="openstack/nova-api-0" Feb 18 09:20:33 crc kubenswrapper[4556]: I0218 09:20:33.226922 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/87c7e959-f5b1-4d11-84d2-908268955e99-internal-tls-certs\") pod \"nova-api-0\" (UID: \"87c7e959-f5b1-4d11-84d2-908268955e99\") " pod="openstack/nova-api-0" Feb 18 09:20:33 crc kubenswrapper[4556]: I0218 09:20:33.226973 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87c7e959-f5b1-4d11-84d2-908268955e99-config-data\") pod \"nova-api-0\" (UID: \"87c7e959-f5b1-4d11-84d2-908268955e99\") " pod="openstack/nova-api-0" Feb 18 09:20:33 crc kubenswrapper[4556]: I0218 09:20:33.294572 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c7a2344-f7ad-40c2-84e9-60abcf73571a" path="/var/lib/kubelet/pods/4c7a2344-f7ad-40c2-84e9-60abcf73571a/volumes" Feb 18 09:20:33 crc kubenswrapper[4556]: I0218 09:20:33.329107 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87c7e959-f5b1-4d11-84d2-908268955e99-logs\") pod \"nova-api-0\" (UID: \"87c7e959-f5b1-4d11-84d2-908268955e99\") " pod="openstack/nova-api-0" Feb 18 09:20:33 crc kubenswrapper[4556]: I0218 09:20:33.329224 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/87c7e959-f5b1-4d11-84d2-908268955e99-public-tls-certs\") pod \"nova-api-0\" (UID: \"87c7e959-f5b1-4d11-84d2-908268955e99\") " pod="openstack/nova-api-0" Feb 18 09:20:33 crc kubenswrapper[4556]: I0218 09:20:33.329339 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6xdk\" (UniqueName: \"kubernetes.io/projected/87c7e959-f5b1-4d11-84d2-908268955e99-kube-api-access-s6xdk\") pod \"nova-api-0\" (UID: \"87c7e959-f5b1-4d11-84d2-908268955e99\") " pod="openstack/nova-api-0" Feb 18 09:20:33 crc kubenswrapper[4556]: I0218 09:20:33.329395 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87c7e959-f5b1-4d11-84d2-908268955e99-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"87c7e959-f5b1-4d11-84d2-908268955e99\") " pod="openstack/nova-api-0" Feb 18 09:20:33 crc kubenswrapper[4556]: I0218 09:20:33.329413 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/87c7e959-f5b1-4d11-84d2-908268955e99-internal-tls-certs\") pod \"nova-api-0\" (UID: \"87c7e959-f5b1-4d11-84d2-908268955e99\") " pod="openstack/nova-api-0" Feb 18 09:20:33 crc kubenswrapper[4556]: I0218 09:20:33.329445 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87c7e959-f5b1-4d11-84d2-908268955e99-config-data\") pod \"nova-api-0\" (UID: \"87c7e959-f5b1-4d11-84d2-908268955e99\") " pod="openstack/nova-api-0" Feb 18 09:20:33 crc kubenswrapper[4556]: I0218 09:20:33.329486 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87c7e959-f5b1-4d11-84d2-908268955e99-logs\") pod \"nova-api-0\" (UID: \"87c7e959-f5b1-4d11-84d2-908268955e99\") " pod="openstack/nova-api-0" Feb 18 09:20:33 crc kubenswrapper[4556]: I0218 09:20:33.333397 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/87c7e959-f5b1-4d11-84d2-908268955e99-public-tls-certs\") pod \"nova-api-0\" (UID: \"87c7e959-f5b1-4d11-84d2-908268955e99\") " pod="openstack/nova-api-0" Feb 18 09:20:33 crc kubenswrapper[4556]: I0218 09:20:33.333915 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87c7e959-f5b1-4d11-84d2-908268955e99-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"87c7e959-f5b1-4d11-84d2-908268955e99\") " pod="openstack/nova-api-0" Feb 18 09:20:33 crc kubenswrapper[4556]: I0218 09:20:33.334038 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/87c7e959-f5b1-4d11-84d2-908268955e99-internal-tls-certs\") pod \"nova-api-0\" (UID: \"87c7e959-f5b1-4d11-84d2-908268955e99\") " pod="openstack/nova-api-0" Feb 18 09:20:33 crc kubenswrapper[4556]: I0218 09:20:33.339877 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87c7e959-f5b1-4d11-84d2-908268955e99-config-data\") pod \"nova-api-0\" (UID: \"87c7e959-f5b1-4d11-84d2-908268955e99\") " pod="openstack/nova-api-0" Feb 18 09:20:33 crc kubenswrapper[4556]: I0218 09:20:33.343789 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6xdk\" (UniqueName: \"kubernetes.io/projected/87c7e959-f5b1-4d11-84d2-908268955e99-kube-api-access-s6xdk\") pod \"nova-api-0\" (UID: \"87c7e959-f5b1-4d11-84d2-908268955e99\") " pod="openstack/nova-api-0" Feb 18 09:20:33 crc kubenswrapper[4556]: I0218 09:20:33.535302 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 18 09:20:33 crc kubenswrapper[4556]: I0218 09:20:33.945758 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 18 09:20:34 crc kubenswrapper[4556]: I0218 09:20:34.126420 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"87c7e959-f5b1-4d11-84d2-908268955e99","Type":"ContainerStarted","Data":"142b9b0dbdaf258de5293cb4fcb059abb419b6cc425a754422ba0e2e39a58bbc"} Feb 18 09:20:34 crc kubenswrapper[4556]: I0218 09:20:34.126463 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"87c7e959-f5b1-4d11-84d2-908268955e99","Type":"ContainerStarted","Data":"ea571bd28492885659f7750d9b6dd048aacbb1680da5fa29c14a75d6b35dc4da"} Feb 18 09:20:34 crc kubenswrapper[4556]: I0218 09:20:34.129789 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d105f935-cc14-4293-8101-17adaef31666","Type":"ContainerStarted","Data":"de1fd851b757765b659640539fd0336a4c1a31fc5faa2c59c48b99bd906ff90e"} Feb 18 09:20:34 crc kubenswrapper[4556]: I0218 09:20:34.386986 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 18 09:20:34 crc kubenswrapper[4556]: I0218 09:20:34.387361 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 18 09:20:34 crc kubenswrapper[4556]: I0218 09:20:34.395333 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Feb 18 09:20:34 crc kubenswrapper[4556]: I0218 09:20:34.415055 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Feb 18 09:20:35 crc kubenswrapper[4556]: I0218 09:20:35.140496 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"87c7e959-f5b1-4d11-84d2-908268955e99","Type":"ContainerStarted","Data":"68d4fb9b5beb0194e57ac5290bb9f1fccea67a235b1b0abbace8316ee010a9e9"} Feb 18 09:20:35 crc kubenswrapper[4556]: I0218 09:20:35.144263 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d105f935-cc14-4293-8101-17adaef31666","Type":"ContainerStarted","Data":"25907a65b937a0da29aafe30a8dbe3395adbfa2d720f0763871d25847732b52e"} Feb 18 09:20:35 crc kubenswrapper[4556]: I0218 09:20:35.164477 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.1644593 podStartE2EDuration="2.1644593s" podCreationTimestamp="2026-02-18 09:20:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:20:35.158537497 +0000 UTC m=+992.175498477" watchObservedRunningTime="2026-02-18 09:20:35.1644593 +0000 UTC m=+992.181420280" Feb 18 09:20:35 crc kubenswrapper[4556]: I0218 09:20:35.168238 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Feb 18 09:20:35 crc kubenswrapper[4556]: I0218 09:20:35.201989 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.36323316 podStartE2EDuration="5.201938157s" podCreationTimestamp="2026-02-18 09:20:30 +0000 UTC" firstStartedPulling="2026-02-18 09:20:30.873272245 +0000 UTC m=+987.890233226" lastFinishedPulling="2026-02-18 09:20:34.711977243 +0000 UTC m=+991.728938223" observedRunningTime="2026-02-18 09:20:35.190450562 +0000 UTC m=+992.207411541" watchObservedRunningTime="2026-02-18 09:20:35.201938157 +0000 UTC m=+992.218899136" Feb 18 09:20:35 crc kubenswrapper[4556]: I0218 09:20:35.345856 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-x62wg"] Feb 18 09:20:35 crc kubenswrapper[4556]: I0218 09:20:35.347194 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-x62wg" Feb 18 09:20:35 crc kubenswrapper[4556]: I0218 09:20:35.349006 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Feb 18 09:20:35 crc kubenswrapper[4556]: I0218 09:20:35.349438 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Feb 18 09:20:35 crc kubenswrapper[4556]: I0218 09:20:35.354735 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-x62wg"] Feb 18 09:20:35 crc kubenswrapper[4556]: I0218 09:20:35.377406 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0c20df4-63b1-46f0-9aca-d533c8ed6b8c-scripts\") pod \"nova-cell1-cell-mapping-x62wg\" (UID: \"d0c20df4-63b1-46f0-9aca-d533c8ed6b8c\") " pod="openstack/nova-cell1-cell-mapping-x62wg" Feb 18 09:20:35 crc kubenswrapper[4556]: I0218 09:20:35.377681 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0c20df4-63b1-46f0-9aca-d533c8ed6b8c-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-x62wg\" (UID: \"d0c20df4-63b1-46f0-9aca-d533c8ed6b8c\") " pod="openstack/nova-cell1-cell-mapping-x62wg" Feb 18 09:20:35 crc kubenswrapper[4556]: I0218 09:20:35.377858 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0c20df4-63b1-46f0-9aca-d533c8ed6b8c-config-data\") pod \"nova-cell1-cell-mapping-x62wg\" (UID: \"d0c20df4-63b1-46f0-9aca-d533c8ed6b8c\") " pod="openstack/nova-cell1-cell-mapping-x62wg" Feb 18 09:20:35 crc kubenswrapper[4556]: I0218 09:20:35.378235 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52zmj\" (UniqueName: \"kubernetes.io/projected/d0c20df4-63b1-46f0-9aca-d533c8ed6b8c-kube-api-access-52zmj\") pod \"nova-cell1-cell-mapping-x62wg\" (UID: \"d0c20df4-63b1-46f0-9aca-d533c8ed6b8c\") " pod="openstack/nova-cell1-cell-mapping-x62wg" Feb 18 09:20:35 crc kubenswrapper[4556]: I0218 09:20:35.398284 4556 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="1f3e12da-26fe-4d2a-ae62-9a3e90c5b752" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.199:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 18 09:20:35 crc kubenswrapper[4556]: I0218 09:20:35.398532 4556 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="1f3e12da-26fe-4d2a-ae62-9a3e90c5b752" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.199:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 18 09:20:35 crc kubenswrapper[4556]: I0218 09:20:35.480853 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0c20df4-63b1-46f0-9aca-d533c8ed6b8c-scripts\") pod \"nova-cell1-cell-mapping-x62wg\" (UID: \"d0c20df4-63b1-46f0-9aca-d533c8ed6b8c\") " pod="openstack/nova-cell1-cell-mapping-x62wg" Feb 18 09:20:35 crc kubenswrapper[4556]: I0218 09:20:35.480985 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0c20df4-63b1-46f0-9aca-d533c8ed6b8c-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-x62wg\" (UID: \"d0c20df4-63b1-46f0-9aca-d533c8ed6b8c\") " pod="openstack/nova-cell1-cell-mapping-x62wg" Feb 18 09:20:35 crc kubenswrapper[4556]: I0218 09:20:35.481070 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0c20df4-63b1-46f0-9aca-d533c8ed6b8c-config-data\") pod \"nova-cell1-cell-mapping-x62wg\" (UID: \"d0c20df4-63b1-46f0-9aca-d533c8ed6b8c\") " pod="openstack/nova-cell1-cell-mapping-x62wg" Feb 18 09:20:35 crc kubenswrapper[4556]: I0218 09:20:35.481269 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52zmj\" (UniqueName: \"kubernetes.io/projected/d0c20df4-63b1-46f0-9aca-d533c8ed6b8c-kube-api-access-52zmj\") pod \"nova-cell1-cell-mapping-x62wg\" (UID: \"d0c20df4-63b1-46f0-9aca-d533c8ed6b8c\") " pod="openstack/nova-cell1-cell-mapping-x62wg" Feb 18 09:20:35 crc kubenswrapper[4556]: I0218 09:20:35.487305 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0c20df4-63b1-46f0-9aca-d533c8ed6b8c-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-x62wg\" (UID: \"d0c20df4-63b1-46f0-9aca-d533c8ed6b8c\") " pod="openstack/nova-cell1-cell-mapping-x62wg" Feb 18 09:20:35 crc kubenswrapper[4556]: I0218 09:20:35.487507 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0c20df4-63b1-46f0-9aca-d533c8ed6b8c-scripts\") pod \"nova-cell1-cell-mapping-x62wg\" (UID: \"d0c20df4-63b1-46f0-9aca-d533c8ed6b8c\") " pod="openstack/nova-cell1-cell-mapping-x62wg" Feb 18 09:20:35 crc kubenswrapper[4556]: I0218 09:20:35.487807 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0c20df4-63b1-46f0-9aca-d533c8ed6b8c-config-data\") pod \"nova-cell1-cell-mapping-x62wg\" (UID: \"d0c20df4-63b1-46f0-9aca-d533c8ed6b8c\") " pod="openstack/nova-cell1-cell-mapping-x62wg" Feb 18 09:20:35 crc kubenswrapper[4556]: I0218 09:20:35.496334 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52zmj\" (UniqueName: \"kubernetes.io/projected/d0c20df4-63b1-46f0-9aca-d533c8ed6b8c-kube-api-access-52zmj\") pod \"nova-cell1-cell-mapping-x62wg\" (UID: \"d0c20df4-63b1-46f0-9aca-d533c8ed6b8c\") " pod="openstack/nova-cell1-cell-mapping-x62wg" Feb 18 09:20:35 crc kubenswrapper[4556]: I0218 09:20:35.668738 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-x62wg" Feb 18 09:20:36 crc kubenswrapper[4556]: I0218 09:20:36.100365 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-x62wg"] Feb 18 09:20:36 crc kubenswrapper[4556]: W0218 09:20:36.105729 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd0c20df4_63b1_46f0_9aca_d533c8ed6b8c.slice/crio-f673439ae653b59446569e20042058ce3e514762d49633462e5972c6746716d5 WatchSource:0}: Error finding container f673439ae653b59446569e20042058ce3e514762d49633462e5972c6746716d5: Status 404 returned error can't find the container with id f673439ae653b59446569e20042058ce3e514762d49633462e5972c6746716d5 Feb 18 09:20:36 crc kubenswrapper[4556]: I0218 09:20:36.153108 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-x62wg" event={"ID":"d0c20df4-63b1-46f0-9aca-d533c8ed6b8c","Type":"ContainerStarted","Data":"f673439ae653b59446569e20042058ce3e514762d49633462e5972c6746716d5"} Feb 18 09:20:36 crc kubenswrapper[4556]: I0218 09:20:36.153371 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 18 09:20:36 crc kubenswrapper[4556]: I0218 09:20:36.557303 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-74b99f9475-7n7kb" Feb 18 09:20:36 crc kubenswrapper[4556]: I0218 09:20:36.603697 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-868bc9dc59-827zj"] Feb 18 09:20:36 crc kubenswrapper[4556]: I0218 09:20:36.603904 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-868bc9dc59-827zj" podUID="5c2e5d6b-1b2f-4183-b654-908203d4b1d3" containerName="dnsmasq-dns" containerID="cri-o://557d6c6a6dd07fd9b0ef587eb11d08bfd8f3521c738d7bf4463bac456d3d22b0" gracePeriod=10 Feb 18 09:20:37 crc kubenswrapper[4556]: I0218 09:20:37.090672 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-868bc9dc59-827zj" Feb 18 09:20:37 crc kubenswrapper[4556]: I0218 09:20:37.116121 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5c2e5d6b-1b2f-4183-b654-908203d4b1d3-ovsdbserver-sb\") pod \"5c2e5d6b-1b2f-4183-b654-908203d4b1d3\" (UID: \"5c2e5d6b-1b2f-4183-b654-908203d4b1d3\") " Feb 18 09:20:37 crc kubenswrapper[4556]: I0218 09:20:37.116198 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5c2e5d6b-1b2f-4183-b654-908203d4b1d3-dns-svc\") pod \"5c2e5d6b-1b2f-4183-b654-908203d4b1d3\" (UID: \"5c2e5d6b-1b2f-4183-b654-908203d4b1d3\") " Feb 18 09:20:37 crc kubenswrapper[4556]: I0218 09:20:37.116264 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5c2e5d6b-1b2f-4183-b654-908203d4b1d3-ovsdbserver-nb\") pod \"5c2e5d6b-1b2f-4183-b654-908203d4b1d3\" (UID: \"5c2e5d6b-1b2f-4183-b654-908203d4b1d3\") " Feb 18 09:20:37 crc kubenswrapper[4556]: I0218 09:20:37.116305 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5c2e5d6b-1b2f-4183-b654-908203d4b1d3-dns-swift-storage-0\") pod \"5c2e5d6b-1b2f-4183-b654-908203d4b1d3\" (UID: \"5c2e5d6b-1b2f-4183-b654-908203d4b1d3\") " Feb 18 09:20:37 crc kubenswrapper[4556]: I0218 09:20:37.116354 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2n26g\" (UniqueName: \"kubernetes.io/projected/5c2e5d6b-1b2f-4183-b654-908203d4b1d3-kube-api-access-2n26g\") pod \"5c2e5d6b-1b2f-4183-b654-908203d4b1d3\" (UID: \"5c2e5d6b-1b2f-4183-b654-908203d4b1d3\") " Feb 18 09:20:37 crc kubenswrapper[4556]: I0218 09:20:37.116451 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c2e5d6b-1b2f-4183-b654-908203d4b1d3-config\") pod \"5c2e5d6b-1b2f-4183-b654-908203d4b1d3\" (UID: \"5c2e5d6b-1b2f-4183-b654-908203d4b1d3\") " Feb 18 09:20:37 crc kubenswrapper[4556]: I0218 09:20:37.127297 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c2e5d6b-1b2f-4183-b654-908203d4b1d3-kube-api-access-2n26g" (OuterVolumeSpecName: "kube-api-access-2n26g") pod "5c2e5d6b-1b2f-4183-b654-908203d4b1d3" (UID: "5c2e5d6b-1b2f-4183-b654-908203d4b1d3"). InnerVolumeSpecName "kube-api-access-2n26g". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:20:37 crc kubenswrapper[4556]: I0218 09:20:37.162551 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-x62wg" event={"ID":"d0c20df4-63b1-46f0-9aca-d533c8ed6b8c","Type":"ContainerStarted","Data":"d745e502212fd8d5ef69ed76fecf788d35f7191b8d1d71353d6a73a4736c0524"} Feb 18 09:20:37 crc kubenswrapper[4556]: I0218 09:20:37.164268 4556 generic.go:334] "Generic (PLEG): container finished" podID="5c2e5d6b-1b2f-4183-b654-908203d4b1d3" containerID="557d6c6a6dd07fd9b0ef587eb11d08bfd8f3521c738d7bf4463bac456d3d22b0" exitCode=0 Feb 18 09:20:37 crc kubenswrapper[4556]: I0218 09:20:37.164451 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-868bc9dc59-827zj" event={"ID":"5c2e5d6b-1b2f-4183-b654-908203d4b1d3","Type":"ContainerDied","Data":"557d6c6a6dd07fd9b0ef587eb11d08bfd8f3521c738d7bf4463bac456d3d22b0"} Feb 18 09:20:37 crc kubenswrapper[4556]: I0218 09:20:37.164510 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-868bc9dc59-827zj" event={"ID":"5c2e5d6b-1b2f-4183-b654-908203d4b1d3","Type":"ContainerDied","Data":"e8c414f28520923e6fc1777f21d7e17726a1e57d0e9c1fa2e64991e92ebe9c62"} Feb 18 09:20:37 crc kubenswrapper[4556]: I0218 09:20:37.164531 4556 scope.go:117] "RemoveContainer" containerID="557d6c6a6dd07fd9b0ef587eb11d08bfd8f3521c738d7bf4463bac456d3d22b0" Feb 18 09:20:37 crc kubenswrapper[4556]: I0218 09:20:37.164854 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-868bc9dc59-827zj" Feb 18 09:20:37 crc kubenswrapper[4556]: I0218 09:20:37.170720 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c2e5d6b-1b2f-4183-b654-908203d4b1d3-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5c2e5d6b-1b2f-4183-b654-908203d4b1d3" (UID: "5c2e5d6b-1b2f-4183-b654-908203d4b1d3"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:20:37 crc kubenswrapper[4556]: I0218 09:20:37.171219 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c2e5d6b-1b2f-4183-b654-908203d4b1d3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5c2e5d6b-1b2f-4183-b654-908203d4b1d3" (UID: "5c2e5d6b-1b2f-4183-b654-908203d4b1d3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:20:37 crc kubenswrapper[4556]: I0218 09:20:37.182810 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c2e5d6b-1b2f-4183-b654-908203d4b1d3-config" (OuterVolumeSpecName: "config") pod "5c2e5d6b-1b2f-4183-b654-908203d4b1d3" (UID: "5c2e5d6b-1b2f-4183-b654-908203d4b1d3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:20:37 crc kubenswrapper[4556]: I0218 09:20:37.183727 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c2e5d6b-1b2f-4183-b654-908203d4b1d3-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5c2e5d6b-1b2f-4183-b654-908203d4b1d3" (UID: "5c2e5d6b-1b2f-4183-b654-908203d4b1d3"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:20:37 crc kubenswrapper[4556]: I0218 09:20:37.184756 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-x62wg" podStartSLOduration=2.184741893 podStartE2EDuration="2.184741893s" podCreationTimestamp="2026-02-18 09:20:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:20:37.173886229 +0000 UTC m=+994.190847209" watchObservedRunningTime="2026-02-18 09:20:37.184741893 +0000 UTC m=+994.201702872" Feb 18 09:20:37 crc kubenswrapper[4556]: I0218 09:20:37.186698 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c2e5d6b-1b2f-4183-b654-908203d4b1d3-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "5c2e5d6b-1b2f-4183-b654-908203d4b1d3" (UID: "5c2e5d6b-1b2f-4183-b654-908203d4b1d3"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:20:37 crc kubenswrapper[4556]: I0218 09:20:37.190129 4556 scope.go:117] "RemoveContainer" containerID="bd038c2f8b47818e7f3f6eefba0f29d206ced733a70b887189020a9b0de9f33b" Feb 18 09:20:37 crc kubenswrapper[4556]: I0218 09:20:37.206140 4556 scope.go:117] "RemoveContainer" containerID="557d6c6a6dd07fd9b0ef587eb11d08bfd8f3521c738d7bf4463bac456d3d22b0" Feb 18 09:20:37 crc kubenswrapper[4556]: E0218 09:20:37.210653 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"557d6c6a6dd07fd9b0ef587eb11d08bfd8f3521c738d7bf4463bac456d3d22b0\": container with ID starting with 557d6c6a6dd07fd9b0ef587eb11d08bfd8f3521c738d7bf4463bac456d3d22b0 not found: ID does not exist" containerID="557d6c6a6dd07fd9b0ef587eb11d08bfd8f3521c738d7bf4463bac456d3d22b0" Feb 18 09:20:37 crc kubenswrapper[4556]: I0218 09:20:37.210713 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"557d6c6a6dd07fd9b0ef587eb11d08bfd8f3521c738d7bf4463bac456d3d22b0"} err="failed to get container status \"557d6c6a6dd07fd9b0ef587eb11d08bfd8f3521c738d7bf4463bac456d3d22b0\": rpc error: code = NotFound desc = could not find container \"557d6c6a6dd07fd9b0ef587eb11d08bfd8f3521c738d7bf4463bac456d3d22b0\": container with ID starting with 557d6c6a6dd07fd9b0ef587eb11d08bfd8f3521c738d7bf4463bac456d3d22b0 not found: ID does not exist" Feb 18 09:20:37 crc kubenswrapper[4556]: I0218 09:20:37.210738 4556 scope.go:117] "RemoveContainer" containerID="bd038c2f8b47818e7f3f6eefba0f29d206ced733a70b887189020a9b0de9f33b" Feb 18 09:20:37 crc kubenswrapper[4556]: E0218 09:20:37.211076 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd038c2f8b47818e7f3f6eefba0f29d206ced733a70b887189020a9b0de9f33b\": container with ID starting with bd038c2f8b47818e7f3f6eefba0f29d206ced733a70b887189020a9b0de9f33b not found: ID does not exist" containerID="bd038c2f8b47818e7f3f6eefba0f29d206ced733a70b887189020a9b0de9f33b" Feb 18 09:20:37 crc kubenswrapper[4556]: I0218 09:20:37.211135 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd038c2f8b47818e7f3f6eefba0f29d206ced733a70b887189020a9b0de9f33b"} err="failed to get container status \"bd038c2f8b47818e7f3f6eefba0f29d206ced733a70b887189020a9b0de9f33b\": rpc error: code = NotFound desc = could not find container \"bd038c2f8b47818e7f3f6eefba0f29d206ced733a70b887189020a9b0de9f33b\": container with ID starting with bd038c2f8b47818e7f3f6eefba0f29d206ced733a70b887189020a9b0de9f33b not found: ID does not exist" Feb 18 09:20:37 crc kubenswrapper[4556]: I0218 09:20:37.220295 4556 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5c2e5d6b-1b2f-4183-b654-908203d4b1d3-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 18 09:20:37 crc kubenswrapper[4556]: I0218 09:20:37.220319 4556 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5c2e5d6b-1b2f-4183-b654-908203d4b1d3-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 18 09:20:37 crc kubenswrapper[4556]: I0218 09:20:37.220331 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2n26g\" (UniqueName: \"kubernetes.io/projected/5c2e5d6b-1b2f-4183-b654-908203d4b1d3-kube-api-access-2n26g\") on node \"crc\" DevicePath \"\"" Feb 18 09:20:37 crc kubenswrapper[4556]: I0218 09:20:37.220341 4556 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c2e5d6b-1b2f-4183-b654-908203d4b1d3-config\") on node \"crc\" DevicePath \"\"" Feb 18 09:20:37 crc kubenswrapper[4556]: I0218 09:20:37.220352 4556 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5c2e5d6b-1b2f-4183-b654-908203d4b1d3-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 18 09:20:37 crc kubenswrapper[4556]: I0218 09:20:37.220360 4556 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5c2e5d6b-1b2f-4183-b654-908203d4b1d3-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 18 09:20:37 crc kubenswrapper[4556]: I0218 09:20:37.488096 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-868bc9dc59-827zj"] Feb 18 09:20:37 crc kubenswrapper[4556]: I0218 09:20:37.496721 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-868bc9dc59-827zj"] Feb 18 09:20:39 crc kubenswrapper[4556]: I0218 09:20:39.292912 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c2e5d6b-1b2f-4183-b654-908203d4b1d3" path="/var/lib/kubelet/pods/5c2e5d6b-1b2f-4183-b654-908203d4b1d3/volumes" Feb 18 09:20:40 crc kubenswrapper[4556]: I0218 09:20:40.190577 4556 generic.go:334] "Generic (PLEG): container finished" podID="d0c20df4-63b1-46f0-9aca-d533c8ed6b8c" containerID="d745e502212fd8d5ef69ed76fecf788d35f7191b8d1d71353d6a73a4736c0524" exitCode=0 Feb 18 09:20:40 crc kubenswrapper[4556]: I0218 09:20:40.190618 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-x62wg" event={"ID":"d0c20df4-63b1-46f0-9aca-d533c8ed6b8c","Type":"ContainerDied","Data":"d745e502212fd8d5ef69ed76fecf788d35f7191b8d1d71353d6a73a4736c0524"} Feb 18 09:20:41 crc kubenswrapper[4556]: I0218 09:20:41.590844 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-x62wg" Feb 18 09:20:41 crc kubenswrapper[4556]: I0218 09:20:41.709956 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0c20df4-63b1-46f0-9aca-d533c8ed6b8c-scripts\") pod \"d0c20df4-63b1-46f0-9aca-d533c8ed6b8c\" (UID: \"d0c20df4-63b1-46f0-9aca-d533c8ed6b8c\") " Feb 18 09:20:41 crc kubenswrapper[4556]: I0218 09:20:41.709999 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-52zmj\" (UniqueName: \"kubernetes.io/projected/d0c20df4-63b1-46f0-9aca-d533c8ed6b8c-kube-api-access-52zmj\") pod \"d0c20df4-63b1-46f0-9aca-d533c8ed6b8c\" (UID: \"d0c20df4-63b1-46f0-9aca-d533c8ed6b8c\") " Feb 18 09:20:41 crc kubenswrapper[4556]: I0218 09:20:41.710111 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0c20df4-63b1-46f0-9aca-d533c8ed6b8c-combined-ca-bundle\") pod \"d0c20df4-63b1-46f0-9aca-d533c8ed6b8c\" (UID: \"d0c20df4-63b1-46f0-9aca-d533c8ed6b8c\") " Feb 18 09:20:41 crc kubenswrapper[4556]: I0218 09:20:41.710246 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0c20df4-63b1-46f0-9aca-d533c8ed6b8c-config-data\") pod \"d0c20df4-63b1-46f0-9aca-d533c8ed6b8c\" (UID: \"d0c20df4-63b1-46f0-9aca-d533c8ed6b8c\") " Feb 18 09:20:41 crc kubenswrapper[4556]: I0218 09:20:41.716284 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0c20df4-63b1-46f0-9aca-d533c8ed6b8c-kube-api-access-52zmj" (OuterVolumeSpecName: "kube-api-access-52zmj") pod "d0c20df4-63b1-46f0-9aca-d533c8ed6b8c" (UID: "d0c20df4-63b1-46f0-9aca-d533c8ed6b8c"). InnerVolumeSpecName "kube-api-access-52zmj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:20:41 crc kubenswrapper[4556]: I0218 09:20:41.716314 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0c20df4-63b1-46f0-9aca-d533c8ed6b8c-scripts" (OuterVolumeSpecName: "scripts") pod "d0c20df4-63b1-46f0-9aca-d533c8ed6b8c" (UID: "d0c20df4-63b1-46f0-9aca-d533c8ed6b8c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:20:41 crc kubenswrapper[4556]: I0218 09:20:41.733977 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0c20df4-63b1-46f0-9aca-d533c8ed6b8c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d0c20df4-63b1-46f0-9aca-d533c8ed6b8c" (UID: "d0c20df4-63b1-46f0-9aca-d533c8ed6b8c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:20:41 crc kubenswrapper[4556]: I0218 09:20:41.735074 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0c20df4-63b1-46f0-9aca-d533c8ed6b8c-config-data" (OuterVolumeSpecName: "config-data") pod "d0c20df4-63b1-46f0-9aca-d533c8ed6b8c" (UID: "d0c20df4-63b1-46f0-9aca-d533c8ed6b8c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:20:41 crc kubenswrapper[4556]: I0218 09:20:41.812299 4556 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0c20df4-63b1-46f0-9aca-d533c8ed6b8c-scripts\") on node \"crc\" DevicePath \"\"" Feb 18 09:20:41 crc kubenswrapper[4556]: I0218 09:20:41.812390 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-52zmj\" (UniqueName: \"kubernetes.io/projected/d0c20df4-63b1-46f0-9aca-d533c8ed6b8c-kube-api-access-52zmj\") on node \"crc\" DevicePath \"\"" Feb 18 09:20:41 crc kubenswrapper[4556]: I0218 09:20:41.812528 4556 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0c20df4-63b1-46f0-9aca-d533c8ed6b8c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:20:41 crc kubenswrapper[4556]: I0218 09:20:41.812578 4556 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0c20df4-63b1-46f0-9aca-d533c8ed6b8c-config-data\") on node \"crc\" DevicePath \"\"" Feb 18 09:20:42 crc kubenswrapper[4556]: I0218 09:20:42.208626 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-x62wg" event={"ID":"d0c20df4-63b1-46f0-9aca-d533c8ed6b8c","Type":"ContainerDied","Data":"f673439ae653b59446569e20042058ce3e514762d49633462e5972c6746716d5"} Feb 18 09:20:42 crc kubenswrapper[4556]: I0218 09:20:42.208674 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f673439ae653b59446569e20042058ce3e514762d49633462e5972c6746716d5" Feb 18 09:20:42 crc kubenswrapper[4556]: I0218 09:20:42.208771 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-x62wg" Feb 18 09:20:42 crc kubenswrapper[4556]: I0218 09:20:42.377760 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 18 09:20:42 crc kubenswrapper[4556]: I0218 09:20:42.378046 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="87c7e959-f5b1-4d11-84d2-908268955e99" containerName="nova-api-log" containerID="cri-o://142b9b0dbdaf258de5293cb4fcb059abb419b6cc425a754422ba0e2e39a58bbc" gracePeriod=30 Feb 18 09:20:42 crc kubenswrapper[4556]: I0218 09:20:42.378191 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="87c7e959-f5b1-4d11-84d2-908268955e99" containerName="nova-api-api" containerID="cri-o://68d4fb9b5beb0194e57ac5290bb9f1fccea67a235b1b0abbace8316ee010a9e9" gracePeriod=30 Feb 18 09:20:42 crc kubenswrapper[4556]: I0218 09:20:42.438409 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 18 09:20:42 crc kubenswrapper[4556]: I0218 09:20:42.438712 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="5d34e777-73c9-4dd3-8b9a-1e6fc8f8aefa" containerName="nova-scheduler-scheduler" containerID="cri-o://e233eebe3d83a854854bdd8bd048694b15180c1bbabf93145a0c20a68d8c5e35" gracePeriod=30 Feb 18 09:20:42 crc kubenswrapper[4556]: I0218 09:20:42.441958 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 18 09:20:42 crc kubenswrapper[4556]: I0218 09:20:42.442609 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="1f3e12da-26fe-4d2a-ae62-9a3e90c5b752" containerName="nova-metadata-log" containerID="cri-o://a903312a63e39580644536734e7abd02a75c7922bd9cd0e7e5635d7c7d0e5fca" gracePeriod=30 Feb 18 09:20:42 crc kubenswrapper[4556]: I0218 09:20:42.442689 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="1f3e12da-26fe-4d2a-ae62-9a3e90c5b752" containerName="nova-metadata-metadata" containerID="cri-o://ea88e490cd9bdf056778cbda1d520395ed86855d37a5c3dec73dbb3079f8c795" gracePeriod=30 Feb 18 09:20:42 crc kubenswrapper[4556]: I0218 09:20:42.901306 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 18 09:20:42 crc kubenswrapper[4556]: I0218 09:20:42.945071 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s6xdk\" (UniqueName: \"kubernetes.io/projected/87c7e959-f5b1-4d11-84d2-908268955e99-kube-api-access-s6xdk\") pod \"87c7e959-f5b1-4d11-84d2-908268955e99\" (UID: \"87c7e959-f5b1-4d11-84d2-908268955e99\") " Feb 18 09:20:42 crc kubenswrapper[4556]: I0218 09:20:42.945204 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87c7e959-f5b1-4d11-84d2-908268955e99-logs\") pod \"87c7e959-f5b1-4d11-84d2-908268955e99\" (UID: \"87c7e959-f5b1-4d11-84d2-908268955e99\") " Feb 18 09:20:42 crc kubenswrapper[4556]: I0218 09:20:42.945346 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87c7e959-f5b1-4d11-84d2-908268955e99-combined-ca-bundle\") pod \"87c7e959-f5b1-4d11-84d2-908268955e99\" (UID: \"87c7e959-f5b1-4d11-84d2-908268955e99\") " Feb 18 09:20:42 crc kubenswrapper[4556]: I0218 09:20:42.945434 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/87c7e959-f5b1-4d11-84d2-908268955e99-public-tls-certs\") pod \"87c7e959-f5b1-4d11-84d2-908268955e99\" (UID: \"87c7e959-f5b1-4d11-84d2-908268955e99\") " Feb 18 09:20:42 crc kubenswrapper[4556]: I0218 09:20:42.945543 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/87c7e959-f5b1-4d11-84d2-908268955e99-internal-tls-certs\") pod \"87c7e959-f5b1-4d11-84d2-908268955e99\" (UID: \"87c7e959-f5b1-4d11-84d2-908268955e99\") " Feb 18 09:20:42 crc kubenswrapper[4556]: I0218 09:20:42.945679 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87c7e959-f5b1-4d11-84d2-908268955e99-config-data\") pod \"87c7e959-f5b1-4d11-84d2-908268955e99\" (UID: \"87c7e959-f5b1-4d11-84d2-908268955e99\") " Feb 18 09:20:42 crc kubenswrapper[4556]: I0218 09:20:42.945689 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87c7e959-f5b1-4d11-84d2-908268955e99-logs" (OuterVolumeSpecName: "logs") pod "87c7e959-f5b1-4d11-84d2-908268955e99" (UID: "87c7e959-f5b1-4d11-84d2-908268955e99"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:20:42 crc kubenswrapper[4556]: I0218 09:20:42.946814 4556 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87c7e959-f5b1-4d11-84d2-908268955e99-logs\") on node \"crc\" DevicePath \"\"" Feb 18 09:20:42 crc kubenswrapper[4556]: I0218 09:20:42.967134 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87c7e959-f5b1-4d11-84d2-908268955e99-kube-api-access-s6xdk" (OuterVolumeSpecName: "kube-api-access-s6xdk") pod "87c7e959-f5b1-4d11-84d2-908268955e99" (UID: "87c7e959-f5b1-4d11-84d2-908268955e99"). InnerVolumeSpecName "kube-api-access-s6xdk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:20:42 crc kubenswrapper[4556]: I0218 09:20:42.975381 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87c7e959-f5b1-4d11-84d2-908268955e99-config-data" (OuterVolumeSpecName: "config-data") pod "87c7e959-f5b1-4d11-84d2-908268955e99" (UID: "87c7e959-f5b1-4d11-84d2-908268955e99"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:20:42 crc kubenswrapper[4556]: I0218 09:20:42.982920 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87c7e959-f5b1-4d11-84d2-908268955e99-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "87c7e959-f5b1-4d11-84d2-908268955e99" (UID: "87c7e959-f5b1-4d11-84d2-908268955e99"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:20:42 crc kubenswrapper[4556]: E0218 09:20:42.999844 4556 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/87c7e959-f5b1-4d11-84d2-908268955e99-internal-tls-certs podName:87c7e959-f5b1-4d11-84d2-908268955e99 nodeName:}" failed. No retries permitted until 2026-02-18 09:20:43.499811718 +0000 UTC m=+1000.516772698 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "internal-tls-certs" (UniqueName: "kubernetes.io/secret/87c7e959-f5b1-4d11-84d2-908268955e99-internal-tls-certs") pod "87c7e959-f5b1-4d11-84d2-908268955e99" (UID: "87c7e959-f5b1-4d11-84d2-908268955e99") : error deleting /var/lib/kubelet/pods/87c7e959-f5b1-4d11-84d2-908268955e99/volume-subpaths: remove /var/lib/kubelet/pods/87c7e959-f5b1-4d11-84d2-908268955e99/volume-subpaths: no such file or directory Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.001959 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87c7e959-f5b1-4d11-84d2-908268955e99-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "87c7e959-f5b1-4d11-84d2-908268955e99" (UID: "87c7e959-f5b1-4d11-84d2-908268955e99"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.049469 4556 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87c7e959-f5b1-4d11-84d2-908268955e99-config-data\") on node \"crc\" DevicePath \"\"" Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.049495 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s6xdk\" (UniqueName: \"kubernetes.io/projected/87c7e959-f5b1-4d11-84d2-908268955e99-kube-api-access-s6xdk\") on node \"crc\" DevicePath \"\"" Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.049510 4556 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87c7e959-f5b1-4d11-84d2-908268955e99-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.049519 4556 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/87c7e959-f5b1-4d11-84d2-908268955e99-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.218563 4556 generic.go:334] "Generic (PLEG): container finished" podID="1f3e12da-26fe-4d2a-ae62-9a3e90c5b752" containerID="a903312a63e39580644536734e7abd02a75c7922bd9cd0e7e5635d7c7d0e5fca" exitCode=143 Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.218628 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1f3e12da-26fe-4d2a-ae62-9a3e90c5b752","Type":"ContainerDied","Data":"a903312a63e39580644536734e7abd02a75c7922bd9cd0e7e5635d7c7d0e5fca"} Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.220443 4556 generic.go:334] "Generic (PLEG): container finished" podID="87c7e959-f5b1-4d11-84d2-908268955e99" containerID="68d4fb9b5beb0194e57ac5290bb9f1fccea67a235b1b0abbace8316ee010a9e9" exitCode=0 Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.220471 4556 generic.go:334] "Generic (PLEG): container finished" podID="87c7e959-f5b1-4d11-84d2-908268955e99" containerID="142b9b0dbdaf258de5293cb4fcb059abb419b6cc425a754422ba0e2e39a58bbc" exitCode=143 Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.220490 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"87c7e959-f5b1-4d11-84d2-908268955e99","Type":"ContainerDied","Data":"68d4fb9b5beb0194e57ac5290bb9f1fccea67a235b1b0abbace8316ee010a9e9"} Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.220505 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"87c7e959-f5b1-4d11-84d2-908268955e99","Type":"ContainerDied","Data":"142b9b0dbdaf258de5293cb4fcb059abb419b6cc425a754422ba0e2e39a58bbc"} Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.220504 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.220526 4556 scope.go:117] "RemoveContainer" containerID="68d4fb9b5beb0194e57ac5290bb9f1fccea67a235b1b0abbace8316ee010a9e9" Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.220514 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"87c7e959-f5b1-4d11-84d2-908268955e99","Type":"ContainerDied","Data":"ea571bd28492885659f7750d9b6dd048aacbb1680da5fa29c14a75d6b35dc4da"} Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.243348 4556 scope.go:117] "RemoveContainer" containerID="142b9b0dbdaf258de5293cb4fcb059abb419b6cc425a754422ba0e2e39a58bbc" Feb 18 09:20:43 crc kubenswrapper[4556]: E0218 09:20:43.246873 4556 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5d34e777_73c9_4dd3_8b9a_1e6fc8f8aefa.slice/crio-e233eebe3d83a854854bdd8bd048694b15180c1bbabf93145a0c20a68d8c5e35.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5d34e777_73c9_4dd3_8b9a_1e6fc8f8aefa.slice/crio-conmon-e233eebe3d83a854854bdd8bd048694b15180c1bbabf93145a0c20a68d8c5e35.scope\": RecentStats: unable to find data in memory cache]" Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.270865 4556 scope.go:117] "RemoveContainer" containerID="68d4fb9b5beb0194e57ac5290bb9f1fccea67a235b1b0abbace8316ee010a9e9" Feb 18 09:20:43 crc kubenswrapper[4556]: E0218 09:20:43.271671 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68d4fb9b5beb0194e57ac5290bb9f1fccea67a235b1b0abbace8316ee010a9e9\": container with ID starting with 68d4fb9b5beb0194e57ac5290bb9f1fccea67a235b1b0abbace8316ee010a9e9 not found: ID does not exist" containerID="68d4fb9b5beb0194e57ac5290bb9f1fccea67a235b1b0abbace8316ee010a9e9" Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.271794 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68d4fb9b5beb0194e57ac5290bb9f1fccea67a235b1b0abbace8316ee010a9e9"} err="failed to get container status \"68d4fb9b5beb0194e57ac5290bb9f1fccea67a235b1b0abbace8316ee010a9e9\": rpc error: code = NotFound desc = could not find container \"68d4fb9b5beb0194e57ac5290bb9f1fccea67a235b1b0abbace8316ee010a9e9\": container with ID starting with 68d4fb9b5beb0194e57ac5290bb9f1fccea67a235b1b0abbace8316ee010a9e9 not found: ID does not exist" Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.271886 4556 scope.go:117] "RemoveContainer" containerID="142b9b0dbdaf258de5293cb4fcb059abb419b6cc425a754422ba0e2e39a58bbc" Feb 18 09:20:43 crc kubenswrapper[4556]: E0218 09:20:43.272269 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"142b9b0dbdaf258de5293cb4fcb059abb419b6cc425a754422ba0e2e39a58bbc\": container with ID starting with 142b9b0dbdaf258de5293cb4fcb059abb419b6cc425a754422ba0e2e39a58bbc not found: ID does not exist" containerID="142b9b0dbdaf258de5293cb4fcb059abb419b6cc425a754422ba0e2e39a58bbc" Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.272297 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"142b9b0dbdaf258de5293cb4fcb059abb419b6cc425a754422ba0e2e39a58bbc"} err="failed to get container status \"142b9b0dbdaf258de5293cb4fcb059abb419b6cc425a754422ba0e2e39a58bbc\": rpc error: code = NotFound desc = could not find container \"142b9b0dbdaf258de5293cb4fcb059abb419b6cc425a754422ba0e2e39a58bbc\": container with ID starting with 142b9b0dbdaf258de5293cb4fcb059abb419b6cc425a754422ba0e2e39a58bbc not found: ID does not exist" Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.272317 4556 scope.go:117] "RemoveContainer" containerID="68d4fb9b5beb0194e57ac5290bb9f1fccea67a235b1b0abbace8316ee010a9e9" Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.272587 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68d4fb9b5beb0194e57ac5290bb9f1fccea67a235b1b0abbace8316ee010a9e9"} err="failed to get container status \"68d4fb9b5beb0194e57ac5290bb9f1fccea67a235b1b0abbace8316ee010a9e9\": rpc error: code = NotFound desc = could not find container \"68d4fb9b5beb0194e57ac5290bb9f1fccea67a235b1b0abbace8316ee010a9e9\": container with ID starting with 68d4fb9b5beb0194e57ac5290bb9f1fccea67a235b1b0abbace8316ee010a9e9 not found: ID does not exist" Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.272672 4556 scope.go:117] "RemoveContainer" containerID="142b9b0dbdaf258de5293cb4fcb059abb419b6cc425a754422ba0e2e39a58bbc" Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.272968 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"142b9b0dbdaf258de5293cb4fcb059abb419b6cc425a754422ba0e2e39a58bbc"} err="failed to get container status \"142b9b0dbdaf258de5293cb4fcb059abb419b6cc425a754422ba0e2e39a58bbc\": rpc error: code = NotFound desc = could not find container \"142b9b0dbdaf258de5293cb4fcb059abb419b6cc425a754422ba0e2e39a58bbc\": container with ID starting with 142b9b0dbdaf258de5293cb4fcb059abb419b6cc425a754422ba0e2e39a58bbc not found: ID does not exist" Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.496353 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.560524 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/87c7e959-f5b1-4d11-84d2-908268955e99-internal-tls-certs\") pod \"87c7e959-f5b1-4d11-84d2-908268955e99\" (UID: \"87c7e959-f5b1-4d11-84d2-908268955e99\") " Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.560754 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d34e777-73c9-4dd3-8b9a-1e6fc8f8aefa-config-data\") pod \"5d34e777-73c9-4dd3-8b9a-1e6fc8f8aefa\" (UID: \"5d34e777-73c9-4dd3-8b9a-1e6fc8f8aefa\") " Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.560921 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d34e777-73c9-4dd3-8b9a-1e6fc8f8aefa-combined-ca-bundle\") pod \"5d34e777-73c9-4dd3-8b9a-1e6fc8f8aefa\" (UID: \"5d34e777-73c9-4dd3-8b9a-1e6fc8f8aefa\") " Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.561121 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pgvfb\" (UniqueName: \"kubernetes.io/projected/5d34e777-73c9-4dd3-8b9a-1e6fc8f8aefa-kube-api-access-pgvfb\") pod \"5d34e777-73c9-4dd3-8b9a-1e6fc8f8aefa\" (UID: \"5d34e777-73c9-4dd3-8b9a-1e6fc8f8aefa\") " Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.566300 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87c7e959-f5b1-4d11-84d2-908268955e99-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "87c7e959-f5b1-4d11-84d2-908268955e99" (UID: "87c7e959-f5b1-4d11-84d2-908268955e99"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.568101 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d34e777-73c9-4dd3-8b9a-1e6fc8f8aefa-kube-api-access-pgvfb" (OuterVolumeSpecName: "kube-api-access-pgvfb") pod "5d34e777-73c9-4dd3-8b9a-1e6fc8f8aefa" (UID: "5d34e777-73c9-4dd3-8b9a-1e6fc8f8aefa"). InnerVolumeSpecName "kube-api-access-pgvfb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.583456 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d34e777-73c9-4dd3-8b9a-1e6fc8f8aefa-config-data" (OuterVolumeSpecName: "config-data") pod "5d34e777-73c9-4dd3-8b9a-1e6fc8f8aefa" (UID: "5d34e777-73c9-4dd3-8b9a-1e6fc8f8aefa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.589075 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d34e777-73c9-4dd3-8b9a-1e6fc8f8aefa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5d34e777-73c9-4dd3-8b9a-1e6fc8f8aefa" (UID: "5d34e777-73c9-4dd3-8b9a-1e6fc8f8aefa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.663850 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pgvfb\" (UniqueName: \"kubernetes.io/projected/5d34e777-73c9-4dd3-8b9a-1e6fc8f8aefa-kube-api-access-pgvfb\") on node \"crc\" DevicePath \"\"" Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.663871 4556 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/87c7e959-f5b1-4d11-84d2-908268955e99-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.663880 4556 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d34e777-73c9-4dd3-8b9a-1e6fc8f8aefa-config-data\") on node \"crc\" DevicePath \"\"" Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.663892 4556 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d34e777-73c9-4dd3-8b9a-1e6fc8f8aefa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.848508 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.855568 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.879353 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 18 09:20:43 crc kubenswrapper[4556]: E0218 09:20:43.879883 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87c7e959-f5b1-4d11-84d2-908268955e99" containerName="nova-api-api" Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.879903 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="87c7e959-f5b1-4d11-84d2-908268955e99" containerName="nova-api-api" Feb 18 09:20:43 crc kubenswrapper[4556]: E0218 09:20:43.879917 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c2e5d6b-1b2f-4183-b654-908203d4b1d3" containerName="dnsmasq-dns" Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.879924 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c2e5d6b-1b2f-4183-b654-908203d4b1d3" containerName="dnsmasq-dns" Feb 18 09:20:43 crc kubenswrapper[4556]: E0218 09:20:43.879959 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0c20df4-63b1-46f0-9aca-d533c8ed6b8c" containerName="nova-manage" Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.879965 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0c20df4-63b1-46f0-9aca-d533c8ed6b8c" containerName="nova-manage" Feb 18 09:20:43 crc kubenswrapper[4556]: E0218 09:20:43.879977 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c2e5d6b-1b2f-4183-b654-908203d4b1d3" containerName="init" Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.879983 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c2e5d6b-1b2f-4183-b654-908203d4b1d3" containerName="init" Feb 18 09:20:43 crc kubenswrapper[4556]: E0218 09:20:43.880008 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d34e777-73c9-4dd3-8b9a-1e6fc8f8aefa" containerName="nova-scheduler-scheduler" Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.880014 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d34e777-73c9-4dd3-8b9a-1e6fc8f8aefa" containerName="nova-scheduler-scheduler" Feb 18 09:20:43 crc kubenswrapper[4556]: E0218 09:20:43.880035 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87c7e959-f5b1-4d11-84d2-908268955e99" containerName="nova-api-log" Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.880052 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="87c7e959-f5b1-4d11-84d2-908268955e99" containerName="nova-api-log" Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.880314 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="87c7e959-f5b1-4d11-84d2-908268955e99" containerName="nova-api-log" Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.880356 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c2e5d6b-1b2f-4183-b654-908203d4b1d3" containerName="dnsmasq-dns" Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.880374 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d34e777-73c9-4dd3-8b9a-1e6fc8f8aefa" containerName="nova-scheduler-scheduler" Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.880382 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="87c7e959-f5b1-4d11-84d2-908268955e99" containerName="nova-api-api" Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.880395 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0c20df4-63b1-46f0-9aca-d533c8ed6b8c" containerName="nova-manage" Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.881570 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.883599 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.883941 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.884873 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.899555 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.969198 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e997362-15c5-4052-bf0b-8491c057f317-config-data\") pod \"nova-api-0\" (UID: \"4e997362-15c5-4052-bf0b-8491c057f317\") " pod="openstack/nova-api-0" Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.969484 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e997362-15c5-4052-bf0b-8491c057f317-internal-tls-certs\") pod \"nova-api-0\" (UID: \"4e997362-15c5-4052-bf0b-8491c057f317\") " pod="openstack/nova-api-0" Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.969512 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e997362-15c5-4052-bf0b-8491c057f317-public-tls-certs\") pod \"nova-api-0\" (UID: \"4e997362-15c5-4052-bf0b-8491c057f317\") " pod="openstack/nova-api-0" Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.969547 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4e997362-15c5-4052-bf0b-8491c057f317-logs\") pod \"nova-api-0\" (UID: \"4e997362-15c5-4052-bf0b-8491c057f317\") " pod="openstack/nova-api-0" Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.969571 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8bzg\" (UniqueName: \"kubernetes.io/projected/4e997362-15c5-4052-bf0b-8491c057f317-kube-api-access-d8bzg\") pod \"nova-api-0\" (UID: \"4e997362-15c5-4052-bf0b-8491c057f317\") " pod="openstack/nova-api-0" Feb 18 09:20:43 crc kubenswrapper[4556]: I0218 09:20:43.969611 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e997362-15c5-4052-bf0b-8491c057f317-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4e997362-15c5-4052-bf0b-8491c057f317\") " pod="openstack/nova-api-0" Feb 18 09:20:44 crc kubenswrapper[4556]: I0218 09:20:44.071736 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e997362-15c5-4052-bf0b-8491c057f317-config-data\") pod \"nova-api-0\" (UID: \"4e997362-15c5-4052-bf0b-8491c057f317\") " pod="openstack/nova-api-0" Feb 18 09:20:44 crc kubenswrapper[4556]: I0218 09:20:44.071855 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e997362-15c5-4052-bf0b-8491c057f317-internal-tls-certs\") pod \"nova-api-0\" (UID: \"4e997362-15c5-4052-bf0b-8491c057f317\") " pod="openstack/nova-api-0" Feb 18 09:20:44 crc kubenswrapper[4556]: I0218 09:20:44.071888 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e997362-15c5-4052-bf0b-8491c057f317-public-tls-certs\") pod \"nova-api-0\" (UID: \"4e997362-15c5-4052-bf0b-8491c057f317\") " pod="openstack/nova-api-0" Feb 18 09:20:44 crc kubenswrapper[4556]: I0218 09:20:44.071976 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4e997362-15c5-4052-bf0b-8491c057f317-logs\") pod \"nova-api-0\" (UID: \"4e997362-15c5-4052-bf0b-8491c057f317\") " pod="openstack/nova-api-0" Feb 18 09:20:44 crc kubenswrapper[4556]: I0218 09:20:44.072057 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8bzg\" (UniqueName: \"kubernetes.io/projected/4e997362-15c5-4052-bf0b-8491c057f317-kube-api-access-d8bzg\") pod \"nova-api-0\" (UID: \"4e997362-15c5-4052-bf0b-8491c057f317\") " pod="openstack/nova-api-0" Feb 18 09:20:44 crc kubenswrapper[4556]: I0218 09:20:44.072223 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e997362-15c5-4052-bf0b-8491c057f317-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4e997362-15c5-4052-bf0b-8491c057f317\") " pod="openstack/nova-api-0" Feb 18 09:20:44 crc kubenswrapper[4556]: I0218 09:20:44.072647 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4e997362-15c5-4052-bf0b-8491c057f317-logs\") pod \"nova-api-0\" (UID: \"4e997362-15c5-4052-bf0b-8491c057f317\") " pod="openstack/nova-api-0" Feb 18 09:20:44 crc kubenswrapper[4556]: I0218 09:20:44.077023 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e997362-15c5-4052-bf0b-8491c057f317-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4e997362-15c5-4052-bf0b-8491c057f317\") " pod="openstack/nova-api-0" Feb 18 09:20:44 crc kubenswrapper[4556]: I0218 09:20:44.077576 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e997362-15c5-4052-bf0b-8491c057f317-public-tls-certs\") pod \"nova-api-0\" (UID: \"4e997362-15c5-4052-bf0b-8491c057f317\") " pod="openstack/nova-api-0" Feb 18 09:20:44 crc kubenswrapper[4556]: I0218 09:20:44.077855 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e997362-15c5-4052-bf0b-8491c057f317-internal-tls-certs\") pod \"nova-api-0\" (UID: \"4e997362-15c5-4052-bf0b-8491c057f317\") " pod="openstack/nova-api-0" Feb 18 09:20:44 crc kubenswrapper[4556]: I0218 09:20:44.078019 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e997362-15c5-4052-bf0b-8491c057f317-config-data\") pod \"nova-api-0\" (UID: \"4e997362-15c5-4052-bf0b-8491c057f317\") " pod="openstack/nova-api-0" Feb 18 09:20:44 crc kubenswrapper[4556]: I0218 09:20:44.087112 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8bzg\" (UniqueName: \"kubernetes.io/projected/4e997362-15c5-4052-bf0b-8491c057f317-kube-api-access-d8bzg\") pod \"nova-api-0\" (UID: \"4e997362-15c5-4052-bf0b-8491c057f317\") " pod="openstack/nova-api-0" Feb 18 09:20:44 crc kubenswrapper[4556]: I0218 09:20:44.210162 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 18 09:20:44 crc kubenswrapper[4556]: I0218 09:20:44.240441 4556 generic.go:334] "Generic (PLEG): container finished" podID="5d34e777-73c9-4dd3-8b9a-1e6fc8f8aefa" containerID="e233eebe3d83a854854bdd8bd048694b15180c1bbabf93145a0c20a68d8c5e35" exitCode=0 Feb 18 09:20:44 crc kubenswrapper[4556]: I0218 09:20:44.240488 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"5d34e777-73c9-4dd3-8b9a-1e6fc8f8aefa","Type":"ContainerDied","Data":"e233eebe3d83a854854bdd8bd048694b15180c1bbabf93145a0c20a68d8c5e35"} Feb 18 09:20:44 crc kubenswrapper[4556]: I0218 09:20:44.240516 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"5d34e777-73c9-4dd3-8b9a-1e6fc8f8aefa","Type":"ContainerDied","Data":"111bc0dcb3d47b87ca4e5a40e5ddbbdce67639d8fd5e1088f438944ec5f72af9"} Feb 18 09:20:44 crc kubenswrapper[4556]: I0218 09:20:44.240525 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 18 09:20:44 crc kubenswrapper[4556]: I0218 09:20:44.240535 4556 scope.go:117] "RemoveContainer" containerID="e233eebe3d83a854854bdd8bd048694b15180c1bbabf93145a0c20a68d8c5e35" Feb 18 09:20:44 crc kubenswrapper[4556]: I0218 09:20:44.320948 4556 scope.go:117] "RemoveContainer" containerID="e233eebe3d83a854854bdd8bd048694b15180c1bbabf93145a0c20a68d8c5e35" Feb 18 09:20:44 crc kubenswrapper[4556]: I0218 09:20:44.324366 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 18 09:20:44 crc kubenswrapper[4556]: E0218 09:20:44.326478 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e233eebe3d83a854854bdd8bd048694b15180c1bbabf93145a0c20a68d8c5e35\": container with ID starting with e233eebe3d83a854854bdd8bd048694b15180c1bbabf93145a0c20a68d8c5e35 not found: ID does not exist" containerID="e233eebe3d83a854854bdd8bd048694b15180c1bbabf93145a0c20a68d8c5e35" Feb 18 09:20:44 crc kubenswrapper[4556]: I0218 09:20:44.326601 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e233eebe3d83a854854bdd8bd048694b15180c1bbabf93145a0c20a68d8c5e35"} err="failed to get container status \"e233eebe3d83a854854bdd8bd048694b15180c1bbabf93145a0c20a68d8c5e35\": rpc error: code = NotFound desc = could not find container \"e233eebe3d83a854854bdd8bd048694b15180c1bbabf93145a0c20a68d8c5e35\": container with ID starting with e233eebe3d83a854854bdd8bd048694b15180c1bbabf93145a0c20a68d8c5e35 not found: ID does not exist" Feb 18 09:20:44 crc kubenswrapper[4556]: I0218 09:20:44.344812 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Feb 18 09:20:44 crc kubenswrapper[4556]: I0218 09:20:44.362610 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Feb 18 09:20:44 crc kubenswrapper[4556]: I0218 09:20:44.364373 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 18 09:20:44 crc kubenswrapper[4556]: I0218 09:20:44.366713 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Feb 18 09:20:44 crc kubenswrapper[4556]: I0218 09:20:44.370104 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 18 09:20:44 crc kubenswrapper[4556]: I0218 09:20:44.481308 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db4d9035-57ee-4c5f-9677-cc2174644152-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"db4d9035-57ee-4c5f-9677-cc2174644152\") " pod="openstack/nova-scheduler-0" Feb 18 09:20:44 crc kubenswrapper[4556]: I0218 09:20:44.481419 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g67n9\" (UniqueName: \"kubernetes.io/projected/db4d9035-57ee-4c5f-9677-cc2174644152-kube-api-access-g67n9\") pod \"nova-scheduler-0\" (UID: \"db4d9035-57ee-4c5f-9677-cc2174644152\") " pod="openstack/nova-scheduler-0" Feb 18 09:20:44 crc kubenswrapper[4556]: I0218 09:20:44.481666 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db4d9035-57ee-4c5f-9677-cc2174644152-config-data\") pod \"nova-scheduler-0\" (UID: \"db4d9035-57ee-4c5f-9677-cc2174644152\") " pod="openstack/nova-scheduler-0" Feb 18 09:20:44 crc kubenswrapper[4556]: I0218 09:20:44.583202 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g67n9\" (UniqueName: \"kubernetes.io/projected/db4d9035-57ee-4c5f-9677-cc2174644152-kube-api-access-g67n9\") pod \"nova-scheduler-0\" (UID: \"db4d9035-57ee-4c5f-9677-cc2174644152\") " pod="openstack/nova-scheduler-0" Feb 18 09:20:44 crc kubenswrapper[4556]: I0218 09:20:44.583463 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db4d9035-57ee-4c5f-9677-cc2174644152-config-data\") pod \"nova-scheduler-0\" (UID: \"db4d9035-57ee-4c5f-9677-cc2174644152\") " pod="openstack/nova-scheduler-0" Feb 18 09:20:44 crc kubenswrapper[4556]: I0218 09:20:44.583558 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db4d9035-57ee-4c5f-9677-cc2174644152-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"db4d9035-57ee-4c5f-9677-cc2174644152\") " pod="openstack/nova-scheduler-0" Feb 18 09:20:44 crc kubenswrapper[4556]: I0218 09:20:44.588972 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db4d9035-57ee-4c5f-9677-cc2174644152-config-data\") pod \"nova-scheduler-0\" (UID: \"db4d9035-57ee-4c5f-9677-cc2174644152\") " pod="openstack/nova-scheduler-0" Feb 18 09:20:44 crc kubenswrapper[4556]: I0218 09:20:44.589899 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db4d9035-57ee-4c5f-9677-cc2174644152-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"db4d9035-57ee-4c5f-9677-cc2174644152\") " pod="openstack/nova-scheduler-0" Feb 18 09:20:44 crc kubenswrapper[4556]: I0218 09:20:44.600639 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g67n9\" (UniqueName: \"kubernetes.io/projected/db4d9035-57ee-4c5f-9677-cc2174644152-kube-api-access-g67n9\") pod \"nova-scheduler-0\" (UID: \"db4d9035-57ee-4c5f-9677-cc2174644152\") " pod="openstack/nova-scheduler-0" Feb 18 09:20:44 crc kubenswrapper[4556]: I0218 09:20:44.644475 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 18 09:20:44 crc kubenswrapper[4556]: W0218 09:20:44.646271 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4e997362_15c5_4052_bf0b_8491c057f317.slice/crio-e9931239004b3e767a565bffca543d772382a161f9201675dd68cf84c4cf1649 WatchSource:0}: Error finding container e9931239004b3e767a565bffca543d772382a161f9201675dd68cf84c4cf1649: Status 404 returned error can't find the container with id e9931239004b3e767a565bffca543d772382a161f9201675dd68cf84c4cf1649 Feb 18 09:20:44 crc kubenswrapper[4556]: I0218 09:20:44.683941 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 18 09:20:45 crc kubenswrapper[4556]: I0218 09:20:45.059852 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 18 09:20:45 crc kubenswrapper[4556]: W0218 09:20:45.067199 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddb4d9035_57ee_4c5f_9677_cc2174644152.slice/crio-cad5936da2208737ae182529a7dde17bfd522757360998bca6a7ee940007a0de WatchSource:0}: Error finding container cad5936da2208737ae182529a7dde17bfd522757360998bca6a7ee940007a0de: Status 404 returned error can't find the container with id cad5936da2208737ae182529a7dde17bfd522757360998bca6a7ee940007a0de Feb 18 09:20:45 crc kubenswrapper[4556]: I0218 09:20:45.252560 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4e997362-15c5-4052-bf0b-8491c057f317","Type":"ContainerStarted","Data":"7ad5cb02cca27f4c169529f3572cfe558d81ae87afe92d395b686e319ddbba79"} Feb 18 09:20:45 crc kubenswrapper[4556]: I0218 09:20:45.252624 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4e997362-15c5-4052-bf0b-8491c057f317","Type":"ContainerStarted","Data":"1169a8133b30bf18e40b90e916a4a56c3f8e11b05b282e38141861e44fb4491f"} Feb 18 09:20:45 crc kubenswrapper[4556]: I0218 09:20:45.252637 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4e997362-15c5-4052-bf0b-8491c057f317","Type":"ContainerStarted","Data":"e9931239004b3e767a565bffca543d772382a161f9201675dd68cf84c4cf1649"} Feb 18 09:20:45 crc kubenswrapper[4556]: I0218 09:20:45.254141 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"db4d9035-57ee-4c5f-9677-cc2174644152","Type":"ContainerStarted","Data":"16a96a6c5565a032bb24a50eff498f16c80f887a12a80a41403bc7b5599cf917"} Feb 18 09:20:45 crc kubenswrapper[4556]: I0218 09:20:45.254194 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"db4d9035-57ee-4c5f-9677-cc2174644152","Type":"ContainerStarted","Data":"cad5936da2208737ae182529a7dde17bfd522757360998bca6a7ee940007a0de"} Feb 18 09:20:45 crc kubenswrapper[4556]: I0218 09:20:45.273791 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.273760835 podStartE2EDuration="2.273760835s" podCreationTimestamp="2026-02-18 09:20:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:20:45.273271663 +0000 UTC m=+1002.290232663" watchObservedRunningTime="2026-02-18 09:20:45.273760835 +0000 UTC m=+1002.290721815" Feb 18 09:20:45 crc kubenswrapper[4556]: I0218 09:20:45.292462 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d34e777-73c9-4dd3-8b9a-1e6fc8f8aefa" path="/var/lib/kubelet/pods/5d34e777-73c9-4dd3-8b9a-1e6fc8f8aefa/volumes" Feb 18 09:20:45 crc kubenswrapper[4556]: I0218 09:20:45.293718 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87c7e959-f5b1-4d11-84d2-908268955e99" path="/var/lib/kubelet/pods/87c7e959-f5b1-4d11-84d2-908268955e99/volumes" Feb 18 09:20:45 crc kubenswrapper[4556]: I0218 09:20:45.303741 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=1.303720223 podStartE2EDuration="1.303720223s" podCreationTimestamp="2026-02-18 09:20:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:20:45.292939482 +0000 UTC m=+1002.309900462" watchObservedRunningTime="2026-02-18 09:20:45.303720223 +0000 UTC m=+1002.320681203" Feb 18 09:20:45 crc kubenswrapper[4556]: I0218 09:20:45.899386 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 18 09:20:46 crc kubenswrapper[4556]: I0218 09:20:46.018460 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f3e12da-26fe-4d2a-ae62-9a3e90c5b752-logs\") pod \"1f3e12da-26fe-4d2a-ae62-9a3e90c5b752\" (UID: \"1f3e12da-26fe-4d2a-ae62-9a3e90c5b752\") " Feb 18 09:20:46 crc kubenswrapper[4556]: I0218 09:20:46.018498 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f3e12da-26fe-4d2a-ae62-9a3e90c5b752-nova-metadata-tls-certs\") pod \"1f3e12da-26fe-4d2a-ae62-9a3e90c5b752\" (UID: \"1f3e12da-26fe-4d2a-ae62-9a3e90c5b752\") " Feb 18 09:20:46 crc kubenswrapper[4556]: I0218 09:20:46.018538 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f3e12da-26fe-4d2a-ae62-9a3e90c5b752-config-data\") pod \"1f3e12da-26fe-4d2a-ae62-9a3e90c5b752\" (UID: \"1f3e12da-26fe-4d2a-ae62-9a3e90c5b752\") " Feb 18 09:20:46 crc kubenswrapper[4556]: I0218 09:20:46.018570 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5jczm\" (UniqueName: \"kubernetes.io/projected/1f3e12da-26fe-4d2a-ae62-9a3e90c5b752-kube-api-access-5jczm\") pod \"1f3e12da-26fe-4d2a-ae62-9a3e90c5b752\" (UID: \"1f3e12da-26fe-4d2a-ae62-9a3e90c5b752\") " Feb 18 09:20:46 crc kubenswrapper[4556]: I0218 09:20:46.018592 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f3e12da-26fe-4d2a-ae62-9a3e90c5b752-combined-ca-bundle\") pod \"1f3e12da-26fe-4d2a-ae62-9a3e90c5b752\" (UID: \"1f3e12da-26fe-4d2a-ae62-9a3e90c5b752\") " Feb 18 09:20:46 crc kubenswrapper[4556]: I0218 09:20:46.019052 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f3e12da-26fe-4d2a-ae62-9a3e90c5b752-logs" (OuterVolumeSpecName: "logs") pod "1f3e12da-26fe-4d2a-ae62-9a3e90c5b752" (UID: "1f3e12da-26fe-4d2a-ae62-9a3e90c5b752"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:20:46 crc kubenswrapper[4556]: I0218 09:20:46.028482 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f3e12da-26fe-4d2a-ae62-9a3e90c5b752-kube-api-access-5jczm" (OuterVolumeSpecName: "kube-api-access-5jczm") pod "1f3e12da-26fe-4d2a-ae62-9a3e90c5b752" (UID: "1f3e12da-26fe-4d2a-ae62-9a3e90c5b752"). InnerVolumeSpecName "kube-api-access-5jczm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:20:46 crc kubenswrapper[4556]: I0218 09:20:46.049930 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f3e12da-26fe-4d2a-ae62-9a3e90c5b752-config-data" (OuterVolumeSpecName: "config-data") pod "1f3e12da-26fe-4d2a-ae62-9a3e90c5b752" (UID: "1f3e12da-26fe-4d2a-ae62-9a3e90c5b752"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:20:46 crc kubenswrapper[4556]: I0218 09:20:46.050341 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f3e12da-26fe-4d2a-ae62-9a3e90c5b752-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1f3e12da-26fe-4d2a-ae62-9a3e90c5b752" (UID: "1f3e12da-26fe-4d2a-ae62-9a3e90c5b752"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:20:46 crc kubenswrapper[4556]: I0218 09:20:46.067342 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f3e12da-26fe-4d2a-ae62-9a3e90c5b752-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "1f3e12da-26fe-4d2a-ae62-9a3e90c5b752" (UID: "1f3e12da-26fe-4d2a-ae62-9a3e90c5b752"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:20:46 crc kubenswrapper[4556]: I0218 09:20:46.121580 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5jczm\" (UniqueName: \"kubernetes.io/projected/1f3e12da-26fe-4d2a-ae62-9a3e90c5b752-kube-api-access-5jczm\") on node \"crc\" DevicePath \"\"" Feb 18 09:20:46 crc kubenswrapper[4556]: I0218 09:20:46.121611 4556 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f3e12da-26fe-4d2a-ae62-9a3e90c5b752-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:20:46 crc kubenswrapper[4556]: I0218 09:20:46.121694 4556 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f3e12da-26fe-4d2a-ae62-9a3e90c5b752-logs\") on node \"crc\" DevicePath \"\"" Feb 18 09:20:46 crc kubenswrapper[4556]: I0218 09:20:46.121704 4556 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f3e12da-26fe-4d2a-ae62-9a3e90c5b752-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 18 09:20:46 crc kubenswrapper[4556]: I0218 09:20:46.121712 4556 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f3e12da-26fe-4d2a-ae62-9a3e90c5b752-config-data\") on node \"crc\" DevicePath \"\"" Feb 18 09:20:46 crc kubenswrapper[4556]: I0218 09:20:46.274788 4556 generic.go:334] "Generic (PLEG): container finished" podID="1f3e12da-26fe-4d2a-ae62-9a3e90c5b752" containerID="ea88e490cd9bdf056778cbda1d520395ed86855d37a5c3dec73dbb3079f8c795" exitCode=0 Feb 18 09:20:46 crc kubenswrapper[4556]: I0218 09:20:46.275598 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 18 09:20:46 crc kubenswrapper[4556]: I0218 09:20:46.277014 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1f3e12da-26fe-4d2a-ae62-9a3e90c5b752","Type":"ContainerDied","Data":"ea88e490cd9bdf056778cbda1d520395ed86855d37a5c3dec73dbb3079f8c795"} Feb 18 09:20:46 crc kubenswrapper[4556]: I0218 09:20:46.279141 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1f3e12da-26fe-4d2a-ae62-9a3e90c5b752","Type":"ContainerDied","Data":"b699a8125f463362973c934736b8ceeaefeb9673fd94c80f14929f68464a2e7d"} Feb 18 09:20:46 crc kubenswrapper[4556]: I0218 09:20:46.279201 4556 scope.go:117] "RemoveContainer" containerID="ea88e490cd9bdf056778cbda1d520395ed86855d37a5c3dec73dbb3079f8c795" Feb 18 09:20:46 crc kubenswrapper[4556]: I0218 09:20:46.314495 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 18 09:20:46 crc kubenswrapper[4556]: I0218 09:20:46.320871 4556 scope.go:117] "RemoveContainer" containerID="a903312a63e39580644536734e7abd02a75c7922bd9cd0e7e5635d7c7d0e5fca" Feb 18 09:20:46 crc kubenswrapper[4556]: I0218 09:20:46.326588 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Feb 18 09:20:46 crc kubenswrapper[4556]: I0218 09:20:46.336576 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 18 09:20:46 crc kubenswrapper[4556]: E0218 09:20:46.337064 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f3e12da-26fe-4d2a-ae62-9a3e90c5b752" containerName="nova-metadata-metadata" Feb 18 09:20:46 crc kubenswrapper[4556]: I0218 09:20:46.337084 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f3e12da-26fe-4d2a-ae62-9a3e90c5b752" containerName="nova-metadata-metadata" Feb 18 09:20:46 crc kubenswrapper[4556]: E0218 09:20:46.337127 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f3e12da-26fe-4d2a-ae62-9a3e90c5b752" containerName="nova-metadata-log" Feb 18 09:20:46 crc kubenswrapper[4556]: I0218 09:20:46.337136 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f3e12da-26fe-4d2a-ae62-9a3e90c5b752" containerName="nova-metadata-log" Feb 18 09:20:46 crc kubenswrapper[4556]: I0218 09:20:46.337365 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f3e12da-26fe-4d2a-ae62-9a3e90c5b752" containerName="nova-metadata-log" Feb 18 09:20:46 crc kubenswrapper[4556]: I0218 09:20:46.337391 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f3e12da-26fe-4d2a-ae62-9a3e90c5b752" containerName="nova-metadata-metadata" Feb 18 09:20:46 crc kubenswrapper[4556]: I0218 09:20:46.338422 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 18 09:20:46 crc kubenswrapper[4556]: I0218 09:20:46.341539 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Feb 18 09:20:46 crc kubenswrapper[4556]: I0218 09:20:46.341760 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 18 09:20:46 crc kubenswrapper[4556]: I0218 09:20:46.343710 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 18 09:20:46 crc kubenswrapper[4556]: I0218 09:20:46.369493 4556 scope.go:117] "RemoveContainer" containerID="ea88e490cd9bdf056778cbda1d520395ed86855d37a5c3dec73dbb3079f8c795" Feb 18 09:20:46 crc kubenswrapper[4556]: E0218 09:20:46.370692 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea88e490cd9bdf056778cbda1d520395ed86855d37a5c3dec73dbb3079f8c795\": container with ID starting with ea88e490cd9bdf056778cbda1d520395ed86855d37a5c3dec73dbb3079f8c795 not found: ID does not exist" containerID="ea88e490cd9bdf056778cbda1d520395ed86855d37a5c3dec73dbb3079f8c795" Feb 18 09:20:46 crc kubenswrapper[4556]: I0218 09:20:46.370745 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea88e490cd9bdf056778cbda1d520395ed86855d37a5c3dec73dbb3079f8c795"} err="failed to get container status \"ea88e490cd9bdf056778cbda1d520395ed86855d37a5c3dec73dbb3079f8c795\": rpc error: code = NotFound desc = could not find container \"ea88e490cd9bdf056778cbda1d520395ed86855d37a5c3dec73dbb3079f8c795\": container with ID starting with ea88e490cd9bdf056778cbda1d520395ed86855d37a5c3dec73dbb3079f8c795 not found: ID does not exist" Feb 18 09:20:46 crc kubenswrapper[4556]: I0218 09:20:46.370780 4556 scope.go:117] "RemoveContainer" containerID="a903312a63e39580644536734e7abd02a75c7922bd9cd0e7e5635d7c7d0e5fca" Feb 18 09:20:46 crc kubenswrapper[4556]: E0218 09:20:46.371243 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a903312a63e39580644536734e7abd02a75c7922bd9cd0e7e5635d7c7d0e5fca\": container with ID starting with a903312a63e39580644536734e7abd02a75c7922bd9cd0e7e5635d7c7d0e5fca not found: ID does not exist" containerID="a903312a63e39580644536734e7abd02a75c7922bd9cd0e7e5635d7c7d0e5fca" Feb 18 09:20:46 crc kubenswrapper[4556]: I0218 09:20:46.371338 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a903312a63e39580644536734e7abd02a75c7922bd9cd0e7e5635d7c7d0e5fca"} err="failed to get container status \"a903312a63e39580644536734e7abd02a75c7922bd9cd0e7e5635d7c7d0e5fca\": rpc error: code = NotFound desc = could not find container \"a903312a63e39580644536734e7abd02a75c7922bd9cd0e7e5635d7c7d0e5fca\": container with ID starting with a903312a63e39580644536734e7abd02a75c7922bd9cd0e7e5635d7c7d0e5fca not found: ID does not exist" Feb 18 09:20:46 crc kubenswrapper[4556]: I0218 09:20:46.426739 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cs68h\" (UniqueName: \"kubernetes.io/projected/4424b924-1dd6-4323-863d-53bf1a0f5fc9-kube-api-access-cs68h\") pod \"nova-metadata-0\" (UID: \"4424b924-1dd6-4323-863d-53bf1a0f5fc9\") " pod="openstack/nova-metadata-0" Feb 18 09:20:46 crc kubenswrapper[4556]: I0218 09:20:46.426927 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4424b924-1dd6-4323-863d-53bf1a0f5fc9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4424b924-1dd6-4323-863d-53bf1a0f5fc9\") " pod="openstack/nova-metadata-0" Feb 18 09:20:46 crc kubenswrapper[4556]: I0218 09:20:46.427034 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4424b924-1dd6-4323-863d-53bf1a0f5fc9-logs\") pod \"nova-metadata-0\" (UID: \"4424b924-1dd6-4323-863d-53bf1a0f5fc9\") " pod="openstack/nova-metadata-0" Feb 18 09:20:46 crc kubenswrapper[4556]: I0218 09:20:46.427288 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4424b924-1dd6-4323-863d-53bf1a0f5fc9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"4424b924-1dd6-4323-863d-53bf1a0f5fc9\") " pod="openstack/nova-metadata-0" Feb 18 09:20:46 crc kubenswrapper[4556]: I0218 09:20:46.427522 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4424b924-1dd6-4323-863d-53bf1a0f5fc9-config-data\") pod \"nova-metadata-0\" (UID: \"4424b924-1dd6-4323-863d-53bf1a0f5fc9\") " pod="openstack/nova-metadata-0" Feb 18 09:20:46 crc kubenswrapper[4556]: I0218 09:20:46.529889 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4424b924-1dd6-4323-863d-53bf1a0f5fc9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"4424b924-1dd6-4323-863d-53bf1a0f5fc9\") " pod="openstack/nova-metadata-0" Feb 18 09:20:46 crc kubenswrapper[4556]: I0218 09:20:46.530139 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4424b924-1dd6-4323-863d-53bf1a0f5fc9-config-data\") pod \"nova-metadata-0\" (UID: \"4424b924-1dd6-4323-863d-53bf1a0f5fc9\") " pod="openstack/nova-metadata-0" Feb 18 09:20:46 crc kubenswrapper[4556]: I0218 09:20:46.530307 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cs68h\" (UniqueName: \"kubernetes.io/projected/4424b924-1dd6-4323-863d-53bf1a0f5fc9-kube-api-access-cs68h\") pod \"nova-metadata-0\" (UID: \"4424b924-1dd6-4323-863d-53bf1a0f5fc9\") " pod="openstack/nova-metadata-0" Feb 18 09:20:46 crc kubenswrapper[4556]: I0218 09:20:46.530343 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4424b924-1dd6-4323-863d-53bf1a0f5fc9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4424b924-1dd6-4323-863d-53bf1a0f5fc9\") " pod="openstack/nova-metadata-0" Feb 18 09:20:46 crc kubenswrapper[4556]: I0218 09:20:46.530389 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4424b924-1dd6-4323-863d-53bf1a0f5fc9-logs\") pod \"nova-metadata-0\" (UID: \"4424b924-1dd6-4323-863d-53bf1a0f5fc9\") " pod="openstack/nova-metadata-0" Feb 18 09:20:46 crc kubenswrapper[4556]: I0218 09:20:46.530944 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4424b924-1dd6-4323-863d-53bf1a0f5fc9-logs\") pod \"nova-metadata-0\" (UID: \"4424b924-1dd6-4323-863d-53bf1a0f5fc9\") " pod="openstack/nova-metadata-0" Feb 18 09:20:46 crc kubenswrapper[4556]: I0218 09:20:46.533713 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4424b924-1dd6-4323-863d-53bf1a0f5fc9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"4424b924-1dd6-4323-863d-53bf1a0f5fc9\") " pod="openstack/nova-metadata-0" Feb 18 09:20:46 crc kubenswrapper[4556]: I0218 09:20:46.534226 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4424b924-1dd6-4323-863d-53bf1a0f5fc9-config-data\") pod \"nova-metadata-0\" (UID: \"4424b924-1dd6-4323-863d-53bf1a0f5fc9\") " pod="openstack/nova-metadata-0" Feb 18 09:20:46 crc kubenswrapper[4556]: I0218 09:20:46.534769 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4424b924-1dd6-4323-863d-53bf1a0f5fc9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4424b924-1dd6-4323-863d-53bf1a0f5fc9\") " pod="openstack/nova-metadata-0" Feb 18 09:20:46 crc kubenswrapper[4556]: I0218 09:20:46.545388 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cs68h\" (UniqueName: \"kubernetes.io/projected/4424b924-1dd6-4323-863d-53bf1a0f5fc9-kube-api-access-cs68h\") pod \"nova-metadata-0\" (UID: \"4424b924-1dd6-4323-863d-53bf1a0f5fc9\") " pod="openstack/nova-metadata-0" Feb 18 09:20:46 crc kubenswrapper[4556]: I0218 09:20:46.656273 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 18 09:20:47 crc kubenswrapper[4556]: W0218 09:20:47.074428 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4424b924_1dd6_4323_863d_53bf1a0f5fc9.slice/crio-90e2a1a3388b20c395073f944c2cf6b4ed8dd101b79d57676ba5dc0296b4d544 WatchSource:0}: Error finding container 90e2a1a3388b20c395073f944c2cf6b4ed8dd101b79d57676ba5dc0296b4d544: Status 404 returned error can't find the container with id 90e2a1a3388b20c395073f944c2cf6b4ed8dd101b79d57676ba5dc0296b4d544 Feb 18 09:20:47 crc kubenswrapper[4556]: I0218 09:20:47.079472 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 18 09:20:47 crc kubenswrapper[4556]: I0218 09:20:47.303552 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f3e12da-26fe-4d2a-ae62-9a3e90c5b752" path="/var/lib/kubelet/pods/1f3e12da-26fe-4d2a-ae62-9a3e90c5b752/volumes" Feb 18 09:20:47 crc kubenswrapper[4556]: I0218 09:20:47.304896 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4424b924-1dd6-4323-863d-53bf1a0f5fc9","Type":"ContainerStarted","Data":"ad18e02698cf48dc67c6ac6a82b5f82f004986b5e536a3ec7d2f886bb0ad6fbc"} Feb 18 09:20:47 crc kubenswrapper[4556]: I0218 09:20:47.304938 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4424b924-1dd6-4323-863d-53bf1a0f5fc9","Type":"ContainerStarted","Data":"90e2a1a3388b20c395073f944c2cf6b4ed8dd101b79d57676ba5dc0296b4d544"} Feb 18 09:20:48 crc kubenswrapper[4556]: I0218 09:20:48.300464 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4424b924-1dd6-4323-863d-53bf1a0f5fc9","Type":"ContainerStarted","Data":"ee514c14b63db5ddac93f18f6107f7f43a68f05bf54f1234c43bf0504e4d124c"} Feb 18 09:20:48 crc kubenswrapper[4556]: I0218 09:20:48.329200 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.329170747 podStartE2EDuration="2.329170747s" podCreationTimestamp="2026-02-18 09:20:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:20:48.316968484 +0000 UTC m=+1005.333929464" watchObservedRunningTime="2026-02-18 09:20:48.329170747 +0000 UTC m=+1005.346131727" Feb 18 09:20:49 crc kubenswrapper[4556]: I0218 09:20:49.684685 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Feb 18 09:20:51 crc kubenswrapper[4556]: I0218 09:20:51.657027 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 18 09:20:51 crc kubenswrapper[4556]: I0218 09:20:51.657099 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 18 09:20:54 crc kubenswrapper[4556]: I0218 09:20:54.211460 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 18 09:20:54 crc kubenswrapper[4556]: I0218 09:20:54.212305 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 18 09:20:54 crc kubenswrapper[4556]: I0218 09:20:54.684806 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Feb 18 09:20:54 crc kubenswrapper[4556]: I0218 09:20:54.708650 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Feb 18 09:20:55 crc kubenswrapper[4556]: I0218 09:20:55.228379 4556 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="4e997362-15c5-4052-bf0b-8491c057f317" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.205:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 18 09:20:55 crc kubenswrapper[4556]: I0218 09:20:55.228391 4556 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="4e997362-15c5-4052-bf0b-8491c057f317" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.205:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 18 09:20:55 crc kubenswrapper[4556]: I0218 09:20:55.381203 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Feb 18 09:20:56 crc kubenswrapper[4556]: I0218 09:20:56.656996 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 18 09:20:56 crc kubenswrapper[4556]: I0218 09:20:56.657271 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 18 09:20:57 crc kubenswrapper[4556]: I0218 09:20:57.671274 4556 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="4424b924-1dd6-4323-863d-53bf1a0f5fc9" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.207:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 18 09:20:57 crc kubenswrapper[4556]: I0218 09:20:57.671303 4556 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="4424b924-1dd6-4323-863d-53bf1a0f5fc9" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.207:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 18 09:21:00 crc kubenswrapper[4556]: I0218 09:21:00.493420 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Feb 18 09:21:04 crc kubenswrapper[4556]: I0218 09:21:04.215856 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 18 09:21:04 crc kubenswrapper[4556]: I0218 09:21:04.216625 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 18 09:21:04 crc kubenswrapper[4556]: I0218 09:21:04.216862 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 18 09:21:04 crc kubenswrapper[4556]: I0218 09:21:04.216904 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 18 09:21:04 crc kubenswrapper[4556]: I0218 09:21:04.221597 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 18 09:21:04 crc kubenswrapper[4556]: I0218 09:21:04.221983 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 18 09:21:06 crc kubenswrapper[4556]: I0218 09:21:06.662068 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 18 09:21:06 crc kubenswrapper[4556]: I0218 09:21:06.662847 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 18 09:21:06 crc kubenswrapper[4556]: I0218 09:21:06.669692 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 18 09:21:07 crc kubenswrapper[4556]: I0218 09:21:07.451262 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 18 09:21:13 crc kubenswrapper[4556]: I0218 09:21:13.792567 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 18 09:21:14 crc kubenswrapper[4556]: I0218 09:21:14.927004 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 18 09:21:17 crc kubenswrapper[4556]: I0218 09:21:17.110788 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="1d7961e1-6f96-4707-907d-1fa0c30641e2" containerName="rabbitmq" containerID="cri-o://845e8aa21dc1c9ff225cb48767ac3a68bdf567f6d2263ac90aef2e165de13da8" gracePeriod=604797 Feb 18 09:21:18 crc kubenswrapper[4556]: I0218 09:21:18.127194 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="03023501-772f-4798-a953-f56dee586cb1" containerName="rabbitmq" containerID="cri-o://b6d70d7c016d0deed03db8dc523b6b6d3d816ebdd2eb672bcd3ebfca6f608a64" gracePeriod=604797 Feb 18 09:21:22 crc kubenswrapper[4556]: I0218 09:21:22.495368 4556 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="1d7961e1-6f96-4707-907d-1fa0c30641e2" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.95:5671: connect: connection refused" Feb 18 09:21:22 crc kubenswrapper[4556]: I0218 09:21:22.533201 4556 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="03023501-772f-4798-a953-f56dee586cb1" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.96:5671: connect: connection refused" Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.501978 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.561747 4556 generic.go:334] "Generic (PLEG): container finished" podID="1d7961e1-6f96-4707-907d-1fa0c30641e2" containerID="845e8aa21dc1c9ff225cb48767ac3a68bdf567f6d2263ac90aef2e165de13da8" exitCode=0 Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.561786 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1d7961e1-6f96-4707-907d-1fa0c30641e2","Type":"ContainerDied","Data":"845e8aa21dc1c9ff225cb48767ac3a68bdf567f6d2263ac90aef2e165de13da8"} Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.561812 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1d7961e1-6f96-4707-907d-1fa0c30641e2","Type":"ContainerDied","Data":"25cc6087f58009fb241691501ea8c5e422f1f3c289c1bb7f9ea20ab54db055a6"} Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.561837 4556 scope.go:117] "RemoveContainer" containerID="845e8aa21dc1c9ff225cb48767ac3a68bdf567f6d2263ac90aef2e165de13da8" Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.561926 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.588268 4556 scope.go:117] "RemoveContainer" containerID="98251fd58b2a29ac981c57fb4437b319daf966b2b07731b525037e5dccf051cf" Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.607837 4556 scope.go:117] "RemoveContainer" containerID="845e8aa21dc1c9ff225cb48767ac3a68bdf567f6d2263ac90aef2e165de13da8" Feb 18 09:21:23 crc kubenswrapper[4556]: E0218 09:21:23.608342 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"845e8aa21dc1c9ff225cb48767ac3a68bdf567f6d2263ac90aef2e165de13da8\": container with ID starting with 845e8aa21dc1c9ff225cb48767ac3a68bdf567f6d2263ac90aef2e165de13da8 not found: ID does not exist" containerID="845e8aa21dc1c9ff225cb48767ac3a68bdf567f6d2263ac90aef2e165de13da8" Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.608381 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"845e8aa21dc1c9ff225cb48767ac3a68bdf567f6d2263ac90aef2e165de13da8"} err="failed to get container status \"845e8aa21dc1c9ff225cb48767ac3a68bdf567f6d2263ac90aef2e165de13da8\": rpc error: code = NotFound desc = could not find container \"845e8aa21dc1c9ff225cb48767ac3a68bdf567f6d2263ac90aef2e165de13da8\": container with ID starting with 845e8aa21dc1c9ff225cb48767ac3a68bdf567f6d2263ac90aef2e165de13da8 not found: ID does not exist" Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.608407 4556 scope.go:117] "RemoveContainer" containerID="98251fd58b2a29ac981c57fb4437b319daf966b2b07731b525037e5dccf051cf" Feb 18 09:21:23 crc kubenswrapper[4556]: E0218 09:21:23.608800 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98251fd58b2a29ac981c57fb4437b319daf966b2b07731b525037e5dccf051cf\": container with ID starting with 98251fd58b2a29ac981c57fb4437b319daf966b2b07731b525037e5dccf051cf not found: ID does not exist" containerID="98251fd58b2a29ac981c57fb4437b319daf966b2b07731b525037e5dccf051cf" Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.608822 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98251fd58b2a29ac981c57fb4437b319daf966b2b07731b525037e5dccf051cf"} err="failed to get container status \"98251fd58b2a29ac981c57fb4437b319daf966b2b07731b525037e5dccf051cf\": rpc error: code = NotFound desc = could not find container \"98251fd58b2a29ac981c57fb4437b319daf966b2b07731b525037e5dccf051cf\": container with ID starting with 98251fd58b2a29ac981c57fb4437b319daf966b2b07731b525037e5dccf051cf not found: ID does not exist" Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.644631 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1d7961e1-6f96-4707-907d-1fa0c30641e2-pod-info\") pod \"1d7961e1-6f96-4707-907d-1fa0c30641e2\" (UID: \"1d7961e1-6f96-4707-907d-1fa0c30641e2\") " Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.644768 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6n8rf\" (UniqueName: \"kubernetes.io/projected/1d7961e1-6f96-4707-907d-1fa0c30641e2-kube-api-access-6n8rf\") pod \"1d7961e1-6f96-4707-907d-1fa0c30641e2\" (UID: \"1d7961e1-6f96-4707-907d-1fa0c30641e2\") " Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.644860 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1d7961e1-6f96-4707-907d-1fa0c30641e2-server-conf\") pod \"1d7961e1-6f96-4707-907d-1fa0c30641e2\" (UID: \"1d7961e1-6f96-4707-907d-1fa0c30641e2\") " Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.644932 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1d7961e1-6f96-4707-907d-1fa0c30641e2-plugins-conf\") pod \"1d7961e1-6f96-4707-907d-1fa0c30641e2\" (UID: \"1d7961e1-6f96-4707-907d-1fa0c30641e2\") " Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.645008 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1d7961e1-6f96-4707-907d-1fa0c30641e2-erlang-cookie-secret\") pod \"1d7961e1-6f96-4707-907d-1fa0c30641e2\" (UID: \"1d7961e1-6f96-4707-907d-1fa0c30641e2\") " Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.645115 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"1d7961e1-6f96-4707-907d-1fa0c30641e2\" (UID: \"1d7961e1-6f96-4707-907d-1fa0c30641e2\") " Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.645506 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1d7961e1-6f96-4707-907d-1fa0c30641e2-rabbitmq-tls\") pod \"1d7961e1-6f96-4707-907d-1fa0c30641e2\" (UID: \"1d7961e1-6f96-4707-907d-1fa0c30641e2\") " Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.645552 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d7961e1-6f96-4707-907d-1fa0c30641e2-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "1d7961e1-6f96-4707-907d-1fa0c30641e2" (UID: "1d7961e1-6f96-4707-907d-1fa0c30641e2"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.645651 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1d7961e1-6f96-4707-907d-1fa0c30641e2-rabbitmq-erlang-cookie\") pod \"1d7961e1-6f96-4707-907d-1fa0c30641e2\" (UID: \"1d7961e1-6f96-4707-907d-1fa0c30641e2\") " Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.645749 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1d7961e1-6f96-4707-907d-1fa0c30641e2-rabbitmq-confd\") pod \"1d7961e1-6f96-4707-907d-1fa0c30641e2\" (UID: \"1d7961e1-6f96-4707-907d-1fa0c30641e2\") " Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.645863 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1d7961e1-6f96-4707-907d-1fa0c30641e2-config-data\") pod \"1d7961e1-6f96-4707-907d-1fa0c30641e2\" (UID: \"1d7961e1-6f96-4707-907d-1fa0c30641e2\") " Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.645996 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1d7961e1-6f96-4707-907d-1fa0c30641e2-rabbitmq-plugins\") pod \"1d7961e1-6f96-4707-907d-1fa0c30641e2\" (UID: \"1d7961e1-6f96-4707-907d-1fa0c30641e2\") " Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.646052 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d7961e1-6f96-4707-907d-1fa0c30641e2-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "1d7961e1-6f96-4707-907d-1fa0c30641e2" (UID: "1d7961e1-6f96-4707-907d-1fa0c30641e2"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.646772 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d7961e1-6f96-4707-907d-1fa0c30641e2-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "1d7961e1-6f96-4707-907d-1fa0c30641e2" (UID: "1d7961e1-6f96-4707-907d-1fa0c30641e2"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.646883 4556 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1d7961e1-6f96-4707-907d-1fa0c30641e2-plugins-conf\") on node \"crc\" DevicePath \"\"" Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.646945 4556 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1d7961e1-6f96-4707-907d-1fa0c30641e2-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.650575 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d7961e1-6f96-4707-907d-1fa0c30641e2-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "1d7961e1-6f96-4707-907d-1fa0c30641e2" (UID: "1d7961e1-6f96-4707-907d-1fa0c30641e2"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.650651 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d7961e1-6f96-4707-907d-1fa0c30641e2-kube-api-access-6n8rf" (OuterVolumeSpecName: "kube-api-access-6n8rf") pod "1d7961e1-6f96-4707-907d-1fa0c30641e2" (UID: "1d7961e1-6f96-4707-907d-1fa0c30641e2"). InnerVolumeSpecName "kube-api-access-6n8rf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.651013 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d7961e1-6f96-4707-907d-1fa0c30641e2-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "1d7961e1-6f96-4707-907d-1fa0c30641e2" (UID: "1d7961e1-6f96-4707-907d-1fa0c30641e2"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.651281 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "persistence") pod "1d7961e1-6f96-4707-907d-1fa0c30641e2" (UID: "1d7961e1-6f96-4707-907d-1fa0c30641e2"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.651469 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/1d7961e1-6f96-4707-907d-1fa0c30641e2-pod-info" (OuterVolumeSpecName: "pod-info") pod "1d7961e1-6f96-4707-907d-1fa0c30641e2" (UID: "1d7961e1-6f96-4707-907d-1fa0c30641e2"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.668585 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d7961e1-6f96-4707-907d-1fa0c30641e2-config-data" (OuterVolumeSpecName: "config-data") pod "1d7961e1-6f96-4707-907d-1fa0c30641e2" (UID: "1d7961e1-6f96-4707-907d-1fa0c30641e2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.690006 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d7961e1-6f96-4707-907d-1fa0c30641e2-server-conf" (OuterVolumeSpecName: "server-conf") pod "1d7961e1-6f96-4707-907d-1fa0c30641e2" (UID: "1d7961e1-6f96-4707-907d-1fa0c30641e2"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.751615 4556 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1d7961e1-6f96-4707-907d-1fa0c30641e2-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.751724 4556 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1d7961e1-6f96-4707-907d-1fa0c30641e2-config-data\") on node \"crc\" DevicePath \"\"" Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.751831 4556 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1d7961e1-6f96-4707-907d-1fa0c30641e2-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.751841 4556 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1d7961e1-6f96-4707-907d-1fa0c30641e2-pod-info\") on node \"crc\" DevicePath \"\"" Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.751900 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6n8rf\" (UniqueName: \"kubernetes.io/projected/1d7961e1-6f96-4707-907d-1fa0c30641e2-kube-api-access-6n8rf\") on node \"crc\" DevicePath \"\"" Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.751928 4556 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1d7961e1-6f96-4707-907d-1fa0c30641e2-server-conf\") on node \"crc\" DevicePath \"\"" Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.751937 4556 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1d7961e1-6f96-4707-907d-1fa0c30641e2-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.752028 4556 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.754624 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d7961e1-6f96-4707-907d-1fa0c30641e2-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "1d7961e1-6f96-4707-907d-1fa0c30641e2" (UID: "1d7961e1-6f96-4707-907d-1fa0c30641e2"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.768869 4556 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.853891 4556 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1d7961e1-6f96-4707-907d-1fa0c30641e2-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.853915 4556 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.886204 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.893798 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.903686 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Feb 18 09:21:23 crc kubenswrapper[4556]: E0218 09:21:23.903971 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d7961e1-6f96-4707-907d-1fa0c30641e2" containerName="setup-container" Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.903987 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d7961e1-6f96-4707-907d-1fa0c30641e2" containerName="setup-container" Feb 18 09:21:23 crc kubenswrapper[4556]: E0218 09:21:23.904004 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d7961e1-6f96-4707-907d-1fa0c30641e2" containerName="rabbitmq" Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.904009 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d7961e1-6f96-4707-907d-1fa0c30641e2" containerName="rabbitmq" Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.904203 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d7961e1-6f96-4707-907d-1fa0c30641e2" containerName="rabbitmq" Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.904957 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.907346 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.907377 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.907460 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.907714 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-rmfvs" Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.907823 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.907946 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.908228 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.919059 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.959996 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0f0ef5d2-a692-435f-a79d-a1af3a294e73-pod-info\") pod \"rabbitmq-server-0\" (UID: \"0f0ef5d2-a692-435f-a79d-a1af3a294e73\") " pod="openstack/rabbitmq-server-0" Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.960046 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0f0ef5d2-a692-435f-a79d-a1af3a294e73-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"0f0ef5d2-a692-435f-a79d-a1af3a294e73\") " pod="openstack/rabbitmq-server-0" Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.960125 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0f0ef5d2-a692-435f-a79d-a1af3a294e73-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"0f0ef5d2-a692-435f-a79d-a1af3a294e73\") " pod="openstack/rabbitmq-server-0" Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.961088 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxf9f\" (UniqueName: \"kubernetes.io/projected/0f0ef5d2-a692-435f-a79d-a1af3a294e73-kube-api-access-fxf9f\") pod \"rabbitmq-server-0\" (UID: \"0f0ef5d2-a692-435f-a79d-a1af3a294e73\") " pod="openstack/rabbitmq-server-0" Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.961230 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0f0ef5d2-a692-435f-a79d-a1af3a294e73-config-data\") pod \"rabbitmq-server-0\" (UID: \"0f0ef5d2-a692-435f-a79d-a1af3a294e73\") " pod="openstack/rabbitmq-server-0" Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.961262 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0f0ef5d2-a692-435f-a79d-a1af3a294e73-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"0f0ef5d2-a692-435f-a79d-a1af3a294e73\") " pod="openstack/rabbitmq-server-0" Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.961296 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0f0ef5d2-a692-435f-a79d-a1af3a294e73-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"0f0ef5d2-a692-435f-a79d-a1af3a294e73\") " pod="openstack/rabbitmq-server-0" Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.961408 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"0f0ef5d2-a692-435f-a79d-a1af3a294e73\") " pod="openstack/rabbitmq-server-0" Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.961448 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0f0ef5d2-a692-435f-a79d-a1af3a294e73-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"0f0ef5d2-a692-435f-a79d-a1af3a294e73\") " pod="openstack/rabbitmq-server-0" Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.961561 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0f0ef5d2-a692-435f-a79d-a1af3a294e73-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"0f0ef5d2-a692-435f-a79d-a1af3a294e73\") " pod="openstack/rabbitmq-server-0" Feb 18 09:21:23 crc kubenswrapper[4556]: I0218 09:21:23.961637 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0f0ef5d2-a692-435f-a79d-a1af3a294e73-server-conf\") pod \"rabbitmq-server-0\" (UID: \"0f0ef5d2-a692-435f-a79d-a1af3a294e73\") " pod="openstack/rabbitmq-server-0" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.062960 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0f0ef5d2-a692-435f-a79d-a1af3a294e73-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"0f0ef5d2-a692-435f-a79d-a1af3a294e73\") " pod="openstack/rabbitmq-server-0" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.063033 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0f0ef5d2-a692-435f-a79d-a1af3a294e73-server-conf\") pod \"rabbitmq-server-0\" (UID: \"0f0ef5d2-a692-435f-a79d-a1af3a294e73\") " pod="openstack/rabbitmq-server-0" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.063095 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0f0ef5d2-a692-435f-a79d-a1af3a294e73-pod-info\") pod \"rabbitmq-server-0\" (UID: \"0f0ef5d2-a692-435f-a79d-a1af3a294e73\") " pod="openstack/rabbitmq-server-0" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.063127 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0f0ef5d2-a692-435f-a79d-a1af3a294e73-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"0f0ef5d2-a692-435f-a79d-a1af3a294e73\") " pod="openstack/rabbitmq-server-0" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.063193 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0f0ef5d2-a692-435f-a79d-a1af3a294e73-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"0f0ef5d2-a692-435f-a79d-a1af3a294e73\") " pod="openstack/rabbitmq-server-0" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.063226 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxf9f\" (UniqueName: \"kubernetes.io/projected/0f0ef5d2-a692-435f-a79d-a1af3a294e73-kube-api-access-fxf9f\") pod \"rabbitmq-server-0\" (UID: \"0f0ef5d2-a692-435f-a79d-a1af3a294e73\") " pod="openstack/rabbitmq-server-0" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.063269 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0f0ef5d2-a692-435f-a79d-a1af3a294e73-config-data\") pod \"rabbitmq-server-0\" (UID: \"0f0ef5d2-a692-435f-a79d-a1af3a294e73\") " pod="openstack/rabbitmq-server-0" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.063283 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0f0ef5d2-a692-435f-a79d-a1af3a294e73-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"0f0ef5d2-a692-435f-a79d-a1af3a294e73\") " pod="openstack/rabbitmq-server-0" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.063303 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0f0ef5d2-a692-435f-a79d-a1af3a294e73-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"0f0ef5d2-a692-435f-a79d-a1af3a294e73\") " pod="openstack/rabbitmq-server-0" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.063349 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"0f0ef5d2-a692-435f-a79d-a1af3a294e73\") " pod="openstack/rabbitmq-server-0" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.063369 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0f0ef5d2-a692-435f-a79d-a1af3a294e73-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"0f0ef5d2-a692-435f-a79d-a1af3a294e73\") " pod="openstack/rabbitmq-server-0" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.063833 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0f0ef5d2-a692-435f-a79d-a1af3a294e73-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"0f0ef5d2-a692-435f-a79d-a1af3a294e73\") " pod="openstack/rabbitmq-server-0" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.064071 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0f0ef5d2-a692-435f-a79d-a1af3a294e73-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"0f0ef5d2-a692-435f-a79d-a1af3a294e73\") " pod="openstack/rabbitmq-server-0" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.065118 4556 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"0f0ef5d2-a692-435f-a79d-a1af3a294e73\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-server-0" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.065530 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0f0ef5d2-a692-435f-a79d-a1af3a294e73-config-data\") pod \"rabbitmq-server-0\" (UID: \"0f0ef5d2-a692-435f-a79d-a1af3a294e73\") " pod="openstack/rabbitmq-server-0" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.065861 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0f0ef5d2-a692-435f-a79d-a1af3a294e73-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"0f0ef5d2-a692-435f-a79d-a1af3a294e73\") " pod="openstack/rabbitmq-server-0" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.065921 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0f0ef5d2-a692-435f-a79d-a1af3a294e73-server-conf\") pod \"rabbitmq-server-0\" (UID: \"0f0ef5d2-a692-435f-a79d-a1af3a294e73\") " pod="openstack/rabbitmq-server-0" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.067438 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0f0ef5d2-a692-435f-a79d-a1af3a294e73-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"0f0ef5d2-a692-435f-a79d-a1af3a294e73\") " pod="openstack/rabbitmq-server-0" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.067936 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0f0ef5d2-a692-435f-a79d-a1af3a294e73-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"0f0ef5d2-a692-435f-a79d-a1af3a294e73\") " pod="openstack/rabbitmq-server-0" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.068595 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0f0ef5d2-a692-435f-a79d-a1af3a294e73-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"0f0ef5d2-a692-435f-a79d-a1af3a294e73\") " pod="openstack/rabbitmq-server-0" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.069131 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0f0ef5d2-a692-435f-a79d-a1af3a294e73-pod-info\") pod \"rabbitmq-server-0\" (UID: \"0f0ef5d2-a692-435f-a79d-a1af3a294e73\") " pod="openstack/rabbitmq-server-0" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.078467 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxf9f\" (UniqueName: \"kubernetes.io/projected/0f0ef5d2-a692-435f-a79d-a1af3a294e73-kube-api-access-fxf9f\") pod \"rabbitmq-server-0\" (UID: \"0f0ef5d2-a692-435f-a79d-a1af3a294e73\") " pod="openstack/rabbitmq-server-0" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.096093 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"0f0ef5d2-a692-435f-a79d-a1af3a294e73\") " pod="openstack/rabbitmq-server-0" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.227548 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.502233 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6c564ccb4c-s54zm"] Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.504167 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c564ccb4c-s54zm" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.506767 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.542242 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6c564ccb4c-s54zm"] Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.579056 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d21c134-833c-4f8d-b2d9-a3c25b2ca597-config\") pod \"dnsmasq-dns-6c564ccb4c-s54zm\" (UID: \"5d21c134-833c-4f8d-b2d9-a3c25b2ca597\") " pod="openstack/dnsmasq-dns-6c564ccb4c-s54zm" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.579145 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5d21c134-833c-4f8d-b2d9-a3c25b2ca597-dns-swift-storage-0\") pod \"dnsmasq-dns-6c564ccb4c-s54zm\" (UID: \"5d21c134-833c-4f8d-b2d9-a3c25b2ca597\") " pod="openstack/dnsmasq-dns-6c564ccb4c-s54zm" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.579231 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5d21c134-833c-4f8d-b2d9-a3c25b2ca597-dns-svc\") pod \"dnsmasq-dns-6c564ccb4c-s54zm\" (UID: \"5d21c134-833c-4f8d-b2d9-a3c25b2ca597\") " pod="openstack/dnsmasq-dns-6c564ccb4c-s54zm" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.579492 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/5d21c134-833c-4f8d-b2d9-a3c25b2ca597-openstack-edpm-ipam\") pod \"dnsmasq-dns-6c564ccb4c-s54zm\" (UID: \"5d21c134-833c-4f8d-b2d9-a3c25b2ca597\") " pod="openstack/dnsmasq-dns-6c564ccb4c-s54zm" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.579561 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5d21c134-833c-4f8d-b2d9-a3c25b2ca597-ovsdbserver-sb\") pod \"dnsmasq-dns-6c564ccb4c-s54zm\" (UID: \"5d21c134-833c-4f8d-b2d9-a3c25b2ca597\") " pod="openstack/dnsmasq-dns-6c564ccb4c-s54zm" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.579641 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5d21c134-833c-4f8d-b2d9-a3c25b2ca597-ovsdbserver-nb\") pod \"dnsmasq-dns-6c564ccb4c-s54zm\" (UID: \"5d21c134-833c-4f8d-b2d9-a3c25b2ca597\") " pod="openstack/dnsmasq-dns-6c564ccb4c-s54zm" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.579722 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5cpfl\" (UniqueName: \"kubernetes.io/projected/5d21c134-833c-4f8d-b2d9-a3c25b2ca597-kube-api-access-5cpfl\") pod \"dnsmasq-dns-6c564ccb4c-s54zm\" (UID: \"5d21c134-833c-4f8d-b2d9-a3c25b2ca597\") " pod="openstack/dnsmasq-dns-6c564ccb4c-s54zm" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.583334 4556 generic.go:334] "Generic (PLEG): container finished" podID="03023501-772f-4798-a953-f56dee586cb1" containerID="b6d70d7c016d0deed03db8dc523b6b6d3d816ebdd2eb672bcd3ebfca6f608a64" exitCode=0 Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.583378 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"03023501-772f-4798-a953-f56dee586cb1","Type":"ContainerDied","Data":"b6d70d7c016d0deed03db8dc523b6b6d3d816ebdd2eb672bcd3ebfca6f608a64"} Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.628314 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c564ccb4c-s54zm"] Feb 18 09:21:24 crc kubenswrapper[4556]: E0218 09:21:24.629223 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc dns-swift-storage-0 kube-api-access-5cpfl openstack-edpm-ipam ovsdbserver-nb ovsdbserver-sb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-6c564ccb4c-s54zm" podUID="5d21c134-833c-4f8d-b2d9-a3c25b2ca597" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.663947 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-56455489f5-9pk6d"] Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.665615 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56455489f5-9pk6d" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.666083 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.681528 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56455489f5-9pk6d"] Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.682681 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d21c134-833c-4f8d-b2d9-a3c25b2ca597-config\") pod \"dnsmasq-dns-6c564ccb4c-s54zm\" (UID: \"5d21c134-833c-4f8d-b2d9-a3c25b2ca597\") " pod="openstack/dnsmasq-dns-6c564ccb4c-s54zm" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.682750 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5d21c134-833c-4f8d-b2d9-a3c25b2ca597-dns-swift-storage-0\") pod \"dnsmasq-dns-6c564ccb4c-s54zm\" (UID: \"5d21c134-833c-4f8d-b2d9-a3c25b2ca597\") " pod="openstack/dnsmasq-dns-6c564ccb4c-s54zm" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.682783 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5d21c134-833c-4f8d-b2d9-a3c25b2ca597-dns-svc\") pod \"dnsmasq-dns-6c564ccb4c-s54zm\" (UID: \"5d21c134-833c-4f8d-b2d9-a3c25b2ca597\") " pod="openstack/dnsmasq-dns-6c564ccb4c-s54zm" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.682904 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/5d21c134-833c-4f8d-b2d9-a3c25b2ca597-openstack-edpm-ipam\") pod \"dnsmasq-dns-6c564ccb4c-s54zm\" (UID: \"5d21c134-833c-4f8d-b2d9-a3c25b2ca597\") " pod="openstack/dnsmasq-dns-6c564ccb4c-s54zm" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.682949 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5d21c134-833c-4f8d-b2d9-a3c25b2ca597-ovsdbserver-sb\") pod \"dnsmasq-dns-6c564ccb4c-s54zm\" (UID: \"5d21c134-833c-4f8d-b2d9-a3c25b2ca597\") " pod="openstack/dnsmasq-dns-6c564ccb4c-s54zm" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.682965 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5d21c134-833c-4f8d-b2d9-a3c25b2ca597-ovsdbserver-nb\") pod \"dnsmasq-dns-6c564ccb4c-s54zm\" (UID: \"5d21c134-833c-4f8d-b2d9-a3c25b2ca597\") " pod="openstack/dnsmasq-dns-6c564ccb4c-s54zm" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.683027 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5cpfl\" (UniqueName: \"kubernetes.io/projected/5d21c134-833c-4f8d-b2d9-a3c25b2ca597-kube-api-access-5cpfl\") pod \"dnsmasq-dns-6c564ccb4c-s54zm\" (UID: \"5d21c134-833c-4f8d-b2d9-a3c25b2ca597\") " pod="openstack/dnsmasq-dns-6c564ccb4c-s54zm" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.683767 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d21c134-833c-4f8d-b2d9-a3c25b2ca597-config\") pod \"dnsmasq-dns-6c564ccb4c-s54zm\" (UID: \"5d21c134-833c-4f8d-b2d9-a3c25b2ca597\") " pod="openstack/dnsmasq-dns-6c564ccb4c-s54zm" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.683780 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5d21c134-833c-4f8d-b2d9-a3c25b2ca597-dns-svc\") pod \"dnsmasq-dns-6c564ccb4c-s54zm\" (UID: \"5d21c134-833c-4f8d-b2d9-a3c25b2ca597\") " pod="openstack/dnsmasq-dns-6c564ccb4c-s54zm" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.683986 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5d21c134-833c-4f8d-b2d9-a3c25b2ca597-ovsdbserver-sb\") pod \"dnsmasq-dns-6c564ccb4c-s54zm\" (UID: \"5d21c134-833c-4f8d-b2d9-a3c25b2ca597\") " pod="openstack/dnsmasq-dns-6c564ccb4c-s54zm" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.685969 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5d21c134-833c-4f8d-b2d9-a3c25b2ca597-dns-swift-storage-0\") pod \"dnsmasq-dns-6c564ccb4c-s54zm\" (UID: \"5d21c134-833c-4f8d-b2d9-a3c25b2ca597\") " pod="openstack/dnsmasq-dns-6c564ccb4c-s54zm" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.686776 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5d21c134-833c-4f8d-b2d9-a3c25b2ca597-ovsdbserver-nb\") pod \"dnsmasq-dns-6c564ccb4c-s54zm\" (UID: \"5d21c134-833c-4f8d-b2d9-a3c25b2ca597\") " pod="openstack/dnsmasq-dns-6c564ccb4c-s54zm" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.687378 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/5d21c134-833c-4f8d-b2d9-a3c25b2ca597-openstack-edpm-ipam\") pod \"dnsmasq-dns-6c564ccb4c-s54zm\" (UID: \"5d21c134-833c-4f8d-b2d9-a3c25b2ca597\") " pod="openstack/dnsmasq-dns-6c564ccb4c-s54zm" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.721986 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5cpfl\" (UniqueName: \"kubernetes.io/projected/5d21c134-833c-4f8d-b2d9-a3c25b2ca597-kube-api-access-5cpfl\") pod \"dnsmasq-dns-6c564ccb4c-s54zm\" (UID: \"5d21c134-833c-4f8d-b2d9-a3c25b2ca597\") " pod="openstack/dnsmasq-dns-6c564ccb4c-s54zm" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.784253 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/03023501-772f-4798-a953-f56dee586cb1-config-data\") pod \"03023501-772f-4798-a953-f56dee586cb1\" (UID: \"03023501-772f-4798-a953-f56dee586cb1\") " Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.784308 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/03023501-772f-4798-a953-f56dee586cb1-rabbitmq-confd\") pod \"03023501-772f-4798-a953-f56dee586cb1\" (UID: \"03023501-772f-4798-a953-f56dee586cb1\") " Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.784385 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"03023501-772f-4798-a953-f56dee586cb1\" (UID: \"03023501-772f-4798-a953-f56dee586cb1\") " Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.784445 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/03023501-772f-4798-a953-f56dee586cb1-rabbitmq-tls\") pod \"03023501-772f-4798-a953-f56dee586cb1\" (UID: \"03023501-772f-4798-a953-f56dee586cb1\") " Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.784469 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/03023501-772f-4798-a953-f56dee586cb1-erlang-cookie-secret\") pod \"03023501-772f-4798-a953-f56dee586cb1\" (UID: \"03023501-772f-4798-a953-f56dee586cb1\") " Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.784488 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/03023501-772f-4798-a953-f56dee586cb1-pod-info\") pod \"03023501-772f-4798-a953-f56dee586cb1\" (UID: \"03023501-772f-4798-a953-f56dee586cb1\") " Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.784544 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/03023501-772f-4798-a953-f56dee586cb1-plugins-conf\") pod \"03023501-772f-4798-a953-f56dee586cb1\" (UID: \"03023501-772f-4798-a953-f56dee586cb1\") " Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.784618 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4q52k\" (UniqueName: \"kubernetes.io/projected/03023501-772f-4798-a953-f56dee586cb1-kube-api-access-4q52k\") pod \"03023501-772f-4798-a953-f56dee586cb1\" (UID: \"03023501-772f-4798-a953-f56dee586cb1\") " Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.784645 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/03023501-772f-4798-a953-f56dee586cb1-server-conf\") pod \"03023501-772f-4798-a953-f56dee586cb1\" (UID: \"03023501-772f-4798-a953-f56dee586cb1\") " Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.784689 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/03023501-772f-4798-a953-f56dee586cb1-rabbitmq-erlang-cookie\") pod \"03023501-772f-4798-a953-f56dee586cb1\" (UID: \"03023501-772f-4798-a953-f56dee586cb1\") " Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.788052 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03023501-772f-4798-a953-f56dee586cb1-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "03023501-772f-4798-a953-f56dee586cb1" (UID: "03023501-772f-4798-a953-f56dee586cb1"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.789391 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/03023501-772f-4798-a953-f56dee586cb1-rabbitmq-plugins\") pod \"03023501-772f-4798-a953-f56dee586cb1\" (UID: \"03023501-772f-4798-a953-f56dee586cb1\") " Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.789701 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d6426d5e-fb37-42e3-8392-3db7d31e81e7-dns-svc\") pod \"dnsmasq-dns-56455489f5-9pk6d\" (UID: \"d6426d5e-fb37-42e3-8392-3db7d31e81e7\") " pod="openstack/dnsmasq-dns-56455489f5-9pk6d" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.789954 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4wmh\" (UniqueName: \"kubernetes.io/projected/d6426d5e-fb37-42e3-8392-3db7d31e81e7-kube-api-access-m4wmh\") pod \"dnsmasq-dns-56455489f5-9pk6d\" (UID: \"d6426d5e-fb37-42e3-8392-3db7d31e81e7\") " pod="openstack/dnsmasq-dns-56455489f5-9pk6d" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.790177 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d6426d5e-fb37-42e3-8392-3db7d31e81e7-dns-swift-storage-0\") pod \"dnsmasq-dns-56455489f5-9pk6d\" (UID: \"d6426d5e-fb37-42e3-8392-3db7d31e81e7\") " pod="openstack/dnsmasq-dns-56455489f5-9pk6d" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.790205 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d6426d5e-fb37-42e3-8392-3db7d31e81e7-ovsdbserver-sb\") pod \"dnsmasq-dns-56455489f5-9pk6d\" (UID: \"d6426d5e-fb37-42e3-8392-3db7d31e81e7\") " pod="openstack/dnsmasq-dns-56455489f5-9pk6d" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.790222 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d6426d5e-fb37-42e3-8392-3db7d31e81e7-ovsdbserver-nb\") pod \"dnsmasq-dns-56455489f5-9pk6d\" (UID: \"d6426d5e-fb37-42e3-8392-3db7d31e81e7\") " pod="openstack/dnsmasq-dns-56455489f5-9pk6d" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.790254 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/d6426d5e-fb37-42e3-8392-3db7d31e81e7-openstack-edpm-ipam\") pod \"dnsmasq-dns-56455489f5-9pk6d\" (UID: \"d6426d5e-fb37-42e3-8392-3db7d31e81e7\") " pod="openstack/dnsmasq-dns-56455489f5-9pk6d" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.790294 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6426d5e-fb37-42e3-8392-3db7d31e81e7-config\") pod \"dnsmasq-dns-56455489f5-9pk6d\" (UID: \"d6426d5e-fb37-42e3-8392-3db7d31e81e7\") " pod="openstack/dnsmasq-dns-56455489f5-9pk6d" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.790365 4556 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/03023501-772f-4798-a953-f56dee586cb1-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.790493 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "persistence") pod "03023501-772f-4798-a953-f56dee586cb1" (UID: "03023501-772f-4798-a953-f56dee586cb1"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.790720 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03023501-772f-4798-a953-f56dee586cb1-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "03023501-772f-4798-a953-f56dee586cb1" (UID: "03023501-772f-4798-a953-f56dee586cb1"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.791444 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03023501-772f-4798-a953-f56dee586cb1-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "03023501-772f-4798-a953-f56dee586cb1" (UID: "03023501-772f-4798-a953-f56dee586cb1"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.792192 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03023501-772f-4798-a953-f56dee586cb1-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "03023501-772f-4798-a953-f56dee586cb1" (UID: "03023501-772f-4798-a953-f56dee586cb1"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.794311 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03023501-772f-4798-a953-f56dee586cb1-kube-api-access-4q52k" (OuterVolumeSpecName: "kube-api-access-4q52k") pod "03023501-772f-4798-a953-f56dee586cb1" (UID: "03023501-772f-4798-a953-f56dee586cb1"). InnerVolumeSpecName "kube-api-access-4q52k". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.796482 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03023501-772f-4798-a953-f56dee586cb1-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "03023501-772f-4798-a953-f56dee586cb1" (UID: "03023501-772f-4798-a953-f56dee586cb1"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.800683 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/03023501-772f-4798-a953-f56dee586cb1-pod-info" (OuterVolumeSpecName: "pod-info") pod "03023501-772f-4798-a953-f56dee586cb1" (UID: "03023501-772f-4798-a953-f56dee586cb1"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.807342 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.818593 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03023501-772f-4798-a953-f56dee586cb1-config-data" (OuterVolumeSpecName: "config-data") pod "03023501-772f-4798-a953-f56dee586cb1" (UID: "03023501-772f-4798-a953-f56dee586cb1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.833618 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03023501-772f-4798-a953-f56dee586cb1-server-conf" (OuterVolumeSpecName: "server-conf") pod "03023501-772f-4798-a953-f56dee586cb1" (UID: "03023501-772f-4798-a953-f56dee586cb1"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.870475 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03023501-772f-4798-a953-f56dee586cb1-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "03023501-772f-4798-a953-f56dee586cb1" (UID: "03023501-772f-4798-a953-f56dee586cb1"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.892084 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6426d5e-fb37-42e3-8392-3db7d31e81e7-config\") pod \"dnsmasq-dns-56455489f5-9pk6d\" (UID: \"d6426d5e-fb37-42e3-8392-3db7d31e81e7\") " pod="openstack/dnsmasq-dns-56455489f5-9pk6d" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.892166 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d6426d5e-fb37-42e3-8392-3db7d31e81e7-dns-svc\") pod \"dnsmasq-dns-56455489f5-9pk6d\" (UID: \"d6426d5e-fb37-42e3-8392-3db7d31e81e7\") " pod="openstack/dnsmasq-dns-56455489f5-9pk6d" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.892323 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4wmh\" (UniqueName: \"kubernetes.io/projected/d6426d5e-fb37-42e3-8392-3db7d31e81e7-kube-api-access-m4wmh\") pod \"dnsmasq-dns-56455489f5-9pk6d\" (UID: \"d6426d5e-fb37-42e3-8392-3db7d31e81e7\") " pod="openstack/dnsmasq-dns-56455489f5-9pk6d" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.892451 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d6426d5e-fb37-42e3-8392-3db7d31e81e7-dns-swift-storage-0\") pod \"dnsmasq-dns-56455489f5-9pk6d\" (UID: \"d6426d5e-fb37-42e3-8392-3db7d31e81e7\") " pod="openstack/dnsmasq-dns-56455489f5-9pk6d" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.892473 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d6426d5e-fb37-42e3-8392-3db7d31e81e7-ovsdbserver-sb\") pod \"dnsmasq-dns-56455489f5-9pk6d\" (UID: \"d6426d5e-fb37-42e3-8392-3db7d31e81e7\") " pod="openstack/dnsmasq-dns-56455489f5-9pk6d" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.892488 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d6426d5e-fb37-42e3-8392-3db7d31e81e7-ovsdbserver-nb\") pod \"dnsmasq-dns-56455489f5-9pk6d\" (UID: \"d6426d5e-fb37-42e3-8392-3db7d31e81e7\") " pod="openstack/dnsmasq-dns-56455489f5-9pk6d" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.892509 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/d6426d5e-fb37-42e3-8392-3db7d31e81e7-openstack-edpm-ipam\") pod \"dnsmasq-dns-56455489f5-9pk6d\" (UID: \"d6426d5e-fb37-42e3-8392-3db7d31e81e7\") " pod="openstack/dnsmasq-dns-56455489f5-9pk6d" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.892568 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4q52k\" (UniqueName: \"kubernetes.io/projected/03023501-772f-4798-a953-f56dee586cb1-kube-api-access-4q52k\") on node \"crc\" DevicePath \"\"" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.892585 4556 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/03023501-772f-4798-a953-f56dee586cb1-server-conf\") on node \"crc\" DevicePath \"\"" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.892594 4556 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/03023501-772f-4798-a953-f56dee586cb1-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.892601 4556 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/03023501-772f-4798-a953-f56dee586cb1-config-data\") on node \"crc\" DevicePath \"\"" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.892609 4556 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/03023501-772f-4798-a953-f56dee586cb1-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.892631 4556 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.892640 4556 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/03023501-772f-4798-a953-f56dee586cb1-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.892650 4556 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/03023501-772f-4798-a953-f56dee586cb1-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.892659 4556 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/03023501-772f-4798-a953-f56dee586cb1-pod-info\") on node \"crc\" DevicePath \"\"" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.892667 4556 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/03023501-772f-4798-a953-f56dee586cb1-plugins-conf\") on node \"crc\" DevicePath \"\"" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.893382 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d6426d5e-fb37-42e3-8392-3db7d31e81e7-dns-swift-storage-0\") pod \"dnsmasq-dns-56455489f5-9pk6d\" (UID: \"d6426d5e-fb37-42e3-8392-3db7d31e81e7\") " pod="openstack/dnsmasq-dns-56455489f5-9pk6d" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.893504 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d6426d5e-fb37-42e3-8392-3db7d31e81e7-ovsdbserver-sb\") pod \"dnsmasq-dns-56455489f5-9pk6d\" (UID: \"d6426d5e-fb37-42e3-8392-3db7d31e81e7\") " pod="openstack/dnsmasq-dns-56455489f5-9pk6d" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.893914 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6426d5e-fb37-42e3-8392-3db7d31e81e7-config\") pod \"dnsmasq-dns-56455489f5-9pk6d\" (UID: \"d6426d5e-fb37-42e3-8392-3db7d31e81e7\") " pod="openstack/dnsmasq-dns-56455489f5-9pk6d" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.894537 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/d6426d5e-fb37-42e3-8392-3db7d31e81e7-openstack-edpm-ipam\") pod \"dnsmasq-dns-56455489f5-9pk6d\" (UID: \"d6426d5e-fb37-42e3-8392-3db7d31e81e7\") " pod="openstack/dnsmasq-dns-56455489f5-9pk6d" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.894892 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d6426d5e-fb37-42e3-8392-3db7d31e81e7-dns-svc\") pod \"dnsmasq-dns-56455489f5-9pk6d\" (UID: \"d6426d5e-fb37-42e3-8392-3db7d31e81e7\") " pod="openstack/dnsmasq-dns-56455489f5-9pk6d" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.895041 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d6426d5e-fb37-42e3-8392-3db7d31e81e7-ovsdbserver-nb\") pod \"dnsmasq-dns-56455489f5-9pk6d\" (UID: \"d6426d5e-fb37-42e3-8392-3db7d31e81e7\") " pod="openstack/dnsmasq-dns-56455489f5-9pk6d" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.910198 4556 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.911545 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4wmh\" (UniqueName: \"kubernetes.io/projected/d6426d5e-fb37-42e3-8392-3db7d31e81e7-kube-api-access-m4wmh\") pod \"dnsmasq-dns-56455489f5-9pk6d\" (UID: \"d6426d5e-fb37-42e3-8392-3db7d31e81e7\") " pod="openstack/dnsmasq-dns-56455489f5-9pk6d" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.985023 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56455489f5-9pk6d" Feb 18 09:21:24 crc kubenswrapper[4556]: I0218 09:21:24.994222 4556 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.290561 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d7961e1-6f96-4707-907d-1fa0c30641e2" path="/var/lib/kubelet/pods/1d7961e1-6f96-4707-907d-1fa0c30641e2/volumes" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.365245 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56455489f5-9pk6d"] Feb 18 09:21:25 crc kubenswrapper[4556]: W0218 09:21:25.405053 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd6426d5e_fb37_42e3_8392_3db7d31e81e7.slice/crio-ef5b272fcfcfd7711d96419fbde89db27373305beaecb4a753f514fcb14bc129 WatchSource:0}: Error finding container ef5b272fcfcfd7711d96419fbde89db27373305beaecb4a753f514fcb14bc129: Status 404 returned error can't find the container with id ef5b272fcfcfd7711d96419fbde89db27373305beaecb4a753f514fcb14bc129 Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.592589 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56455489f5-9pk6d" event={"ID":"d6426d5e-fb37-42e3-8392-3db7d31e81e7","Type":"ContainerStarted","Data":"ef5b272fcfcfd7711d96419fbde89db27373305beaecb4a753f514fcb14bc129"} Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.593815 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0f0ef5d2-a692-435f-a79d-a1af3a294e73","Type":"ContainerStarted","Data":"32487845c1d1b05c5181ce8b04472ee558b009029bd068e06c1e36ee72493e24"} Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.596049 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"03023501-772f-4798-a953-f56dee586cb1","Type":"ContainerDied","Data":"c24a2bd28a1265077ec721a2e2df3311427addb6fdceba6e220564059fc2a066"} Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.596097 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.596120 4556 scope.go:117] "RemoveContainer" containerID="b6d70d7c016d0deed03db8dc523b6b6d3d816ebdd2eb672bcd3ebfca6f608a64" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.596070 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c564ccb4c-s54zm" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.626265 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c564ccb4c-s54zm" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.636778 4556 scope.go:117] "RemoveContainer" containerID="7bf62d08939139f2182730d0a4d9b37f729f09b7b5649c1d303400e1b02489b7" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.641063 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.655896 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.696628 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 18 09:21:25 crc kubenswrapper[4556]: E0218 09:21:25.697055 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03023501-772f-4798-a953-f56dee586cb1" containerName="rabbitmq" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.697076 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="03023501-772f-4798-a953-f56dee586cb1" containerName="rabbitmq" Feb 18 09:21:25 crc kubenswrapper[4556]: E0218 09:21:25.697102 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03023501-772f-4798-a953-f56dee586cb1" containerName="setup-container" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.697121 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="03023501-772f-4798-a953-f56dee586cb1" containerName="setup-container" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.697335 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="03023501-772f-4798-a953-f56dee586cb1" containerName="rabbitmq" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.698322 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.701379 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.701893 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.702129 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.704804 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.704907 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.706442 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.706459 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.706757 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5cpfl\" (UniqueName: \"kubernetes.io/projected/5d21c134-833c-4f8d-b2d9-a3c25b2ca597-kube-api-access-5cpfl\") pod \"5d21c134-833c-4f8d-b2d9-a3c25b2ca597\" (UID: \"5d21c134-833c-4f8d-b2d9-a3c25b2ca597\") " Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.706919 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/5d21c134-833c-4f8d-b2d9-a3c25b2ca597-openstack-edpm-ipam\") pod \"5d21c134-833c-4f8d-b2d9-a3c25b2ca597\" (UID: \"5d21c134-833c-4f8d-b2d9-a3c25b2ca597\") " Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.706985 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5d21c134-833c-4f8d-b2d9-a3c25b2ca597-ovsdbserver-nb\") pod \"5d21c134-833c-4f8d-b2d9-a3c25b2ca597\" (UID: \"5d21c134-833c-4f8d-b2d9-a3c25b2ca597\") " Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.707012 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d21c134-833c-4f8d-b2d9-a3c25b2ca597-config\") pod \"5d21c134-833c-4f8d-b2d9-a3c25b2ca597\" (UID: \"5d21c134-833c-4f8d-b2d9-a3c25b2ca597\") " Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.707065 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5d21c134-833c-4f8d-b2d9-a3c25b2ca597-dns-swift-storage-0\") pod \"5d21c134-833c-4f8d-b2d9-a3c25b2ca597\" (UID: \"5d21c134-833c-4f8d-b2d9-a3c25b2ca597\") " Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.707206 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5d21c134-833c-4f8d-b2d9-a3c25b2ca597-dns-svc\") pod \"5d21c134-833c-4f8d-b2d9-a3c25b2ca597\" (UID: \"5d21c134-833c-4f8d-b2d9-a3c25b2ca597\") " Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.707295 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5d21c134-833c-4f8d-b2d9-a3c25b2ca597-ovsdbserver-sb\") pod \"5d21c134-833c-4f8d-b2d9-a3c25b2ca597\" (UID: \"5d21c134-833c-4f8d-b2d9-a3c25b2ca597\") " Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.708082 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d21c134-833c-4f8d-b2d9-a3c25b2ca597-config" (OuterVolumeSpecName: "config") pod "5d21c134-833c-4f8d-b2d9-a3c25b2ca597" (UID: "5d21c134-833c-4f8d-b2d9-a3c25b2ca597"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.708392 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d21c134-833c-4f8d-b2d9-a3c25b2ca597-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "5d21c134-833c-4f8d-b2d9-a3c25b2ca597" (UID: "5d21c134-833c-4f8d-b2d9-a3c25b2ca597"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.708410 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d21c134-833c-4f8d-b2d9-a3c25b2ca597-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5d21c134-833c-4f8d-b2d9-a3c25b2ca597" (UID: "5d21c134-833c-4f8d-b2d9-a3c25b2ca597"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.708467 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d21c134-833c-4f8d-b2d9-a3c25b2ca597-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5d21c134-833c-4f8d-b2d9-a3c25b2ca597" (UID: "5d21c134-833c-4f8d-b2d9-a3c25b2ca597"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.708921 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d21c134-833c-4f8d-b2d9-a3c25b2ca597-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5d21c134-833c-4f8d-b2d9-a3c25b2ca597" (UID: "5d21c134-833c-4f8d-b2d9-a3c25b2ca597"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.712048 4556 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5d21c134-833c-4f8d-b2d9-a3c25b2ca597-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.712075 4556 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5d21c134-833c-4f8d-b2d9-a3c25b2ca597-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.712085 4556 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5d21c134-833c-4f8d-b2d9-a3c25b2ca597-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.712094 4556 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5d21c134-833c-4f8d-b2d9-a3c25b2ca597-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.712103 4556 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d21c134-833c-4f8d-b2d9-a3c25b2ca597-config\") on node \"crc\" DevicePath \"\"" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.712566 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d21c134-833c-4f8d-b2d9-a3c25b2ca597-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "5d21c134-833c-4f8d-b2d9-a3c25b2ca597" (UID: "5d21c134-833c-4f8d-b2d9-a3c25b2ca597"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.712820 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-ksd85" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.718693 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d21c134-833c-4f8d-b2d9-a3c25b2ca597-kube-api-access-5cpfl" (OuterVolumeSpecName: "kube-api-access-5cpfl") pod "5d21c134-833c-4f8d-b2d9-a3c25b2ca597" (UID: "5d21c134-833c-4f8d-b2d9-a3c25b2ca597"). InnerVolumeSpecName "kube-api-access-5cpfl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.813603 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8b1c0c56-94a0-4ab7-ae4c-5f4035e37359-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"8b1c0c56-94a0-4ab7-ae4c-5f4035e37359\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.813661 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8b1c0c56-94a0-4ab7-ae4c-5f4035e37359-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"8b1c0c56-94a0-4ab7-ae4c-5f4035e37359\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.813689 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8b1c0c56-94a0-4ab7-ae4c-5f4035e37359-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"8b1c0c56-94a0-4ab7-ae4c-5f4035e37359\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.813720 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8b1c0c56-94a0-4ab7-ae4c-5f4035e37359-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8b1c0c56-94a0-4ab7-ae4c-5f4035e37359\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.813737 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jp9xh\" (UniqueName: \"kubernetes.io/projected/8b1c0c56-94a0-4ab7-ae4c-5f4035e37359-kube-api-access-jp9xh\") pod \"rabbitmq-cell1-server-0\" (UID: \"8b1c0c56-94a0-4ab7-ae4c-5f4035e37359\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.813775 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"8b1c0c56-94a0-4ab7-ae4c-5f4035e37359\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.813818 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8b1c0c56-94a0-4ab7-ae4c-5f4035e37359-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"8b1c0c56-94a0-4ab7-ae4c-5f4035e37359\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.813875 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8b1c0c56-94a0-4ab7-ae4c-5f4035e37359-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"8b1c0c56-94a0-4ab7-ae4c-5f4035e37359\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.813918 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8b1c0c56-94a0-4ab7-ae4c-5f4035e37359-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"8b1c0c56-94a0-4ab7-ae4c-5f4035e37359\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.814042 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8b1c0c56-94a0-4ab7-ae4c-5f4035e37359-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"8b1c0c56-94a0-4ab7-ae4c-5f4035e37359\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.814134 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8b1c0c56-94a0-4ab7-ae4c-5f4035e37359-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8b1c0c56-94a0-4ab7-ae4c-5f4035e37359\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.814310 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5cpfl\" (UniqueName: \"kubernetes.io/projected/5d21c134-833c-4f8d-b2d9-a3c25b2ca597-kube-api-access-5cpfl\") on node \"crc\" DevicePath \"\"" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.814323 4556 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/5d21c134-833c-4f8d-b2d9-a3c25b2ca597-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.915509 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8b1c0c56-94a0-4ab7-ae4c-5f4035e37359-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"8b1c0c56-94a0-4ab7-ae4c-5f4035e37359\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.915579 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8b1c0c56-94a0-4ab7-ae4c-5f4035e37359-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"8b1c0c56-94a0-4ab7-ae4c-5f4035e37359\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.915620 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8b1c0c56-94a0-4ab7-ae4c-5f4035e37359-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"8b1c0c56-94a0-4ab7-ae4c-5f4035e37359\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.915643 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8b1c0c56-94a0-4ab7-ae4c-5f4035e37359-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8b1c0c56-94a0-4ab7-ae4c-5f4035e37359\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.915844 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8b1c0c56-94a0-4ab7-ae4c-5f4035e37359-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"8b1c0c56-94a0-4ab7-ae4c-5f4035e37359\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.915898 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8b1c0c56-94a0-4ab7-ae4c-5f4035e37359-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"8b1c0c56-94a0-4ab7-ae4c-5f4035e37359\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.916234 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8b1c0c56-94a0-4ab7-ae4c-5f4035e37359-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"8b1c0c56-94a0-4ab7-ae4c-5f4035e37359\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.917052 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8b1c0c56-94a0-4ab7-ae4c-5f4035e37359-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8b1c0c56-94a0-4ab7-ae4c-5f4035e37359\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.917177 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8b1c0c56-94a0-4ab7-ae4c-5f4035e37359-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"8b1c0c56-94a0-4ab7-ae4c-5f4035e37359\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.917439 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8b1c0c56-94a0-4ab7-ae4c-5f4035e37359-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"8b1c0c56-94a0-4ab7-ae4c-5f4035e37359\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.917483 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8b1c0c56-94a0-4ab7-ae4c-5f4035e37359-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8b1c0c56-94a0-4ab7-ae4c-5f4035e37359\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.917503 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jp9xh\" (UniqueName: \"kubernetes.io/projected/8b1c0c56-94a0-4ab7-ae4c-5f4035e37359-kube-api-access-jp9xh\") pod \"rabbitmq-cell1-server-0\" (UID: \"8b1c0c56-94a0-4ab7-ae4c-5f4035e37359\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.918075 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8b1c0c56-94a0-4ab7-ae4c-5f4035e37359-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8b1c0c56-94a0-4ab7-ae4c-5f4035e37359\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.918271 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"8b1c0c56-94a0-4ab7-ae4c-5f4035e37359\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.918346 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8b1c0c56-94a0-4ab7-ae4c-5f4035e37359-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"8b1c0c56-94a0-4ab7-ae4c-5f4035e37359\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.918667 4556 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"8b1c0c56-94a0-4ab7-ae4c-5f4035e37359\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.919059 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8b1c0c56-94a0-4ab7-ae4c-5f4035e37359-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"8b1c0c56-94a0-4ab7-ae4c-5f4035e37359\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.919103 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8b1c0c56-94a0-4ab7-ae4c-5f4035e37359-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"8b1c0c56-94a0-4ab7-ae4c-5f4035e37359\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.919267 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8b1c0c56-94a0-4ab7-ae4c-5f4035e37359-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"8b1c0c56-94a0-4ab7-ae4c-5f4035e37359\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.919497 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8b1c0c56-94a0-4ab7-ae4c-5f4035e37359-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"8b1c0c56-94a0-4ab7-ae4c-5f4035e37359\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.920543 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8b1c0c56-94a0-4ab7-ae4c-5f4035e37359-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"8b1c0c56-94a0-4ab7-ae4c-5f4035e37359\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.932794 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jp9xh\" (UniqueName: \"kubernetes.io/projected/8b1c0c56-94a0-4ab7-ae4c-5f4035e37359-kube-api-access-jp9xh\") pod \"rabbitmq-cell1-server-0\" (UID: \"8b1c0c56-94a0-4ab7-ae4c-5f4035e37359\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:21:25 crc kubenswrapper[4556]: I0218 09:21:25.941447 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"8b1c0c56-94a0-4ab7-ae4c-5f4035e37359\") " pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:21:26 crc kubenswrapper[4556]: I0218 09:21:26.036309 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:21:26 crc kubenswrapper[4556]: I0218 09:21:26.409936 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 18 09:21:26 crc kubenswrapper[4556]: W0218 09:21:26.416330 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8b1c0c56_94a0_4ab7_ae4c_5f4035e37359.slice/crio-3c070ff86af17addebaee8374b9a1e9f0f244fca84d1f3800b0b2a3fe817d423 WatchSource:0}: Error finding container 3c070ff86af17addebaee8374b9a1e9f0f244fca84d1f3800b0b2a3fe817d423: Status 404 returned error can't find the container with id 3c070ff86af17addebaee8374b9a1e9f0f244fca84d1f3800b0b2a3fe817d423 Feb 18 09:21:26 crc kubenswrapper[4556]: I0218 09:21:26.607489 4556 generic.go:334] "Generic (PLEG): container finished" podID="d6426d5e-fb37-42e3-8392-3db7d31e81e7" containerID="78322359ebeeaab2795cb339950124c5ad9b1cc67df634f8240e46128c152acb" exitCode=0 Feb 18 09:21:26 crc kubenswrapper[4556]: I0218 09:21:26.607597 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56455489f5-9pk6d" event={"ID":"d6426d5e-fb37-42e3-8392-3db7d31e81e7","Type":"ContainerDied","Data":"78322359ebeeaab2795cb339950124c5ad9b1cc67df634f8240e46128c152acb"} Feb 18 09:21:26 crc kubenswrapper[4556]: I0218 09:21:26.609789 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0f0ef5d2-a692-435f-a79d-a1af3a294e73","Type":"ContainerStarted","Data":"bdf9db0bd2af3a9f1ae04d7d3e5141384e3f08f6e87ba80b025fd6e363ac6cbd"} Feb 18 09:21:26 crc kubenswrapper[4556]: I0218 09:21:26.611138 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"8b1c0c56-94a0-4ab7-ae4c-5f4035e37359","Type":"ContainerStarted","Data":"3c070ff86af17addebaee8374b9a1e9f0f244fca84d1f3800b0b2a3fe817d423"} Feb 18 09:21:26 crc kubenswrapper[4556]: I0218 09:21:26.611193 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c564ccb4c-s54zm" Feb 18 09:21:26 crc kubenswrapper[4556]: I0218 09:21:26.720282 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c564ccb4c-s54zm"] Feb 18 09:21:26 crc kubenswrapper[4556]: I0218 09:21:26.733590 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6c564ccb4c-s54zm"] Feb 18 09:21:27 crc kubenswrapper[4556]: I0218 09:21:27.291713 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03023501-772f-4798-a953-f56dee586cb1" path="/var/lib/kubelet/pods/03023501-772f-4798-a953-f56dee586cb1/volumes" Feb 18 09:21:27 crc kubenswrapper[4556]: I0218 09:21:27.292327 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d21c134-833c-4f8d-b2d9-a3c25b2ca597" path="/var/lib/kubelet/pods/5d21c134-833c-4f8d-b2d9-a3c25b2ca597/volumes" Feb 18 09:21:27 crc kubenswrapper[4556]: I0218 09:21:27.621227 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56455489f5-9pk6d" event={"ID":"d6426d5e-fb37-42e3-8392-3db7d31e81e7","Type":"ContainerStarted","Data":"cc08af353d8359d6e06fcf936e03d1e94f3bee7e238f5e76903f96c48634b229"} Feb 18 09:21:27 crc kubenswrapper[4556]: I0218 09:21:27.621688 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-56455489f5-9pk6d" Feb 18 09:21:27 crc kubenswrapper[4556]: I0218 09:21:27.623459 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"8b1c0c56-94a0-4ab7-ae4c-5f4035e37359","Type":"ContainerStarted","Data":"6b2d9f74db15e0ecc2f743a46576423497be226ad6428f0d7d28a74220a1ac91"} Feb 18 09:21:27 crc kubenswrapper[4556]: I0218 09:21:27.640663 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-56455489f5-9pk6d" podStartSLOduration=3.640648049 podStartE2EDuration="3.640648049s" podCreationTimestamp="2026-02-18 09:21:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:21:27.63576114 +0000 UTC m=+1044.652722120" watchObservedRunningTime="2026-02-18 09:21:27.640648049 +0000 UTC m=+1044.657609029" Feb 18 09:21:34 crc kubenswrapper[4556]: I0218 09:21:34.986977 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-56455489f5-9pk6d" Feb 18 09:21:35 crc kubenswrapper[4556]: I0218 09:21:35.027036 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74b99f9475-7n7kb"] Feb 18 09:21:35 crc kubenswrapper[4556]: I0218 09:21:35.030523 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-74b99f9475-7n7kb" podUID="a7122b24-4132-4967-a132-631e9ec2a11a" containerName="dnsmasq-dns" containerID="cri-o://8bf3c5eb66360572cb8dde27f0e8c31413e0e5c2c108fc1a73999f8e43241056" gracePeriod=10 Feb 18 09:21:35 crc kubenswrapper[4556]: I0218 09:21:35.432724 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74b99f9475-7n7kb" Feb 18 09:21:35 crc kubenswrapper[4556]: I0218 09:21:35.480211 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7122b24-4132-4967-a132-631e9ec2a11a-config\") pod \"a7122b24-4132-4967-a132-631e9ec2a11a\" (UID: \"a7122b24-4132-4967-a132-631e9ec2a11a\") " Feb 18 09:21:35 crc kubenswrapper[4556]: I0218 09:21:35.480329 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a7122b24-4132-4967-a132-631e9ec2a11a-dns-swift-storage-0\") pod \"a7122b24-4132-4967-a132-631e9ec2a11a\" (UID: \"a7122b24-4132-4967-a132-631e9ec2a11a\") " Feb 18 09:21:35 crc kubenswrapper[4556]: I0218 09:21:35.480351 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a7122b24-4132-4967-a132-631e9ec2a11a-ovsdbserver-sb\") pod \"a7122b24-4132-4967-a132-631e9ec2a11a\" (UID: \"a7122b24-4132-4967-a132-631e9ec2a11a\") " Feb 18 09:21:35 crc kubenswrapper[4556]: I0218 09:21:35.480395 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2flzl\" (UniqueName: \"kubernetes.io/projected/a7122b24-4132-4967-a132-631e9ec2a11a-kube-api-access-2flzl\") pod \"a7122b24-4132-4967-a132-631e9ec2a11a\" (UID: \"a7122b24-4132-4967-a132-631e9ec2a11a\") " Feb 18 09:21:35 crc kubenswrapper[4556]: I0218 09:21:35.480518 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a7122b24-4132-4967-a132-631e9ec2a11a-dns-svc\") pod \"a7122b24-4132-4967-a132-631e9ec2a11a\" (UID: \"a7122b24-4132-4967-a132-631e9ec2a11a\") " Feb 18 09:21:35 crc kubenswrapper[4556]: I0218 09:21:35.480574 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a7122b24-4132-4967-a132-631e9ec2a11a-ovsdbserver-nb\") pod \"a7122b24-4132-4967-a132-631e9ec2a11a\" (UID: \"a7122b24-4132-4967-a132-631e9ec2a11a\") " Feb 18 09:21:35 crc kubenswrapper[4556]: I0218 09:21:35.486853 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7122b24-4132-4967-a132-631e9ec2a11a-kube-api-access-2flzl" (OuterVolumeSpecName: "kube-api-access-2flzl") pod "a7122b24-4132-4967-a132-631e9ec2a11a" (UID: "a7122b24-4132-4967-a132-631e9ec2a11a"). InnerVolumeSpecName "kube-api-access-2flzl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:21:35 crc kubenswrapper[4556]: I0218 09:21:35.521130 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7122b24-4132-4967-a132-631e9ec2a11a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a7122b24-4132-4967-a132-631e9ec2a11a" (UID: "a7122b24-4132-4967-a132-631e9ec2a11a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:21:35 crc kubenswrapper[4556]: I0218 09:21:35.522706 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7122b24-4132-4967-a132-631e9ec2a11a-config" (OuterVolumeSpecName: "config") pod "a7122b24-4132-4967-a132-631e9ec2a11a" (UID: "a7122b24-4132-4967-a132-631e9ec2a11a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:21:35 crc kubenswrapper[4556]: I0218 09:21:35.522909 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7122b24-4132-4967-a132-631e9ec2a11a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a7122b24-4132-4967-a132-631e9ec2a11a" (UID: "a7122b24-4132-4967-a132-631e9ec2a11a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:21:35 crc kubenswrapper[4556]: I0218 09:21:35.523292 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7122b24-4132-4967-a132-631e9ec2a11a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a7122b24-4132-4967-a132-631e9ec2a11a" (UID: "a7122b24-4132-4967-a132-631e9ec2a11a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:21:35 crc kubenswrapper[4556]: I0218 09:21:35.530492 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7122b24-4132-4967-a132-631e9ec2a11a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "a7122b24-4132-4967-a132-631e9ec2a11a" (UID: "a7122b24-4132-4967-a132-631e9ec2a11a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:21:35 crc kubenswrapper[4556]: I0218 09:21:35.583345 4556 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a7122b24-4132-4967-a132-631e9ec2a11a-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 18 09:21:35 crc kubenswrapper[4556]: I0218 09:21:35.583371 4556 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a7122b24-4132-4967-a132-631e9ec2a11a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 18 09:21:35 crc kubenswrapper[4556]: I0218 09:21:35.583381 4556 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7122b24-4132-4967-a132-631e9ec2a11a-config\") on node \"crc\" DevicePath \"\"" Feb 18 09:21:35 crc kubenswrapper[4556]: I0218 09:21:35.583388 4556 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a7122b24-4132-4967-a132-631e9ec2a11a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 18 09:21:35 crc kubenswrapper[4556]: I0218 09:21:35.583398 4556 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a7122b24-4132-4967-a132-631e9ec2a11a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 18 09:21:35 crc kubenswrapper[4556]: I0218 09:21:35.583406 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2flzl\" (UniqueName: \"kubernetes.io/projected/a7122b24-4132-4967-a132-631e9ec2a11a-kube-api-access-2flzl\") on node \"crc\" DevicePath \"\"" Feb 18 09:21:35 crc kubenswrapper[4556]: I0218 09:21:35.681271 4556 generic.go:334] "Generic (PLEG): container finished" podID="a7122b24-4132-4967-a132-631e9ec2a11a" containerID="8bf3c5eb66360572cb8dde27f0e8c31413e0e5c2c108fc1a73999f8e43241056" exitCode=0 Feb 18 09:21:35 crc kubenswrapper[4556]: I0218 09:21:35.681320 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74b99f9475-7n7kb" event={"ID":"a7122b24-4132-4967-a132-631e9ec2a11a","Type":"ContainerDied","Data":"8bf3c5eb66360572cb8dde27f0e8c31413e0e5c2c108fc1a73999f8e43241056"} Feb 18 09:21:35 crc kubenswrapper[4556]: I0218 09:21:35.681452 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74b99f9475-7n7kb" event={"ID":"a7122b24-4132-4967-a132-631e9ec2a11a","Type":"ContainerDied","Data":"eec59aeb42f6a1cfdcc9f8ebe6b85c1d01c207de75c7cbc46c24134ad8d66349"} Feb 18 09:21:35 crc kubenswrapper[4556]: I0218 09:21:35.681474 4556 scope.go:117] "RemoveContainer" containerID="8bf3c5eb66360572cb8dde27f0e8c31413e0e5c2c108fc1a73999f8e43241056" Feb 18 09:21:35 crc kubenswrapper[4556]: I0218 09:21:35.681343 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74b99f9475-7n7kb" Feb 18 09:21:35 crc kubenswrapper[4556]: I0218 09:21:35.697734 4556 scope.go:117] "RemoveContainer" containerID="b7eeec31510c3a6176773eb9808a367f07b7ee25d6e1167b0475f5e8add8aaba" Feb 18 09:21:35 crc kubenswrapper[4556]: I0218 09:21:35.709506 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74b99f9475-7n7kb"] Feb 18 09:21:35 crc kubenswrapper[4556]: I0218 09:21:35.717135 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-74b99f9475-7n7kb"] Feb 18 09:21:35 crc kubenswrapper[4556]: I0218 09:21:35.738128 4556 scope.go:117] "RemoveContainer" containerID="8bf3c5eb66360572cb8dde27f0e8c31413e0e5c2c108fc1a73999f8e43241056" Feb 18 09:21:35 crc kubenswrapper[4556]: E0218 09:21:35.738599 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8bf3c5eb66360572cb8dde27f0e8c31413e0e5c2c108fc1a73999f8e43241056\": container with ID starting with 8bf3c5eb66360572cb8dde27f0e8c31413e0e5c2c108fc1a73999f8e43241056 not found: ID does not exist" containerID="8bf3c5eb66360572cb8dde27f0e8c31413e0e5c2c108fc1a73999f8e43241056" Feb 18 09:21:35 crc kubenswrapper[4556]: I0218 09:21:35.738679 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8bf3c5eb66360572cb8dde27f0e8c31413e0e5c2c108fc1a73999f8e43241056"} err="failed to get container status \"8bf3c5eb66360572cb8dde27f0e8c31413e0e5c2c108fc1a73999f8e43241056\": rpc error: code = NotFound desc = could not find container \"8bf3c5eb66360572cb8dde27f0e8c31413e0e5c2c108fc1a73999f8e43241056\": container with ID starting with 8bf3c5eb66360572cb8dde27f0e8c31413e0e5c2c108fc1a73999f8e43241056 not found: ID does not exist" Feb 18 09:21:35 crc kubenswrapper[4556]: I0218 09:21:35.738750 4556 scope.go:117] "RemoveContainer" containerID="b7eeec31510c3a6176773eb9808a367f07b7ee25d6e1167b0475f5e8add8aaba" Feb 18 09:21:35 crc kubenswrapper[4556]: E0218 09:21:35.739134 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7eeec31510c3a6176773eb9808a367f07b7ee25d6e1167b0475f5e8add8aaba\": container with ID starting with b7eeec31510c3a6176773eb9808a367f07b7ee25d6e1167b0475f5e8add8aaba not found: ID does not exist" containerID="b7eeec31510c3a6176773eb9808a367f07b7ee25d6e1167b0475f5e8add8aaba" Feb 18 09:21:35 crc kubenswrapper[4556]: I0218 09:21:35.739184 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7eeec31510c3a6176773eb9808a367f07b7ee25d6e1167b0475f5e8add8aaba"} err="failed to get container status \"b7eeec31510c3a6176773eb9808a367f07b7ee25d6e1167b0475f5e8add8aaba\": rpc error: code = NotFound desc = could not find container \"b7eeec31510c3a6176773eb9808a367f07b7ee25d6e1167b0475f5e8add8aaba\": container with ID starting with b7eeec31510c3a6176773eb9808a367f07b7ee25d6e1167b0475f5e8add8aaba not found: ID does not exist" Feb 18 09:21:37 crc kubenswrapper[4556]: I0218 09:21:37.290396 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7122b24-4132-4967-a132-631e9ec2a11a" path="/var/lib/kubelet/pods/a7122b24-4132-4967-a132-631e9ec2a11a/volumes" Feb 18 09:21:43 crc kubenswrapper[4556]: I0218 09:21:43.436531 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-sqqbc"] Feb 18 09:21:43 crc kubenswrapper[4556]: E0218 09:21:43.436881 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7122b24-4132-4967-a132-631e9ec2a11a" containerName="dnsmasq-dns" Feb 18 09:21:43 crc kubenswrapper[4556]: I0218 09:21:43.436894 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7122b24-4132-4967-a132-631e9ec2a11a" containerName="dnsmasq-dns" Feb 18 09:21:43 crc kubenswrapper[4556]: E0218 09:21:43.436913 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7122b24-4132-4967-a132-631e9ec2a11a" containerName="init" Feb 18 09:21:43 crc kubenswrapper[4556]: I0218 09:21:43.436919 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7122b24-4132-4967-a132-631e9ec2a11a" containerName="init" Feb 18 09:21:43 crc kubenswrapper[4556]: I0218 09:21:43.437097 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7122b24-4132-4967-a132-631e9ec2a11a" containerName="dnsmasq-dns" Feb 18 09:21:43 crc kubenswrapper[4556]: I0218 09:21:43.437689 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-sqqbc" Feb 18 09:21:43 crc kubenswrapper[4556]: I0218 09:21:43.439804 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-54z7h" Feb 18 09:21:43 crc kubenswrapper[4556]: I0218 09:21:43.441178 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 18 09:21:43 crc kubenswrapper[4556]: I0218 09:21:43.441200 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 18 09:21:43 crc kubenswrapper[4556]: I0218 09:21:43.441840 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 18 09:21:43 crc kubenswrapper[4556]: I0218 09:21:43.451840 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-sqqbc"] Feb 18 09:21:43 crc kubenswrapper[4556]: I0218 09:21:43.515185 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2b34eeb-f211-4454-a74d-207de2fa9f13-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-sqqbc\" (UID: \"c2b34eeb-f211-4454-a74d-207de2fa9f13\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-sqqbc" Feb 18 09:21:43 crc kubenswrapper[4556]: I0218 09:21:43.515251 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c2b34eeb-f211-4454-a74d-207de2fa9f13-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-sqqbc\" (UID: \"c2b34eeb-f211-4454-a74d-207de2fa9f13\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-sqqbc" Feb 18 09:21:43 crc kubenswrapper[4556]: I0218 09:21:43.515351 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2b34eeb-f211-4454-a74d-207de2fa9f13-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-sqqbc\" (UID: \"c2b34eeb-f211-4454-a74d-207de2fa9f13\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-sqqbc" Feb 18 09:21:43 crc kubenswrapper[4556]: I0218 09:21:43.515382 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjgpl\" (UniqueName: \"kubernetes.io/projected/c2b34eeb-f211-4454-a74d-207de2fa9f13-kube-api-access-zjgpl\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-sqqbc\" (UID: \"c2b34eeb-f211-4454-a74d-207de2fa9f13\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-sqqbc" Feb 18 09:21:43 crc kubenswrapper[4556]: I0218 09:21:43.617245 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2b34eeb-f211-4454-a74d-207de2fa9f13-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-sqqbc\" (UID: \"c2b34eeb-f211-4454-a74d-207de2fa9f13\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-sqqbc" Feb 18 09:21:43 crc kubenswrapper[4556]: I0218 09:21:43.617656 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c2b34eeb-f211-4454-a74d-207de2fa9f13-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-sqqbc\" (UID: \"c2b34eeb-f211-4454-a74d-207de2fa9f13\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-sqqbc" Feb 18 09:21:43 crc kubenswrapper[4556]: I0218 09:21:43.617766 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2b34eeb-f211-4454-a74d-207de2fa9f13-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-sqqbc\" (UID: \"c2b34eeb-f211-4454-a74d-207de2fa9f13\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-sqqbc" Feb 18 09:21:43 crc kubenswrapper[4556]: I0218 09:21:43.617802 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjgpl\" (UniqueName: \"kubernetes.io/projected/c2b34eeb-f211-4454-a74d-207de2fa9f13-kube-api-access-zjgpl\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-sqqbc\" (UID: \"c2b34eeb-f211-4454-a74d-207de2fa9f13\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-sqqbc" Feb 18 09:21:43 crc kubenswrapper[4556]: I0218 09:21:43.624062 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2b34eeb-f211-4454-a74d-207de2fa9f13-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-sqqbc\" (UID: \"c2b34eeb-f211-4454-a74d-207de2fa9f13\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-sqqbc" Feb 18 09:21:43 crc kubenswrapper[4556]: I0218 09:21:43.624328 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2b34eeb-f211-4454-a74d-207de2fa9f13-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-sqqbc\" (UID: \"c2b34eeb-f211-4454-a74d-207de2fa9f13\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-sqqbc" Feb 18 09:21:43 crc kubenswrapper[4556]: I0218 09:21:43.624878 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c2b34eeb-f211-4454-a74d-207de2fa9f13-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-sqqbc\" (UID: \"c2b34eeb-f211-4454-a74d-207de2fa9f13\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-sqqbc" Feb 18 09:21:43 crc kubenswrapper[4556]: I0218 09:21:43.632882 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjgpl\" (UniqueName: \"kubernetes.io/projected/c2b34eeb-f211-4454-a74d-207de2fa9f13-kube-api-access-zjgpl\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-sqqbc\" (UID: \"c2b34eeb-f211-4454-a74d-207de2fa9f13\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-sqqbc" Feb 18 09:21:43 crc kubenswrapper[4556]: I0218 09:21:43.756239 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-sqqbc" Feb 18 09:21:44 crc kubenswrapper[4556]: I0218 09:21:44.188921 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-sqqbc"] Feb 18 09:21:44 crc kubenswrapper[4556]: I0218 09:21:44.761873 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-sqqbc" event={"ID":"c2b34eeb-f211-4454-a74d-207de2fa9f13","Type":"ContainerStarted","Data":"53511d05331a8461b3c14e3ecd6ab9345a584724d750b0fdc0e0db29b7627dcc"} Feb 18 09:21:51 crc kubenswrapper[4556]: I0218 09:21:51.816250 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-sqqbc" event={"ID":"c2b34eeb-f211-4454-a74d-207de2fa9f13","Type":"ContainerStarted","Data":"1a7a94cc61c56352ec2c761a2eb014bc120030d2fedeb4030ae7812a599700cd"} Feb 18 09:21:51 crc kubenswrapper[4556]: I0218 09:21:51.833693 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-sqqbc" podStartSLOduration=2.236300125 podStartE2EDuration="8.83367921s" podCreationTimestamp="2026-02-18 09:21:43 +0000 UTC" firstStartedPulling="2026-02-18 09:21:44.193308605 +0000 UTC m=+1061.210269585" lastFinishedPulling="2026-02-18 09:21:50.790687691 +0000 UTC m=+1067.807648670" observedRunningTime="2026-02-18 09:21:51.827168397 +0000 UTC m=+1068.844129378" watchObservedRunningTime="2026-02-18 09:21:51.83367921 +0000 UTC m=+1068.850640190" Feb 18 09:21:57 crc kubenswrapper[4556]: I0218 09:21:57.866366 4556 generic.go:334] "Generic (PLEG): container finished" podID="0f0ef5d2-a692-435f-a79d-a1af3a294e73" containerID="bdf9db0bd2af3a9f1ae04d7d3e5141384e3f08f6e87ba80b025fd6e363ac6cbd" exitCode=0 Feb 18 09:21:57 crc kubenswrapper[4556]: I0218 09:21:57.866450 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0f0ef5d2-a692-435f-a79d-a1af3a294e73","Type":"ContainerDied","Data":"bdf9db0bd2af3a9f1ae04d7d3e5141384e3f08f6e87ba80b025fd6e363ac6cbd"} Feb 18 09:21:58 crc kubenswrapper[4556]: I0218 09:21:58.877734 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0f0ef5d2-a692-435f-a79d-a1af3a294e73","Type":"ContainerStarted","Data":"7d78b26a37fb8e319133263963d6e666537c3ef095e86b70f5a7e961192e56d9"} Feb 18 09:21:58 crc kubenswrapper[4556]: I0218 09:21:58.878242 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Feb 18 09:21:58 crc kubenswrapper[4556]: I0218 09:21:58.900732 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=35.900711964 podStartE2EDuration="35.900711964s" podCreationTimestamp="2026-02-18 09:21:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:21:58.897571481 +0000 UTC m=+1075.914532461" watchObservedRunningTime="2026-02-18 09:21:58.900711964 +0000 UTC m=+1075.917672945" Feb 18 09:21:59 crc kubenswrapper[4556]: I0218 09:21:59.887824 4556 generic.go:334] "Generic (PLEG): container finished" podID="8b1c0c56-94a0-4ab7-ae4c-5f4035e37359" containerID="6b2d9f74db15e0ecc2f743a46576423497be226ad6428f0d7d28a74220a1ac91" exitCode=0 Feb 18 09:21:59 crc kubenswrapper[4556]: I0218 09:21:59.888286 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"8b1c0c56-94a0-4ab7-ae4c-5f4035e37359","Type":"ContainerDied","Data":"6b2d9f74db15e0ecc2f743a46576423497be226ad6428f0d7d28a74220a1ac91"} Feb 18 09:22:00 crc kubenswrapper[4556]: I0218 09:22:00.901122 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"8b1c0c56-94a0-4ab7-ae4c-5f4035e37359","Type":"ContainerStarted","Data":"f934110d3643b900e8371f4301e98c88b6703c6f048dddc6435b791a3886315a"} Feb 18 09:22:00 crc kubenswrapper[4556]: I0218 09:22:00.902181 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:22:00 crc kubenswrapper[4556]: I0218 09:22:00.929638 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=35.929622426 podStartE2EDuration="35.929622426s" podCreationTimestamp="2026-02-18 09:21:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:22:00.922987138 +0000 UTC m=+1077.939948119" watchObservedRunningTime="2026-02-18 09:22:00.929622426 +0000 UTC m=+1077.946583396" Feb 18 09:22:01 crc kubenswrapper[4556]: I0218 09:22:01.727349 4556 patch_prober.go:28] interesting pod/machine-config-daemon-f76hs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 09:22:01 crc kubenswrapper[4556]: I0218 09:22:01.727703 4556 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 09:22:01 crc kubenswrapper[4556]: I0218 09:22:01.914860 4556 generic.go:334] "Generic (PLEG): container finished" podID="c2b34eeb-f211-4454-a74d-207de2fa9f13" containerID="1a7a94cc61c56352ec2c761a2eb014bc120030d2fedeb4030ae7812a599700cd" exitCode=0 Feb 18 09:22:01 crc kubenswrapper[4556]: I0218 09:22:01.914958 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-sqqbc" event={"ID":"c2b34eeb-f211-4454-a74d-207de2fa9f13","Type":"ContainerDied","Data":"1a7a94cc61c56352ec2c761a2eb014bc120030d2fedeb4030ae7812a599700cd"} Feb 18 09:22:03 crc kubenswrapper[4556]: I0218 09:22:03.308798 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-sqqbc" Feb 18 09:22:03 crc kubenswrapper[4556]: I0218 09:22:03.405022 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zjgpl\" (UniqueName: \"kubernetes.io/projected/c2b34eeb-f211-4454-a74d-207de2fa9f13-kube-api-access-zjgpl\") pod \"c2b34eeb-f211-4454-a74d-207de2fa9f13\" (UID: \"c2b34eeb-f211-4454-a74d-207de2fa9f13\") " Feb 18 09:22:03 crc kubenswrapper[4556]: I0218 09:22:03.405348 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2b34eeb-f211-4454-a74d-207de2fa9f13-inventory\") pod \"c2b34eeb-f211-4454-a74d-207de2fa9f13\" (UID: \"c2b34eeb-f211-4454-a74d-207de2fa9f13\") " Feb 18 09:22:03 crc kubenswrapper[4556]: I0218 09:22:03.405475 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c2b34eeb-f211-4454-a74d-207de2fa9f13-ssh-key-openstack-edpm-ipam\") pod \"c2b34eeb-f211-4454-a74d-207de2fa9f13\" (UID: \"c2b34eeb-f211-4454-a74d-207de2fa9f13\") " Feb 18 09:22:03 crc kubenswrapper[4556]: I0218 09:22:03.405506 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2b34eeb-f211-4454-a74d-207de2fa9f13-repo-setup-combined-ca-bundle\") pod \"c2b34eeb-f211-4454-a74d-207de2fa9f13\" (UID: \"c2b34eeb-f211-4454-a74d-207de2fa9f13\") " Feb 18 09:22:03 crc kubenswrapper[4556]: I0218 09:22:03.411682 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2b34eeb-f211-4454-a74d-207de2fa9f13-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "c2b34eeb-f211-4454-a74d-207de2fa9f13" (UID: "c2b34eeb-f211-4454-a74d-207de2fa9f13"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:22:03 crc kubenswrapper[4556]: I0218 09:22:03.421377 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2b34eeb-f211-4454-a74d-207de2fa9f13-kube-api-access-zjgpl" (OuterVolumeSpecName: "kube-api-access-zjgpl") pod "c2b34eeb-f211-4454-a74d-207de2fa9f13" (UID: "c2b34eeb-f211-4454-a74d-207de2fa9f13"). InnerVolumeSpecName "kube-api-access-zjgpl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:22:03 crc kubenswrapper[4556]: I0218 09:22:03.434511 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2b34eeb-f211-4454-a74d-207de2fa9f13-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "c2b34eeb-f211-4454-a74d-207de2fa9f13" (UID: "c2b34eeb-f211-4454-a74d-207de2fa9f13"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:22:03 crc kubenswrapper[4556]: I0218 09:22:03.437183 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2b34eeb-f211-4454-a74d-207de2fa9f13-inventory" (OuterVolumeSpecName: "inventory") pod "c2b34eeb-f211-4454-a74d-207de2fa9f13" (UID: "c2b34eeb-f211-4454-a74d-207de2fa9f13"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:22:03 crc kubenswrapper[4556]: I0218 09:22:03.508751 4556 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2b34eeb-f211-4454-a74d-207de2fa9f13-inventory\") on node \"crc\" DevicePath \"\"" Feb 18 09:22:03 crc kubenswrapper[4556]: I0218 09:22:03.508791 4556 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c2b34eeb-f211-4454-a74d-207de2fa9f13-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 18 09:22:03 crc kubenswrapper[4556]: I0218 09:22:03.508803 4556 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2b34eeb-f211-4454-a74d-207de2fa9f13-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:22:03 crc kubenswrapper[4556]: I0218 09:22:03.508812 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zjgpl\" (UniqueName: \"kubernetes.io/projected/c2b34eeb-f211-4454-a74d-207de2fa9f13-kube-api-access-zjgpl\") on node \"crc\" DevicePath \"\"" Feb 18 09:22:03 crc kubenswrapper[4556]: I0218 09:22:03.937715 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-sqqbc" event={"ID":"c2b34eeb-f211-4454-a74d-207de2fa9f13","Type":"ContainerDied","Data":"53511d05331a8461b3c14e3ecd6ab9345a584724d750b0fdc0e0db29b7627dcc"} Feb 18 09:22:03 crc kubenswrapper[4556]: I0218 09:22:03.938090 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="53511d05331a8461b3c14e3ecd6ab9345a584724d750b0fdc0e0db29b7627dcc" Feb 18 09:22:03 crc kubenswrapper[4556]: I0218 09:22:03.937782 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-sqqbc" Feb 18 09:22:04 crc kubenswrapper[4556]: I0218 09:22:04.003250 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-hjswl"] Feb 18 09:22:04 crc kubenswrapper[4556]: E0218 09:22:04.003730 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2b34eeb-f211-4454-a74d-207de2fa9f13" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 18 09:22:04 crc kubenswrapper[4556]: I0218 09:22:04.003746 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2b34eeb-f211-4454-a74d-207de2fa9f13" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 18 09:22:04 crc kubenswrapper[4556]: I0218 09:22:04.003976 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2b34eeb-f211-4454-a74d-207de2fa9f13" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 18 09:22:04 crc kubenswrapper[4556]: I0218 09:22:04.004688 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-hjswl" Feb 18 09:22:04 crc kubenswrapper[4556]: I0218 09:22:04.006505 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 18 09:22:04 crc kubenswrapper[4556]: I0218 09:22:04.008958 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-54z7h" Feb 18 09:22:04 crc kubenswrapper[4556]: I0218 09:22:04.009487 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 18 09:22:04 crc kubenswrapper[4556]: I0218 09:22:04.010355 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 18 09:22:04 crc kubenswrapper[4556]: I0218 09:22:04.015068 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-hjswl"] Feb 18 09:22:04 crc kubenswrapper[4556]: I0218 09:22:04.019678 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3ebf502b-e4a2-4d25-8551-f1deed31b5c9-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-hjswl\" (UID: \"3ebf502b-e4a2-4d25-8551-f1deed31b5c9\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-hjswl" Feb 18 09:22:04 crc kubenswrapper[4556]: I0218 09:22:04.019768 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvvsz\" (UniqueName: \"kubernetes.io/projected/3ebf502b-e4a2-4d25-8551-f1deed31b5c9-kube-api-access-lvvsz\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-hjswl\" (UID: \"3ebf502b-e4a2-4d25-8551-f1deed31b5c9\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-hjswl" Feb 18 09:22:04 crc kubenswrapper[4556]: I0218 09:22:04.019798 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3ebf502b-e4a2-4d25-8551-f1deed31b5c9-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-hjswl\" (UID: \"3ebf502b-e4a2-4d25-8551-f1deed31b5c9\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-hjswl" Feb 18 09:22:04 crc kubenswrapper[4556]: I0218 09:22:04.122920 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvvsz\" (UniqueName: \"kubernetes.io/projected/3ebf502b-e4a2-4d25-8551-f1deed31b5c9-kube-api-access-lvvsz\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-hjswl\" (UID: \"3ebf502b-e4a2-4d25-8551-f1deed31b5c9\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-hjswl" Feb 18 09:22:04 crc kubenswrapper[4556]: I0218 09:22:04.123114 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3ebf502b-e4a2-4d25-8551-f1deed31b5c9-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-hjswl\" (UID: \"3ebf502b-e4a2-4d25-8551-f1deed31b5c9\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-hjswl" Feb 18 09:22:04 crc kubenswrapper[4556]: I0218 09:22:04.123663 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3ebf502b-e4a2-4d25-8551-f1deed31b5c9-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-hjswl\" (UID: \"3ebf502b-e4a2-4d25-8551-f1deed31b5c9\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-hjswl" Feb 18 09:22:04 crc kubenswrapper[4556]: I0218 09:22:04.127327 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3ebf502b-e4a2-4d25-8551-f1deed31b5c9-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-hjswl\" (UID: \"3ebf502b-e4a2-4d25-8551-f1deed31b5c9\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-hjswl" Feb 18 09:22:04 crc kubenswrapper[4556]: I0218 09:22:04.127410 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3ebf502b-e4a2-4d25-8551-f1deed31b5c9-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-hjswl\" (UID: \"3ebf502b-e4a2-4d25-8551-f1deed31b5c9\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-hjswl" Feb 18 09:22:04 crc kubenswrapper[4556]: I0218 09:22:04.138905 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvvsz\" (UniqueName: \"kubernetes.io/projected/3ebf502b-e4a2-4d25-8551-f1deed31b5c9-kube-api-access-lvvsz\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-hjswl\" (UID: \"3ebf502b-e4a2-4d25-8551-f1deed31b5c9\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-hjswl" Feb 18 09:22:04 crc kubenswrapper[4556]: I0218 09:22:04.321008 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-hjswl" Feb 18 09:22:04 crc kubenswrapper[4556]: I0218 09:22:04.814653 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-hjswl"] Feb 18 09:22:04 crc kubenswrapper[4556]: W0218 09:22:04.818856 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3ebf502b_e4a2_4d25_8551_f1deed31b5c9.slice/crio-d30854a28f8152b7436ba3d9671081b16709ee62110af8a1230be2fd054d6207 WatchSource:0}: Error finding container d30854a28f8152b7436ba3d9671081b16709ee62110af8a1230be2fd054d6207: Status 404 returned error can't find the container with id d30854a28f8152b7436ba3d9671081b16709ee62110af8a1230be2fd054d6207 Feb 18 09:22:04 crc kubenswrapper[4556]: I0218 09:22:04.822999 4556 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 18 09:22:04 crc kubenswrapper[4556]: I0218 09:22:04.950591 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-hjswl" event={"ID":"3ebf502b-e4a2-4d25-8551-f1deed31b5c9","Type":"ContainerStarted","Data":"d30854a28f8152b7436ba3d9671081b16709ee62110af8a1230be2fd054d6207"} Feb 18 09:22:05 crc kubenswrapper[4556]: I0218 09:22:05.966576 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-hjswl" event={"ID":"3ebf502b-e4a2-4d25-8551-f1deed31b5c9","Type":"ContainerStarted","Data":"5b43d36b50ca7615f31c7d14a917e1a41319ad6f1637d05f563f24072ddd3e49"} Feb 18 09:22:05 crc kubenswrapper[4556]: I0218 09:22:05.988624 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-hjswl" podStartSLOduration=2.40385331 podStartE2EDuration="2.988596117s" podCreationTimestamp="2026-02-18 09:22:03 +0000 UTC" firstStartedPulling="2026-02-18 09:22:04.822792761 +0000 UTC m=+1081.839753741" lastFinishedPulling="2026-02-18 09:22:05.407535568 +0000 UTC m=+1082.424496548" observedRunningTime="2026-02-18 09:22:05.986228611 +0000 UTC m=+1083.003189591" watchObservedRunningTime="2026-02-18 09:22:05.988596117 +0000 UTC m=+1083.005557097" Feb 18 09:22:07 crc kubenswrapper[4556]: I0218 09:22:07.986487 4556 generic.go:334] "Generic (PLEG): container finished" podID="3ebf502b-e4a2-4d25-8551-f1deed31b5c9" containerID="5b43d36b50ca7615f31c7d14a917e1a41319ad6f1637d05f563f24072ddd3e49" exitCode=0 Feb 18 09:22:07 crc kubenswrapper[4556]: I0218 09:22:07.986540 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-hjswl" event={"ID":"3ebf502b-e4a2-4d25-8551-f1deed31b5c9","Type":"ContainerDied","Data":"5b43d36b50ca7615f31c7d14a917e1a41319ad6f1637d05f563f24072ddd3e49"} Feb 18 09:22:09 crc kubenswrapper[4556]: I0218 09:22:09.347469 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-hjswl" Feb 18 09:22:09 crc kubenswrapper[4556]: I0218 09:22:09.442549 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3ebf502b-e4a2-4d25-8551-f1deed31b5c9-inventory\") pod \"3ebf502b-e4a2-4d25-8551-f1deed31b5c9\" (UID: \"3ebf502b-e4a2-4d25-8551-f1deed31b5c9\") " Feb 18 09:22:09 crc kubenswrapper[4556]: I0218 09:22:09.442778 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lvvsz\" (UniqueName: \"kubernetes.io/projected/3ebf502b-e4a2-4d25-8551-f1deed31b5c9-kube-api-access-lvvsz\") pod \"3ebf502b-e4a2-4d25-8551-f1deed31b5c9\" (UID: \"3ebf502b-e4a2-4d25-8551-f1deed31b5c9\") " Feb 18 09:22:09 crc kubenswrapper[4556]: I0218 09:22:09.442801 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3ebf502b-e4a2-4d25-8551-f1deed31b5c9-ssh-key-openstack-edpm-ipam\") pod \"3ebf502b-e4a2-4d25-8551-f1deed31b5c9\" (UID: \"3ebf502b-e4a2-4d25-8551-f1deed31b5c9\") " Feb 18 09:22:09 crc kubenswrapper[4556]: I0218 09:22:09.452636 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ebf502b-e4a2-4d25-8551-f1deed31b5c9-kube-api-access-lvvsz" (OuterVolumeSpecName: "kube-api-access-lvvsz") pod "3ebf502b-e4a2-4d25-8551-f1deed31b5c9" (UID: "3ebf502b-e4a2-4d25-8551-f1deed31b5c9"). InnerVolumeSpecName "kube-api-access-lvvsz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:22:09 crc kubenswrapper[4556]: I0218 09:22:09.463399 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ebf502b-e4a2-4d25-8551-f1deed31b5c9-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "3ebf502b-e4a2-4d25-8551-f1deed31b5c9" (UID: "3ebf502b-e4a2-4d25-8551-f1deed31b5c9"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:22:09 crc kubenswrapper[4556]: I0218 09:22:09.464458 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ebf502b-e4a2-4d25-8551-f1deed31b5c9-inventory" (OuterVolumeSpecName: "inventory") pod "3ebf502b-e4a2-4d25-8551-f1deed31b5c9" (UID: "3ebf502b-e4a2-4d25-8551-f1deed31b5c9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:22:09 crc kubenswrapper[4556]: I0218 09:22:09.545617 4556 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3ebf502b-e4a2-4d25-8551-f1deed31b5c9-inventory\") on node \"crc\" DevicePath \"\"" Feb 18 09:22:09 crc kubenswrapper[4556]: I0218 09:22:09.545648 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lvvsz\" (UniqueName: \"kubernetes.io/projected/3ebf502b-e4a2-4d25-8551-f1deed31b5c9-kube-api-access-lvvsz\") on node \"crc\" DevicePath \"\"" Feb 18 09:22:09 crc kubenswrapper[4556]: I0218 09:22:09.545660 4556 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3ebf502b-e4a2-4d25-8551-f1deed31b5c9-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 18 09:22:10 crc kubenswrapper[4556]: I0218 09:22:10.007048 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-hjswl" event={"ID":"3ebf502b-e4a2-4d25-8551-f1deed31b5c9","Type":"ContainerDied","Data":"d30854a28f8152b7436ba3d9671081b16709ee62110af8a1230be2fd054d6207"} Feb 18 09:22:10 crc kubenswrapper[4556]: I0218 09:22:10.007093 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d30854a28f8152b7436ba3d9671081b16709ee62110af8a1230be2fd054d6207" Feb 18 09:22:10 crc kubenswrapper[4556]: I0218 09:22:10.007101 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-hjswl" Feb 18 09:22:10 crc kubenswrapper[4556]: I0218 09:22:10.063246 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pqnqh"] Feb 18 09:22:10 crc kubenswrapper[4556]: E0218 09:22:10.063826 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ebf502b-e4a2-4d25-8551-f1deed31b5c9" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Feb 18 09:22:10 crc kubenswrapper[4556]: I0218 09:22:10.063846 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ebf502b-e4a2-4d25-8551-f1deed31b5c9" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Feb 18 09:22:10 crc kubenswrapper[4556]: I0218 09:22:10.064078 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ebf502b-e4a2-4d25-8551-f1deed31b5c9" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Feb 18 09:22:10 crc kubenswrapper[4556]: I0218 09:22:10.064909 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pqnqh" Feb 18 09:22:10 crc kubenswrapper[4556]: I0218 09:22:10.067397 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 18 09:22:10 crc kubenswrapper[4556]: I0218 09:22:10.067675 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 18 09:22:10 crc kubenswrapper[4556]: I0218 09:22:10.067797 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-54z7h" Feb 18 09:22:10 crc kubenswrapper[4556]: I0218 09:22:10.067852 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 18 09:22:10 crc kubenswrapper[4556]: I0218 09:22:10.074068 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pqnqh"] Feb 18 09:22:10 crc kubenswrapper[4556]: I0218 09:22:10.161354 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gb5q4\" (UniqueName: \"kubernetes.io/projected/91161557-e383-4bba-81d8-d3b0ba3b6840-kube-api-access-gb5q4\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-pqnqh\" (UID: \"91161557-e383-4bba-81d8-d3b0ba3b6840\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pqnqh" Feb 18 09:22:10 crc kubenswrapper[4556]: I0218 09:22:10.161590 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91161557-e383-4bba-81d8-d3b0ba3b6840-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-pqnqh\" (UID: \"91161557-e383-4bba-81d8-d3b0ba3b6840\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pqnqh" Feb 18 09:22:10 crc kubenswrapper[4556]: I0218 09:22:10.161633 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/91161557-e383-4bba-81d8-d3b0ba3b6840-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-pqnqh\" (UID: \"91161557-e383-4bba-81d8-d3b0ba3b6840\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pqnqh" Feb 18 09:22:10 crc kubenswrapper[4556]: I0218 09:22:10.161773 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/91161557-e383-4bba-81d8-d3b0ba3b6840-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-pqnqh\" (UID: \"91161557-e383-4bba-81d8-d3b0ba3b6840\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pqnqh" Feb 18 09:22:10 crc kubenswrapper[4556]: I0218 09:22:10.263834 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91161557-e383-4bba-81d8-d3b0ba3b6840-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-pqnqh\" (UID: \"91161557-e383-4bba-81d8-d3b0ba3b6840\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pqnqh" Feb 18 09:22:10 crc kubenswrapper[4556]: I0218 09:22:10.263884 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/91161557-e383-4bba-81d8-d3b0ba3b6840-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-pqnqh\" (UID: \"91161557-e383-4bba-81d8-d3b0ba3b6840\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pqnqh" Feb 18 09:22:10 crc kubenswrapper[4556]: I0218 09:22:10.263999 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/91161557-e383-4bba-81d8-d3b0ba3b6840-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-pqnqh\" (UID: \"91161557-e383-4bba-81d8-d3b0ba3b6840\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pqnqh" Feb 18 09:22:10 crc kubenswrapper[4556]: I0218 09:22:10.264239 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gb5q4\" (UniqueName: \"kubernetes.io/projected/91161557-e383-4bba-81d8-d3b0ba3b6840-kube-api-access-gb5q4\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-pqnqh\" (UID: \"91161557-e383-4bba-81d8-d3b0ba3b6840\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pqnqh" Feb 18 09:22:10 crc kubenswrapper[4556]: I0218 09:22:10.268771 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/91161557-e383-4bba-81d8-d3b0ba3b6840-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-pqnqh\" (UID: \"91161557-e383-4bba-81d8-d3b0ba3b6840\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pqnqh" Feb 18 09:22:10 crc kubenswrapper[4556]: I0218 09:22:10.268793 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/91161557-e383-4bba-81d8-d3b0ba3b6840-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-pqnqh\" (UID: \"91161557-e383-4bba-81d8-d3b0ba3b6840\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pqnqh" Feb 18 09:22:10 crc kubenswrapper[4556]: I0218 09:22:10.271398 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91161557-e383-4bba-81d8-d3b0ba3b6840-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-pqnqh\" (UID: \"91161557-e383-4bba-81d8-d3b0ba3b6840\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pqnqh" Feb 18 09:22:10 crc kubenswrapper[4556]: I0218 09:22:10.282816 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gb5q4\" (UniqueName: \"kubernetes.io/projected/91161557-e383-4bba-81d8-d3b0ba3b6840-kube-api-access-gb5q4\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-pqnqh\" (UID: \"91161557-e383-4bba-81d8-d3b0ba3b6840\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pqnqh" Feb 18 09:22:10 crc kubenswrapper[4556]: I0218 09:22:10.385527 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pqnqh" Feb 18 09:22:10 crc kubenswrapper[4556]: I0218 09:22:10.857146 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pqnqh"] Feb 18 09:22:10 crc kubenswrapper[4556]: W0218 09:22:10.858650 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod91161557_e383_4bba_81d8_d3b0ba3b6840.slice/crio-51390230281f0e8b584c04e369ee05a76366370558e464901f377468c6ca6227 WatchSource:0}: Error finding container 51390230281f0e8b584c04e369ee05a76366370558e464901f377468c6ca6227: Status 404 returned error can't find the container with id 51390230281f0e8b584c04e369ee05a76366370558e464901f377468c6ca6227 Feb 18 09:22:11 crc kubenswrapper[4556]: I0218 09:22:11.015904 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pqnqh" event={"ID":"91161557-e383-4bba-81d8-d3b0ba3b6840","Type":"ContainerStarted","Data":"51390230281f0e8b584c04e369ee05a76366370558e464901f377468c6ca6227"} Feb 18 09:22:12 crc kubenswrapper[4556]: I0218 09:22:12.027808 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pqnqh" event={"ID":"91161557-e383-4bba-81d8-d3b0ba3b6840","Type":"ContainerStarted","Data":"0585e289cf2002c120aa75eea3724029df6a95f46e301546a5687b3493e217a1"} Feb 18 09:22:12 crc kubenswrapper[4556]: I0218 09:22:12.048001 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pqnqh" podStartSLOduration=1.5407177189999999 podStartE2EDuration="2.047983634s" podCreationTimestamp="2026-02-18 09:22:10 +0000 UTC" firstStartedPulling="2026-02-18 09:22:10.860491402 +0000 UTC m=+1087.877452382" lastFinishedPulling="2026-02-18 09:22:11.367757317 +0000 UTC m=+1088.384718297" observedRunningTime="2026-02-18 09:22:12.042547569 +0000 UTC m=+1089.059508549" watchObservedRunningTime="2026-02-18 09:22:12.047983634 +0000 UTC m=+1089.064944615" Feb 18 09:22:14 crc kubenswrapper[4556]: I0218 09:22:14.230305 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Feb 18 09:22:16 crc kubenswrapper[4556]: I0218 09:22:16.039312 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Feb 18 09:22:31 crc kubenswrapper[4556]: I0218 09:22:31.727578 4556 patch_prober.go:28] interesting pod/machine-config-daemon-f76hs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 09:22:31 crc kubenswrapper[4556]: I0218 09:22:31.728540 4556 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 09:23:01 crc kubenswrapper[4556]: I0218 09:23:01.727299 4556 patch_prober.go:28] interesting pod/machine-config-daemon-f76hs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 09:23:01 crc kubenswrapper[4556]: I0218 09:23:01.727689 4556 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 09:23:01 crc kubenswrapper[4556]: I0218 09:23:01.727733 4556 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" Feb 18 09:23:01 crc kubenswrapper[4556]: I0218 09:23:01.728467 4556 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"16c417a1a589105242e4b889223fc0777a5d7312ffccba6c3cb31cf53da6f65a"} pod="openshift-machine-config-operator/machine-config-daemon-f76hs" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 18 09:23:01 crc kubenswrapper[4556]: I0218 09:23:01.728521 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerName="machine-config-daemon" containerID="cri-o://16c417a1a589105242e4b889223fc0777a5d7312ffccba6c3cb31cf53da6f65a" gracePeriod=600 Feb 18 09:23:02 crc kubenswrapper[4556]: I0218 09:23:02.426806 4556 generic.go:334] "Generic (PLEG): container finished" podID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerID="16c417a1a589105242e4b889223fc0777a5d7312ffccba6c3cb31cf53da6f65a" exitCode=0 Feb 18 09:23:02 crc kubenswrapper[4556]: I0218 09:23:02.426844 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" event={"ID":"8dac7f27-d3d1-4778-9e54-f273035a1d37","Type":"ContainerDied","Data":"16c417a1a589105242e4b889223fc0777a5d7312ffccba6c3cb31cf53da6f65a"} Feb 18 09:23:02 crc kubenswrapper[4556]: I0218 09:23:02.427059 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" event={"ID":"8dac7f27-d3d1-4778-9e54-f273035a1d37","Type":"ContainerStarted","Data":"06650a93598feaae08bc8fc051647e71397e8b9f4be0a24b1b089259c0bc5843"} Feb 18 09:23:02 crc kubenswrapper[4556]: I0218 09:23:02.427081 4556 scope.go:117] "RemoveContainer" containerID="7138060d5bb9e6802973b6e4fec551cb2729ba7d9b32c912776fd108e1b87ace" Feb 18 09:23:09 crc kubenswrapper[4556]: I0218 09:23:09.207925 4556 scope.go:117] "RemoveContainer" containerID="d1da7830e114e7c9d08657a35bdcbad92e1e4bf6c2ebaeedb0d76f08ad71fa0d" Feb 18 09:24:09 crc kubenswrapper[4556]: I0218 09:24:09.249952 4556 scope.go:117] "RemoveContainer" containerID="72fff07c09446faeccec54641807858690433ad99d4ec281b539e023cf897518" Feb 18 09:24:09 crc kubenswrapper[4556]: I0218 09:24:09.281369 4556 scope.go:117] "RemoveContainer" containerID="ca0d831b26e0653479e1280cd1625ab6207b9f101f1bbc574d2caad852610ea1" Feb 18 09:25:01 crc kubenswrapper[4556]: I0218 09:25:01.727307 4556 patch_prober.go:28] interesting pod/machine-config-daemon-f76hs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 09:25:01 crc kubenswrapper[4556]: I0218 09:25:01.727943 4556 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 09:25:22 crc kubenswrapper[4556]: I0218 09:25:22.469997 4556 generic.go:334] "Generic (PLEG): container finished" podID="91161557-e383-4bba-81d8-d3b0ba3b6840" containerID="0585e289cf2002c120aa75eea3724029df6a95f46e301546a5687b3493e217a1" exitCode=0 Feb 18 09:25:22 crc kubenswrapper[4556]: I0218 09:25:22.470059 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pqnqh" event={"ID":"91161557-e383-4bba-81d8-d3b0ba3b6840","Type":"ContainerDied","Data":"0585e289cf2002c120aa75eea3724029df6a95f46e301546a5687b3493e217a1"} Feb 18 09:25:23 crc kubenswrapper[4556]: I0218 09:25:23.802788 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pqnqh" Feb 18 09:25:23 crc kubenswrapper[4556]: I0218 09:25:23.923144 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/91161557-e383-4bba-81d8-d3b0ba3b6840-ssh-key-openstack-edpm-ipam\") pod \"91161557-e383-4bba-81d8-d3b0ba3b6840\" (UID: \"91161557-e383-4bba-81d8-d3b0ba3b6840\") " Feb 18 09:25:23 crc kubenswrapper[4556]: I0218 09:25:23.923580 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gb5q4\" (UniqueName: \"kubernetes.io/projected/91161557-e383-4bba-81d8-d3b0ba3b6840-kube-api-access-gb5q4\") pod \"91161557-e383-4bba-81d8-d3b0ba3b6840\" (UID: \"91161557-e383-4bba-81d8-d3b0ba3b6840\") " Feb 18 09:25:23 crc kubenswrapper[4556]: I0218 09:25:23.923699 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91161557-e383-4bba-81d8-d3b0ba3b6840-bootstrap-combined-ca-bundle\") pod \"91161557-e383-4bba-81d8-d3b0ba3b6840\" (UID: \"91161557-e383-4bba-81d8-d3b0ba3b6840\") " Feb 18 09:25:23 crc kubenswrapper[4556]: I0218 09:25:23.923777 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/91161557-e383-4bba-81d8-d3b0ba3b6840-inventory\") pod \"91161557-e383-4bba-81d8-d3b0ba3b6840\" (UID: \"91161557-e383-4bba-81d8-d3b0ba3b6840\") " Feb 18 09:25:23 crc kubenswrapper[4556]: I0218 09:25:23.929957 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91161557-e383-4bba-81d8-d3b0ba3b6840-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "91161557-e383-4bba-81d8-d3b0ba3b6840" (UID: "91161557-e383-4bba-81d8-d3b0ba3b6840"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:25:23 crc kubenswrapper[4556]: I0218 09:25:23.930335 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91161557-e383-4bba-81d8-d3b0ba3b6840-kube-api-access-gb5q4" (OuterVolumeSpecName: "kube-api-access-gb5q4") pod "91161557-e383-4bba-81d8-d3b0ba3b6840" (UID: "91161557-e383-4bba-81d8-d3b0ba3b6840"). InnerVolumeSpecName "kube-api-access-gb5q4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:25:23 crc kubenswrapper[4556]: I0218 09:25:23.946588 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91161557-e383-4bba-81d8-d3b0ba3b6840-inventory" (OuterVolumeSpecName: "inventory") pod "91161557-e383-4bba-81d8-d3b0ba3b6840" (UID: "91161557-e383-4bba-81d8-d3b0ba3b6840"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:25:23 crc kubenswrapper[4556]: I0218 09:25:23.946876 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91161557-e383-4bba-81d8-d3b0ba3b6840-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "91161557-e383-4bba-81d8-d3b0ba3b6840" (UID: "91161557-e383-4bba-81d8-d3b0ba3b6840"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:25:24 crc kubenswrapper[4556]: I0218 09:25:24.027280 4556 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91161557-e383-4bba-81d8-d3b0ba3b6840-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:25:24 crc kubenswrapper[4556]: I0218 09:25:24.027431 4556 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/91161557-e383-4bba-81d8-d3b0ba3b6840-inventory\") on node \"crc\" DevicePath \"\"" Feb 18 09:25:24 crc kubenswrapper[4556]: I0218 09:25:24.027491 4556 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/91161557-e383-4bba-81d8-d3b0ba3b6840-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 18 09:25:24 crc kubenswrapper[4556]: I0218 09:25:24.027541 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gb5q4\" (UniqueName: \"kubernetes.io/projected/91161557-e383-4bba-81d8-d3b0ba3b6840-kube-api-access-gb5q4\") on node \"crc\" DevicePath \"\"" Feb 18 09:25:24 crc kubenswrapper[4556]: I0218 09:25:24.492214 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pqnqh" event={"ID":"91161557-e383-4bba-81d8-d3b0ba3b6840","Type":"ContainerDied","Data":"51390230281f0e8b584c04e369ee05a76366370558e464901f377468c6ca6227"} Feb 18 09:25:24 crc kubenswrapper[4556]: I0218 09:25:24.492259 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="51390230281f0e8b584c04e369ee05a76366370558e464901f377468c6ca6227" Feb 18 09:25:24 crc kubenswrapper[4556]: I0218 09:25:24.492327 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pqnqh" Feb 18 09:25:24 crc kubenswrapper[4556]: I0218 09:25:24.557978 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qdtn4"] Feb 18 09:25:24 crc kubenswrapper[4556]: E0218 09:25:24.558558 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91161557-e383-4bba-81d8-d3b0ba3b6840" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Feb 18 09:25:24 crc kubenswrapper[4556]: I0218 09:25:24.558582 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="91161557-e383-4bba-81d8-d3b0ba3b6840" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Feb 18 09:25:24 crc kubenswrapper[4556]: I0218 09:25:24.572094 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="91161557-e383-4bba-81d8-d3b0ba3b6840" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Feb 18 09:25:24 crc kubenswrapper[4556]: I0218 09:25:24.574555 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qdtn4" Feb 18 09:25:24 crc kubenswrapper[4556]: I0218 09:25:24.582214 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qdtn4"] Feb 18 09:25:24 crc kubenswrapper[4556]: I0218 09:25:24.582474 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-54z7h" Feb 18 09:25:24 crc kubenswrapper[4556]: I0218 09:25:24.582558 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 18 09:25:24 crc kubenswrapper[4556]: I0218 09:25:24.582794 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 18 09:25:24 crc kubenswrapper[4556]: I0218 09:25:24.586174 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 18 09:25:24 crc kubenswrapper[4556]: I0218 09:25:24.640797 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a33e8704-dcbf-43ff-b665-cff323679c76-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-qdtn4\" (UID: \"a33e8704-dcbf-43ff-b665-cff323679c76\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qdtn4" Feb 18 09:25:24 crc kubenswrapper[4556]: I0218 09:25:24.640875 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a33e8704-dcbf-43ff-b665-cff323679c76-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-qdtn4\" (UID: \"a33e8704-dcbf-43ff-b665-cff323679c76\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qdtn4" Feb 18 09:25:24 crc kubenswrapper[4556]: I0218 09:25:24.640988 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlncv\" (UniqueName: \"kubernetes.io/projected/a33e8704-dcbf-43ff-b665-cff323679c76-kube-api-access-vlncv\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-qdtn4\" (UID: \"a33e8704-dcbf-43ff-b665-cff323679c76\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qdtn4" Feb 18 09:25:24 crc kubenswrapper[4556]: I0218 09:25:24.743757 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlncv\" (UniqueName: \"kubernetes.io/projected/a33e8704-dcbf-43ff-b665-cff323679c76-kube-api-access-vlncv\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-qdtn4\" (UID: \"a33e8704-dcbf-43ff-b665-cff323679c76\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qdtn4" Feb 18 09:25:24 crc kubenswrapper[4556]: I0218 09:25:24.744662 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a33e8704-dcbf-43ff-b665-cff323679c76-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-qdtn4\" (UID: \"a33e8704-dcbf-43ff-b665-cff323679c76\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qdtn4" Feb 18 09:25:24 crc kubenswrapper[4556]: I0218 09:25:24.744775 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a33e8704-dcbf-43ff-b665-cff323679c76-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-qdtn4\" (UID: \"a33e8704-dcbf-43ff-b665-cff323679c76\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qdtn4" Feb 18 09:25:24 crc kubenswrapper[4556]: I0218 09:25:24.748194 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a33e8704-dcbf-43ff-b665-cff323679c76-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-qdtn4\" (UID: \"a33e8704-dcbf-43ff-b665-cff323679c76\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qdtn4" Feb 18 09:25:24 crc kubenswrapper[4556]: I0218 09:25:24.749015 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a33e8704-dcbf-43ff-b665-cff323679c76-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-qdtn4\" (UID: \"a33e8704-dcbf-43ff-b665-cff323679c76\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qdtn4" Feb 18 09:25:24 crc kubenswrapper[4556]: I0218 09:25:24.760809 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlncv\" (UniqueName: \"kubernetes.io/projected/a33e8704-dcbf-43ff-b665-cff323679c76-kube-api-access-vlncv\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-qdtn4\" (UID: \"a33e8704-dcbf-43ff-b665-cff323679c76\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qdtn4" Feb 18 09:25:24 crc kubenswrapper[4556]: I0218 09:25:24.902110 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qdtn4" Feb 18 09:25:25 crc kubenswrapper[4556]: I0218 09:25:25.361250 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qdtn4"] Feb 18 09:25:25 crc kubenswrapper[4556]: I0218 09:25:25.500444 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qdtn4" event={"ID":"a33e8704-dcbf-43ff-b665-cff323679c76","Type":"ContainerStarted","Data":"ec753463be61da54de961f45c0f7bee412c1231e177a42d7aa9716339764a4ff"} Feb 18 09:25:26 crc kubenswrapper[4556]: I0218 09:25:26.511887 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qdtn4" event={"ID":"a33e8704-dcbf-43ff-b665-cff323679c76","Type":"ContainerStarted","Data":"e9bdfa2ef2b6702b9c43460e38dc81af4bdccacc1215ce9d270a389753f0dd93"} Feb 18 09:25:26 crc kubenswrapper[4556]: I0218 09:25:26.531377 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qdtn4" podStartSLOduration=2.003090935 podStartE2EDuration="2.531362356s" podCreationTimestamp="2026-02-18 09:25:24 +0000 UTC" firstStartedPulling="2026-02-18 09:25:25.366263558 +0000 UTC m=+1282.383224538" lastFinishedPulling="2026-02-18 09:25:25.894534979 +0000 UTC m=+1282.911495959" observedRunningTime="2026-02-18 09:25:26.526983696 +0000 UTC m=+1283.543944677" watchObservedRunningTime="2026-02-18 09:25:26.531362356 +0000 UTC m=+1283.548323337" Feb 18 09:25:31 crc kubenswrapper[4556]: I0218 09:25:31.728068 4556 patch_prober.go:28] interesting pod/machine-config-daemon-f76hs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 09:25:31 crc kubenswrapper[4556]: I0218 09:25:31.728684 4556 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 09:26:01 crc kubenswrapper[4556]: I0218 09:26:01.727115 4556 patch_prober.go:28] interesting pod/machine-config-daemon-f76hs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 09:26:01 crc kubenswrapper[4556]: I0218 09:26:01.727555 4556 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 09:26:01 crc kubenswrapper[4556]: I0218 09:26:01.727596 4556 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" Feb 18 09:26:01 crc kubenswrapper[4556]: I0218 09:26:01.728310 4556 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"06650a93598feaae08bc8fc051647e71397e8b9f4be0a24b1b089259c0bc5843"} pod="openshift-machine-config-operator/machine-config-daemon-f76hs" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 18 09:26:01 crc kubenswrapper[4556]: I0218 09:26:01.728357 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerName="machine-config-daemon" containerID="cri-o://06650a93598feaae08bc8fc051647e71397e8b9f4be0a24b1b089259c0bc5843" gracePeriod=600 Feb 18 09:26:02 crc kubenswrapper[4556]: I0218 09:26:02.794316 4556 generic.go:334] "Generic (PLEG): container finished" podID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerID="06650a93598feaae08bc8fc051647e71397e8b9f4be0a24b1b089259c0bc5843" exitCode=0 Feb 18 09:26:02 crc kubenswrapper[4556]: I0218 09:26:02.794379 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" event={"ID":"8dac7f27-d3d1-4778-9e54-f273035a1d37","Type":"ContainerDied","Data":"06650a93598feaae08bc8fc051647e71397e8b9f4be0a24b1b089259c0bc5843"} Feb 18 09:26:02 crc kubenswrapper[4556]: I0218 09:26:02.794702 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" event={"ID":"8dac7f27-d3d1-4778-9e54-f273035a1d37","Type":"ContainerStarted","Data":"507a9134b6862bc8be1ac03b7993c890ef64f16c21975e00aa96c53fe413415d"} Feb 18 09:26:02 crc kubenswrapper[4556]: I0218 09:26:02.794723 4556 scope.go:117] "RemoveContainer" containerID="16c417a1a589105242e4b889223fc0777a5d7312ffccba6c3cb31cf53da6f65a" Feb 18 09:26:09 crc kubenswrapper[4556]: I0218 09:26:09.351356 4556 scope.go:117] "RemoveContainer" containerID="5b0edfc3e404d52fc9c324525b661c589d816bf646b1ce5470138f85ad4b763a" Feb 18 09:26:09 crc kubenswrapper[4556]: I0218 09:26:09.367941 4556 scope.go:117] "RemoveContainer" containerID="1d20caf4c1ff142373c33a04b0efed7d80eb00bca35ccaa547d8f5984917dcef" Feb 18 09:26:09 crc kubenswrapper[4556]: I0218 09:26:09.382406 4556 scope.go:117] "RemoveContainer" containerID="f6a2777c15f011057e9eb148ba4d170bda4c52701b24e2f8ccd0d1db588afa4f" Feb 18 09:26:55 crc kubenswrapper[4556]: I0218 09:26:55.160986 4556 generic.go:334] "Generic (PLEG): container finished" podID="a33e8704-dcbf-43ff-b665-cff323679c76" containerID="e9bdfa2ef2b6702b9c43460e38dc81af4bdccacc1215ce9d270a389753f0dd93" exitCode=0 Feb 18 09:26:55 crc kubenswrapper[4556]: I0218 09:26:55.161054 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qdtn4" event={"ID":"a33e8704-dcbf-43ff-b665-cff323679c76","Type":"ContainerDied","Data":"e9bdfa2ef2b6702b9c43460e38dc81af4bdccacc1215ce9d270a389753f0dd93"} Feb 18 09:26:56 crc kubenswrapper[4556]: I0218 09:26:56.483270 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qdtn4" Feb 18 09:26:56 crc kubenswrapper[4556]: I0218 09:26:56.610104 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vlncv\" (UniqueName: \"kubernetes.io/projected/a33e8704-dcbf-43ff-b665-cff323679c76-kube-api-access-vlncv\") pod \"a33e8704-dcbf-43ff-b665-cff323679c76\" (UID: \"a33e8704-dcbf-43ff-b665-cff323679c76\") " Feb 18 09:26:56 crc kubenswrapper[4556]: I0218 09:26:56.610265 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a33e8704-dcbf-43ff-b665-cff323679c76-inventory\") pod \"a33e8704-dcbf-43ff-b665-cff323679c76\" (UID: \"a33e8704-dcbf-43ff-b665-cff323679c76\") " Feb 18 09:26:56 crc kubenswrapper[4556]: I0218 09:26:56.610290 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a33e8704-dcbf-43ff-b665-cff323679c76-ssh-key-openstack-edpm-ipam\") pod \"a33e8704-dcbf-43ff-b665-cff323679c76\" (UID: \"a33e8704-dcbf-43ff-b665-cff323679c76\") " Feb 18 09:26:56 crc kubenswrapper[4556]: I0218 09:26:56.616293 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a33e8704-dcbf-43ff-b665-cff323679c76-kube-api-access-vlncv" (OuterVolumeSpecName: "kube-api-access-vlncv") pod "a33e8704-dcbf-43ff-b665-cff323679c76" (UID: "a33e8704-dcbf-43ff-b665-cff323679c76"). InnerVolumeSpecName "kube-api-access-vlncv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:26:56 crc kubenswrapper[4556]: I0218 09:26:56.633044 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a33e8704-dcbf-43ff-b665-cff323679c76-inventory" (OuterVolumeSpecName: "inventory") pod "a33e8704-dcbf-43ff-b665-cff323679c76" (UID: "a33e8704-dcbf-43ff-b665-cff323679c76"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:26:56 crc kubenswrapper[4556]: I0218 09:26:56.633797 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a33e8704-dcbf-43ff-b665-cff323679c76-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "a33e8704-dcbf-43ff-b665-cff323679c76" (UID: "a33e8704-dcbf-43ff-b665-cff323679c76"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:26:56 crc kubenswrapper[4556]: I0218 09:26:56.712037 4556 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a33e8704-dcbf-43ff-b665-cff323679c76-inventory\") on node \"crc\" DevicePath \"\"" Feb 18 09:26:56 crc kubenswrapper[4556]: I0218 09:26:56.712066 4556 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a33e8704-dcbf-43ff-b665-cff323679c76-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 18 09:26:56 crc kubenswrapper[4556]: I0218 09:26:56.712078 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vlncv\" (UniqueName: \"kubernetes.io/projected/a33e8704-dcbf-43ff-b665-cff323679c76-kube-api-access-vlncv\") on node \"crc\" DevicePath \"\"" Feb 18 09:26:57 crc kubenswrapper[4556]: I0218 09:26:57.179784 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qdtn4" event={"ID":"a33e8704-dcbf-43ff-b665-cff323679c76","Type":"ContainerDied","Data":"ec753463be61da54de961f45c0f7bee412c1231e177a42d7aa9716339764a4ff"} Feb 18 09:26:57 crc kubenswrapper[4556]: I0218 09:26:57.179826 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ec753463be61da54de961f45c0f7bee412c1231e177a42d7aa9716339764a4ff" Feb 18 09:26:57 crc kubenswrapper[4556]: I0218 09:26:57.179828 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qdtn4" Feb 18 09:26:57 crc kubenswrapper[4556]: I0218 09:26:57.239126 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-snzk4"] Feb 18 09:26:57 crc kubenswrapper[4556]: E0218 09:26:57.239844 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a33e8704-dcbf-43ff-b665-cff323679c76" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Feb 18 09:26:57 crc kubenswrapper[4556]: I0218 09:26:57.239945 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="a33e8704-dcbf-43ff-b665-cff323679c76" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Feb 18 09:26:57 crc kubenswrapper[4556]: I0218 09:26:57.240227 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="a33e8704-dcbf-43ff-b665-cff323679c76" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Feb 18 09:26:57 crc kubenswrapper[4556]: I0218 09:26:57.240990 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-snzk4" Feb 18 09:26:57 crc kubenswrapper[4556]: I0218 09:26:57.243040 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-54z7h" Feb 18 09:26:57 crc kubenswrapper[4556]: I0218 09:26:57.243219 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 18 09:26:57 crc kubenswrapper[4556]: I0218 09:26:57.243402 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 18 09:26:57 crc kubenswrapper[4556]: I0218 09:26:57.243544 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 18 09:26:57 crc kubenswrapper[4556]: I0218 09:26:57.246677 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-snzk4"] Feb 18 09:26:57 crc kubenswrapper[4556]: I0218 09:26:57.320247 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4fdd7d3d-c538-4cb6-9d45-b94723b744b1-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-snzk4\" (UID: \"4fdd7d3d-c538-4cb6-9d45-b94723b744b1\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-snzk4" Feb 18 09:26:57 crc kubenswrapper[4556]: I0218 09:26:57.320299 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghgr2\" (UniqueName: \"kubernetes.io/projected/4fdd7d3d-c538-4cb6-9d45-b94723b744b1-kube-api-access-ghgr2\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-snzk4\" (UID: \"4fdd7d3d-c538-4cb6-9d45-b94723b744b1\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-snzk4" Feb 18 09:26:57 crc kubenswrapper[4556]: I0218 09:26:57.320395 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4fdd7d3d-c538-4cb6-9d45-b94723b744b1-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-snzk4\" (UID: \"4fdd7d3d-c538-4cb6-9d45-b94723b744b1\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-snzk4" Feb 18 09:26:57 crc kubenswrapper[4556]: I0218 09:26:57.421514 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4fdd7d3d-c538-4cb6-9d45-b94723b744b1-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-snzk4\" (UID: \"4fdd7d3d-c538-4cb6-9d45-b94723b744b1\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-snzk4" Feb 18 09:26:57 crc kubenswrapper[4556]: I0218 09:26:57.421668 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghgr2\" (UniqueName: \"kubernetes.io/projected/4fdd7d3d-c538-4cb6-9d45-b94723b744b1-kube-api-access-ghgr2\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-snzk4\" (UID: \"4fdd7d3d-c538-4cb6-9d45-b94723b744b1\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-snzk4" Feb 18 09:26:57 crc kubenswrapper[4556]: I0218 09:26:57.421830 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4fdd7d3d-c538-4cb6-9d45-b94723b744b1-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-snzk4\" (UID: \"4fdd7d3d-c538-4cb6-9d45-b94723b744b1\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-snzk4" Feb 18 09:26:57 crc kubenswrapper[4556]: I0218 09:26:57.425629 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4fdd7d3d-c538-4cb6-9d45-b94723b744b1-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-snzk4\" (UID: \"4fdd7d3d-c538-4cb6-9d45-b94723b744b1\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-snzk4" Feb 18 09:26:57 crc kubenswrapper[4556]: I0218 09:26:57.427432 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4fdd7d3d-c538-4cb6-9d45-b94723b744b1-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-snzk4\" (UID: \"4fdd7d3d-c538-4cb6-9d45-b94723b744b1\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-snzk4" Feb 18 09:26:57 crc kubenswrapper[4556]: I0218 09:26:57.436784 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghgr2\" (UniqueName: \"kubernetes.io/projected/4fdd7d3d-c538-4cb6-9d45-b94723b744b1-kube-api-access-ghgr2\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-snzk4\" (UID: \"4fdd7d3d-c538-4cb6-9d45-b94723b744b1\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-snzk4" Feb 18 09:26:57 crc kubenswrapper[4556]: I0218 09:26:57.554996 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-snzk4" Feb 18 09:26:57 crc kubenswrapper[4556]: I0218 09:26:57.990166 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-snzk4"] Feb 18 09:26:58 crc kubenswrapper[4556]: I0218 09:26:58.188432 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-snzk4" event={"ID":"4fdd7d3d-c538-4cb6-9d45-b94723b744b1","Type":"ContainerStarted","Data":"d1cd92269a3d2373156472928607e26cca6ed13d0011522b2d85c18fb27481b8"} Feb 18 09:26:59 crc kubenswrapper[4556]: I0218 09:26:59.200772 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-snzk4" event={"ID":"4fdd7d3d-c538-4cb6-9d45-b94723b744b1","Type":"ContainerStarted","Data":"958d448d3bb8ff034d9fb39539e0ad02c0cf9c18a1318f26353fab49e435e0dd"} Feb 18 09:26:59 crc kubenswrapper[4556]: I0218 09:26:59.213782 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-snzk4" podStartSLOduration=1.589158125 podStartE2EDuration="2.213762252s" podCreationTimestamp="2026-02-18 09:26:57 +0000 UTC" firstStartedPulling="2026-02-18 09:26:57.997431246 +0000 UTC m=+1375.014392227" lastFinishedPulling="2026-02-18 09:26:58.622035374 +0000 UTC m=+1375.638996354" observedRunningTime="2026-02-18 09:26:59.212518625 +0000 UTC m=+1376.229479606" watchObservedRunningTime="2026-02-18 09:26:59.213762252 +0000 UTC m=+1376.230723232" Feb 18 09:27:00 crc kubenswrapper[4556]: I0218 09:27:00.036454 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-fqkrx"] Feb 18 09:27:00 crc kubenswrapper[4556]: I0218 09:27:00.045908 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-101a-account-create-update-t5fnf"] Feb 18 09:27:00 crc kubenswrapper[4556]: I0218 09:27:00.055127 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-de68-account-create-update-f26q6"] Feb 18 09:27:00 crc kubenswrapper[4556]: I0218 09:27:00.062537 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-fqkrx"] Feb 18 09:27:00 crc kubenswrapper[4556]: I0218 09:27:00.069560 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-101a-account-create-update-t5fnf"] Feb 18 09:27:00 crc kubenswrapper[4556]: I0218 09:27:00.077578 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-de68-account-create-update-f26q6"] Feb 18 09:27:01 crc kubenswrapper[4556]: I0218 09:27:01.024819 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-be94-account-create-update-rqlh4"] Feb 18 09:27:01 crc kubenswrapper[4556]: I0218 09:27:01.031559 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-94t8h"] Feb 18 09:27:01 crc kubenswrapper[4556]: I0218 09:27:01.041500 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-8n5wp"] Feb 18 09:27:01 crc kubenswrapper[4556]: I0218 09:27:01.050233 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-be94-account-create-update-rqlh4"] Feb 18 09:27:01 crc kubenswrapper[4556]: I0218 09:27:01.057125 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-94t8h"] Feb 18 09:27:01 crc kubenswrapper[4556]: I0218 09:27:01.062457 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-8n5wp"] Feb 18 09:27:01 crc kubenswrapper[4556]: I0218 09:27:01.290487 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03d46588-9781-48c2-9ce0-a56681c223c9" path="/var/lib/kubelet/pods/03d46588-9781-48c2-9ce0-a56681c223c9/volumes" Feb 18 09:27:01 crc kubenswrapper[4556]: I0218 09:27:01.291120 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f905f0b-32d9-40bc-8c19-9cfb9e4f739b" path="/var/lib/kubelet/pods/0f905f0b-32d9-40bc-8c19-9cfb9e4f739b/volumes" Feb 18 09:27:01 crc kubenswrapper[4556]: I0218 09:27:01.291676 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45a79e16-1b19-4d21-9edb-ff35eb78d357" path="/var/lib/kubelet/pods/45a79e16-1b19-4d21-9edb-ff35eb78d357/volumes" Feb 18 09:27:01 crc kubenswrapper[4556]: I0218 09:27:01.292215 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f7157d7-e589-4d41-955e-73ad86618b83" path="/var/lib/kubelet/pods/8f7157d7-e589-4d41-955e-73ad86618b83/volumes" Feb 18 09:27:01 crc kubenswrapper[4556]: I0218 09:27:01.293192 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b43cd61b-beae-4ef8-8648-d014d045e56c" path="/var/lib/kubelet/pods/b43cd61b-beae-4ef8-8648-d014d045e56c/volumes" Feb 18 09:27:01 crc kubenswrapper[4556]: I0218 09:27:01.293690 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce838404-6d81-41ca-bdbf-e9af260a0d55" path="/var/lib/kubelet/pods/ce838404-6d81-41ca-bdbf-e9af260a0d55/volumes" Feb 18 09:27:06 crc kubenswrapper[4556]: I0218 09:27:06.021944 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-blbjd"] Feb 18 09:27:06 crc kubenswrapper[4556]: I0218 09:27:06.028628 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-blbjd"] Feb 18 09:27:07 crc kubenswrapper[4556]: I0218 09:27:07.292285 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a185e40-8a02-44a1-ab00-2575bebde1d5" path="/var/lib/kubelet/pods/7a185e40-8a02-44a1-ab00-2575bebde1d5/volumes" Feb 18 09:27:09 crc kubenswrapper[4556]: I0218 09:27:09.436022 4556 scope.go:117] "RemoveContainer" containerID="b40a41ffea6945c4c57ebb77f95f33e5dea3a98e76eb30b2a22fcdaf4d9d3cbf" Feb 18 09:27:09 crc kubenswrapper[4556]: I0218 09:27:09.456622 4556 scope.go:117] "RemoveContainer" containerID="3933b7ed7b0d41f7f8756aeadc1b735fa65bb0952d7a007f2153c06e853c694b" Feb 18 09:27:09 crc kubenswrapper[4556]: I0218 09:27:09.493436 4556 scope.go:117] "RemoveContainer" containerID="acb3bfeafbc1f1a549401dc430a37eb857f13ba1b502240d540b8a9b1f9225f4" Feb 18 09:27:09 crc kubenswrapper[4556]: I0218 09:27:09.527052 4556 scope.go:117] "RemoveContainer" containerID="b6e2dd2b3609e35ad44eba00703cf72f2e7c5e572459442f7fd721df57ee204e" Feb 18 09:27:09 crc kubenswrapper[4556]: I0218 09:27:09.558733 4556 scope.go:117] "RemoveContainer" containerID="513593e18b400b5a6800acf343dac7b309280f2ceafe96d0317252c54847bc60" Feb 18 09:27:09 crc kubenswrapper[4556]: I0218 09:27:09.591195 4556 scope.go:117] "RemoveContainer" containerID="61012520598e8019f9b5f64bf562fa2834f8935a374317944e30a9c6c3ca3a3c" Feb 18 09:27:09 crc kubenswrapper[4556]: I0218 09:27:09.616890 4556 scope.go:117] "RemoveContainer" containerID="e053ef5cf1cf63f13f45392300ae28da28077133bd12a872132c875a0fc3bab0" Feb 18 09:27:23 crc kubenswrapper[4556]: I0218 09:27:23.028992 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-fxcc2"] Feb 18 09:27:23 crc kubenswrapper[4556]: I0218 09:27:23.037616 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-fxcc2"] Feb 18 09:27:23 crc kubenswrapper[4556]: I0218 09:27:23.289951 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4731e0af-7aff-41c9-bc33-ad533ee750ce" path="/var/lib/kubelet/pods/4731e0af-7aff-41c9-bc33-ad533ee750ce/volumes" Feb 18 09:27:36 crc kubenswrapper[4556]: I0218 09:27:36.023826 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-v7c5w"] Feb 18 09:27:36 crc kubenswrapper[4556]: I0218 09:27:36.032162 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-v7c5w"] Feb 18 09:27:37 crc kubenswrapper[4556]: I0218 09:27:37.023726 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-2d25r"] Feb 18 09:27:37 crc kubenswrapper[4556]: I0218 09:27:37.030699 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-2d25r"] Feb 18 09:27:37 crc kubenswrapper[4556]: I0218 09:27:37.291915 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f46d1fa-3786-4dde-ba6f-d6775782faeb" path="/var/lib/kubelet/pods/9f46d1fa-3786-4dde-ba6f-d6775782faeb/volumes" Feb 18 09:27:37 crc kubenswrapper[4556]: I0218 09:27:37.292633 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a037e8b8-1769-4728-a30c-dcb8671d2e1b" path="/var/lib/kubelet/pods/a037e8b8-1769-4728-a30c-dcb8671d2e1b/volumes" Feb 18 09:27:40 crc kubenswrapper[4556]: I0218 09:27:40.029419 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-fbv4q"] Feb 18 09:27:40 crc kubenswrapper[4556]: I0218 09:27:40.036728 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-ecdb-account-create-update-52wl7"] Feb 18 09:27:40 crc kubenswrapper[4556]: I0218 09:27:40.043805 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-fbv4q"] Feb 18 09:27:40 crc kubenswrapper[4556]: I0218 09:27:40.051209 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-7946-account-create-update-7t9xc"] Feb 18 09:27:40 crc kubenswrapper[4556]: I0218 09:27:40.063009 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-8f0c-account-create-update-l4dkr"] Feb 18 09:27:40 crc kubenswrapper[4556]: I0218 09:27:40.071556 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-ecdb-account-create-update-52wl7"] Feb 18 09:27:40 crc kubenswrapper[4556]: I0218 09:27:40.078393 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-7946-account-create-update-7t9xc"] Feb 18 09:27:40 crc kubenswrapper[4556]: I0218 09:27:40.084862 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-8f0c-account-create-update-l4dkr"] Feb 18 09:27:41 crc kubenswrapper[4556]: I0218 09:27:41.291611 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2856804d-f222-49c2-abc3-153227545c0a" path="/var/lib/kubelet/pods/2856804d-f222-49c2-abc3-153227545c0a/volumes" Feb 18 09:27:41 crc kubenswrapper[4556]: I0218 09:27:41.292515 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45043340-ef17-4b6c-a6e0-a87eac1942ec" path="/var/lib/kubelet/pods/45043340-ef17-4b6c-a6e0-a87eac1942ec/volumes" Feb 18 09:27:41 crc kubenswrapper[4556]: I0218 09:27:41.293285 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6bb6b9c-b2a6-4908-8da1-c519224d8ef2" path="/var/lib/kubelet/pods/a6bb6b9c-b2a6-4908-8da1-c519224d8ef2/volumes" Feb 18 09:27:41 crc kubenswrapper[4556]: I0218 09:27:41.294045 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1b45040-901f-4eab-a55d-46a9842686a8" path="/var/lib/kubelet/pods/f1b45040-901f-4eab-a55d-46a9842686a8/volumes" Feb 18 09:27:43 crc kubenswrapper[4556]: I0218 09:27:43.022533 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-hzc26"] Feb 18 09:27:43 crc kubenswrapper[4556]: I0218 09:27:43.030053 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-hzc26"] Feb 18 09:27:43 crc kubenswrapper[4556]: I0218 09:27:43.290756 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43bc6089-0cfa-478c-8edd-77e667cbb137" path="/var/lib/kubelet/pods/43bc6089-0cfa-478c-8edd-77e667cbb137/volumes" Feb 18 09:27:56 crc kubenswrapper[4556]: I0218 09:27:56.627964 4556 generic.go:334] "Generic (PLEG): container finished" podID="4fdd7d3d-c538-4cb6-9d45-b94723b744b1" containerID="958d448d3bb8ff034d9fb39539e0ad02c0cf9c18a1318f26353fab49e435e0dd" exitCode=0 Feb 18 09:27:56 crc kubenswrapper[4556]: I0218 09:27:56.628041 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-snzk4" event={"ID":"4fdd7d3d-c538-4cb6-9d45-b94723b744b1","Type":"ContainerDied","Data":"958d448d3bb8ff034d9fb39539e0ad02c0cf9c18a1318f26353fab49e435e0dd"} Feb 18 09:27:57 crc kubenswrapper[4556]: I0218 09:27:57.947123 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-snzk4" Feb 18 09:27:57 crc kubenswrapper[4556]: I0218 09:27:57.966611 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ghgr2\" (UniqueName: \"kubernetes.io/projected/4fdd7d3d-c538-4cb6-9d45-b94723b744b1-kube-api-access-ghgr2\") pod \"4fdd7d3d-c538-4cb6-9d45-b94723b744b1\" (UID: \"4fdd7d3d-c538-4cb6-9d45-b94723b744b1\") " Feb 18 09:27:57 crc kubenswrapper[4556]: I0218 09:27:57.966813 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4fdd7d3d-c538-4cb6-9d45-b94723b744b1-inventory\") pod \"4fdd7d3d-c538-4cb6-9d45-b94723b744b1\" (UID: \"4fdd7d3d-c538-4cb6-9d45-b94723b744b1\") " Feb 18 09:27:57 crc kubenswrapper[4556]: I0218 09:27:57.966880 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4fdd7d3d-c538-4cb6-9d45-b94723b744b1-ssh-key-openstack-edpm-ipam\") pod \"4fdd7d3d-c538-4cb6-9d45-b94723b744b1\" (UID: \"4fdd7d3d-c538-4cb6-9d45-b94723b744b1\") " Feb 18 09:27:57 crc kubenswrapper[4556]: I0218 09:27:57.971798 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4fdd7d3d-c538-4cb6-9d45-b94723b744b1-kube-api-access-ghgr2" (OuterVolumeSpecName: "kube-api-access-ghgr2") pod "4fdd7d3d-c538-4cb6-9d45-b94723b744b1" (UID: "4fdd7d3d-c538-4cb6-9d45-b94723b744b1"). InnerVolumeSpecName "kube-api-access-ghgr2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:27:57 crc kubenswrapper[4556]: I0218 09:27:57.989074 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4fdd7d3d-c538-4cb6-9d45-b94723b744b1-inventory" (OuterVolumeSpecName: "inventory") pod "4fdd7d3d-c538-4cb6-9d45-b94723b744b1" (UID: "4fdd7d3d-c538-4cb6-9d45-b94723b744b1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:27:57 crc kubenswrapper[4556]: I0218 09:27:57.991354 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4fdd7d3d-c538-4cb6-9d45-b94723b744b1-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "4fdd7d3d-c538-4cb6-9d45-b94723b744b1" (UID: "4fdd7d3d-c538-4cb6-9d45-b94723b744b1"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:27:58 crc kubenswrapper[4556]: I0218 09:27:58.069067 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ghgr2\" (UniqueName: \"kubernetes.io/projected/4fdd7d3d-c538-4cb6-9d45-b94723b744b1-kube-api-access-ghgr2\") on node \"crc\" DevicePath \"\"" Feb 18 09:27:58 crc kubenswrapper[4556]: I0218 09:27:58.069097 4556 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4fdd7d3d-c538-4cb6-9d45-b94723b744b1-inventory\") on node \"crc\" DevicePath \"\"" Feb 18 09:27:58 crc kubenswrapper[4556]: I0218 09:27:58.069107 4556 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4fdd7d3d-c538-4cb6-9d45-b94723b744b1-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 18 09:27:58 crc kubenswrapper[4556]: I0218 09:27:58.646097 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-snzk4" event={"ID":"4fdd7d3d-c538-4cb6-9d45-b94723b744b1","Type":"ContainerDied","Data":"d1cd92269a3d2373156472928607e26cca6ed13d0011522b2d85c18fb27481b8"} Feb 18 09:27:58 crc kubenswrapper[4556]: I0218 09:27:58.646413 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d1cd92269a3d2373156472928607e26cca6ed13d0011522b2d85c18fb27481b8" Feb 18 09:27:58 crc kubenswrapper[4556]: I0218 09:27:58.646221 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-snzk4" Feb 18 09:27:58 crc kubenswrapper[4556]: I0218 09:27:58.707704 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6jjjt"] Feb 18 09:27:58 crc kubenswrapper[4556]: E0218 09:27:58.708032 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fdd7d3d-c538-4cb6-9d45-b94723b744b1" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Feb 18 09:27:58 crc kubenswrapper[4556]: I0218 09:27:58.708050 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fdd7d3d-c538-4cb6-9d45-b94723b744b1" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Feb 18 09:27:58 crc kubenswrapper[4556]: I0218 09:27:58.708277 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="4fdd7d3d-c538-4cb6-9d45-b94723b744b1" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Feb 18 09:27:58 crc kubenswrapper[4556]: I0218 09:27:58.708804 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6jjjt" Feb 18 09:27:58 crc kubenswrapper[4556]: I0218 09:27:58.712991 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 18 09:27:58 crc kubenswrapper[4556]: I0218 09:27:58.713226 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 18 09:27:58 crc kubenswrapper[4556]: I0218 09:27:58.713363 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-54z7h" Feb 18 09:27:58 crc kubenswrapper[4556]: I0218 09:27:58.716909 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 18 09:27:58 crc kubenswrapper[4556]: I0218 09:27:58.719823 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6jjjt"] Feb 18 09:27:58 crc kubenswrapper[4556]: I0218 09:27:58.782377 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a5a47273-47da-41d2-a98f-dc1fba7e1102-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6jjjt\" (UID: \"a5a47273-47da-41d2-a98f-dc1fba7e1102\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6jjjt" Feb 18 09:27:58 crc kubenswrapper[4556]: I0218 09:27:58.782531 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gjks4\" (UniqueName: \"kubernetes.io/projected/a5a47273-47da-41d2-a98f-dc1fba7e1102-kube-api-access-gjks4\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6jjjt\" (UID: \"a5a47273-47da-41d2-a98f-dc1fba7e1102\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6jjjt" Feb 18 09:27:58 crc kubenswrapper[4556]: I0218 09:27:58.782800 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a5a47273-47da-41d2-a98f-dc1fba7e1102-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6jjjt\" (UID: \"a5a47273-47da-41d2-a98f-dc1fba7e1102\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6jjjt" Feb 18 09:27:58 crc kubenswrapper[4556]: I0218 09:27:58.884196 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a5a47273-47da-41d2-a98f-dc1fba7e1102-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6jjjt\" (UID: \"a5a47273-47da-41d2-a98f-dc1fba7e1102\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6jjjt" Feb 18 09:27:58 crc kubenswrapper[4556]: I0218 09:27:58.884339 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a5a47273-47da-41d2-a98f-dc1fba7e1102-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6jjjt\" (UID: \"a5a47273-47da-41d2-a98f-dc1fba7e1102\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6jjjt" Feb 18 09:27:58 crc kubenswrapper[4556]: I0218 09:27:58.884428 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gjks4\" (UniqueName: \"kubernetes.io/projected/a5a47273-47da-41d2-a98f-dc1fba7e1102-kube-api-access-gjks4\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6jjjt\" (UID: \"a5a47273-47da-41d2-a98f-dc1fba7e1102\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6jjjt" Feb 18 09:27:58 crc kubenswrapper[4556]: I0218 09:27:58.890217 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a5a47273-47da-41d2-a98f-dc1fba7e1102-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6jjjt\" (UID: \"a5a47273-47da-41d2-a98f-dc1fba7e1102\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6jjjt" Feb 18 09:27:58 crc kubenswrapper[4556]: I0218 09:27:58.891764 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a5a47273-47da-41d2-a98f-dc1fba7e1102-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6jjjt\" (UID: \"a5a47273-47da-41d2-a98f-dc1fba7e1102\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6jjjt" Feb 18 09:27:58 crc kubenswrapper[4556]: I0218 09:27:58.900767 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gjks4\" (UniqueName: \"kubernetes.io/projected/a5a47273-47da-41d2-a98f-dc1fba7e1102-kube-api-access-gjks4\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6jjjt\" (UID: \"a5a47273-47da-41d2-a98f-dc1fba7e1102\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6jjjt" Feb 18 09:27:59 crc kubenswrapper[4556]: I0218 09:27:59.025354 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6jjjt" Feb 18 09:27:59 crc kubenswrapper[4556]: I0218 09:27:59.473167 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6jjjt"] Feb 18 09:27:59 crc kubenswrapper[4556]: I0218 09:27:59.475359 4556 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 18 09:27:59 crc kubenswrapper[4556]: I0218 09:27:59.655248 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6jjjt" event={"ID":"a5a47273-47da-41d2-a98f-dc1fba7e1102","Type":"ContainerStarted","Data":"e45b4dba7476540f5eedc96f4e98ce9a7cb98631a048178c4b5731d873c2d104"} Feb 18 09:28:00 crc kubenswrapper[4556]: I0218 09:28:00.027257 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-4k7vg"] Feb 18 09:28:00 crc kubenswrapper[4556]: I0218 09:28:00.037103 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-4k7vg"] Feb 18 09:28:00 crc kubenswrapper[4556]: I0218 09:28:00.663993 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6jjjt" event={"ID":"a5a47273-47da-41d2-a98f-dc1fba7e1102","Type":"ContainerStarted","Data":"739ddfc762d0fdea7e932d8c8ed872e3ef2feb599a1363aa12b6f914bca2e296"} Feb 18 09:28:00 crc kubenswrapper[4556]: I0218 09:28:00.681456 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6jjjt" podStartSLOduration=2.063399666 podStartE2EDuration="2.681431463s" podCreationTimestamp="2026-02-18 09:27:58 +0000 UTC" firstStartedPulling="2026-02-18 09:27:59.474981823 +0000 UTC m=+1436.491942803" lastFinishedPulling="2026-02-18 09:28:00.09301362 +0000 UTC m=+1437.109974600" observedRunningTime="2026-02-18 09:28:00.67980183 +0000 UTC m=+1437.696762810" watchObservedRunningTime="2026-02-18 09:28:00.681431463 +0000 UTC m=+1437.698392444" Feb 18 09:28:01 crc kubenswrapper[4556]: I0218 09:28:01.291726 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6bf90aa4-605f-4b7d-abde-2a1f3236a6af" path="/var/lib/kubelet/pods/6bf90aa4-605f-4b7d-abde-2a1f3236a6af/volumes" Feb 18 09:28:01 crc kubenswrapper[4556]: I0218 09:28:01.727144 4556 patch_prober.go:28] interesting pod/machine-config-daemon-f76hs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 09:28:01 crc kubenswrapper[4556]: I0218 09:28:01.727702 4556 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 09:28:03 crc kubenswrapper[4556]: I0218 09:28:03.690195 4556 generic.go:334] "Generic (PLEG): container finished" podID="a5a47273-47da-41d2-a98f-dc1fba7e1102" containerID="739ddfc762d0fdea7e932d8c8ed872e3ef2feb599a1363aa12b6f914bca2e296" exitCode=0 Feb 18 09:28:03 crc kubenswrapper[4556]: I0218 09:28:03.690303 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6jjjt" event={"ID":"a5a47273-47da-41d2-a98f-dc1fba7e1102","Type":"ContainerDied","Data":"739ddfc762d0fdea7e932d8c8ed872e3ef2feb599a1363aa12b6f914bca2e296"} Feb 18 09:28:05 crc kubenswrapper[4556]: I0218 09:28:05.017195 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6jjjt" Feb 18 09:28:05 crc kubenswrapper[4556]: I0218 09:28:05.105696 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a5a47273-47da-41d2-a98f-dc1fba7e1102-ssh-key-openstack-edpm-ipam\") pod \"a5a47273-47da-41d2-a98f-dc1fba7e1102\" (UID: \"a5a47273-47da-41d2-a98f-dc1fba7e1102\") " Feb 18 09:28:05 crc kubenswrapper[4556]: I0218 09:28:05.105853 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a5a47273-47da-41d2-a98f-dc1fba7e1102-inventory\") pod \"a5a47273-47da-41d2-a98f-dc1fba7e1102\" (UID: \"a5a47273-47da-41d2-a98f-dc1fba7e1102\") " Feb 18 09:28:05 crc kubenswrapper[4556]: I0218 09:28:05.105914 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gjks4\" (UniqueName: \"kubernetes.io/projected/a5a47273-47da-41d2-a98f-dc1fba7e1102-kube-api-access-gjks4\") pod \"a5a47273-47da-41d2-a98f-dc1fba7e1102\" (UID: \"a5a47273-47da-41d2-a98f-dc1fba7e1102\") " Feb 18 09:28:05 crc kubenswrapper[4556]: I0218 09:28:05.112045 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5a47273-47da-41d2-a98f-dc1fba7e1102-kube-api-access-gjks4" (OuterVolumeSpecName: "kube-api-access-gjks4") pod "a5a47273-47da-41d2-a98f-dc1fba7e1102" (UID: "a5a47273-47da-41d2-a98f-dc1fba7e1102"). InnerVolumeSpecName "kube-api-access-gjks4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:28:05 crc kubenswrapper[4556]: I0218 09:28:05.129106 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5a47273-47da-41d2-a98f-dc1fba7e1102-inventory" (OuterVolumeSpecName: "inventory") pod "a5a47273-47da-41d2-a98f-dc1fba7e1102" (UID: "a5a47273-47da-41d2-a98f-dc1fba7e1102"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:28:05 crc kubenswrapper[4556]: I0218 09:28:05.129184 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5a47273-47da-41d2-a98f-dc1fba7e1102-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "a5a47273-47da-41d2-a98f-dc1fba7e1102" (UID: "a5a47273-47da-41d2-a98f-dc1fba7e1102"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:28:05 crc kubenswrapper[4556]: I0218 09:28:05.209011 4556 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a5a47273-47da-41d2-a98f-dc1fba7e1102-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 18 09:28:05 crc kubenswrapper[4556]: I0218 09:28:05.209060 4556 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a5a47273-47da-41d2-a98f-dc1fba7e1102-inventory\") on node \"crc\" DevicePath \"\"" Feb 18 09:28:05 crc kubenswrapper[4556]: I0218 09:28:05.209072 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gjks4\" (UniqueName: \"kubernetes.io/projected/a5a47273-47da-41d2-a98f-dc1fba7e1102-kube-api-access-gjks4\") on node \"crc\" DevicePath \"\"" Feb 18 09:28:05 crc kubenswrapper[4556]: I0218 09:28:05.704841 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6jjjt" event={"ID":"a5a47273-47da-41d2-a98f-dc1fba7e1102","Type":"ContainerDied","Data":"e45b4dba7476540f5eedc96f4e98ce9a7cb98631a048178c4b5731d873c2d104"} Feb 18 09:28:05 crc kubenswrapper[4556]: I0218 09:28:05.704882 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6jjjt" Feb 18 09:28:05 crc kubenswrapper[4556]: I0218 09:28:05.704884 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e45b4dba7476540f5eedc96f4e98ce9a7cb98631a048178c4b5731d873c2d104" Feb 18 09:28:05 crc kubenswrapper[4556]: I0218 09:28:05.756506 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-zmv42"] Feb 18 09:28:05 crc kubenswrapper[4556]: E0218 09:28:05.756965 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5a47273-47da-41d2-a98f-dc1fba7e1102" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Feb 18 09:28:05 crc kubenswrapper[4556]: I0218 09:28:05.756986 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5a47273-47da-41d2-a98f-dc1fba7e1102" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Feb 18 09:28:05 crc kubenswrapper[4556]: I0218 09:28:05.757188 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5a47273-47da-41d2-a98f-dc1fba7e1102" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Feb 18 09:28:05 crc kubenswrapper[4556]: I0218 09:28:05.757837 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zmv42" Feb 18 09:28:05 crc kubenswrapper[4556]: I0218 09:28:05.759325 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 18 09:28:05 crc kubenswrapper[4556]: I0218 09:28:05.759443 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 18 09:28:05 crc kubenswrapper[4556]: I0218 09:28:05.759546 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-54z7h" Feb 18 09:28:05 crc kubenswrapper[4556]: I0218 09:28:05.759627 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 18 09:28:05 crc kubenswrapper[4556]: I0218 09:28:05.763398 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-zmv42"] Feb 18 09:28:05 crc kubenswrapper[4556]: I0218 09:28:05.819996 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xhlp\" (UniqueName: \"kubernetes.io/projected/d1d95573-7bea-4ee6-b0c1-13ec18b10244-kube-api-access-4xhlp\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-zmv42\" (UID: \"d1d95573-7bea-4ee6-b0c1-13ec18b10244\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zmv42" Feb 18 09:28:05 crc kubenswrapper[4556]: I0218 09:28:05.820392 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d1d95573-7bea-4ee6-b0c1-13ec18b10244-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-zmv42\" (UID: \"d1d95573-7bea-4ee6-b0c1-13ec18b10244\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zmv42" Feb 18 09:28:05 crc kubenswrapper[4556]: I0218 09:28:05.820492 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d1d95573-7bea-4ee6-b0c1-13ec18b10244-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-zmv42\" (UID: \"d1d95573-7bea-4ee6-b0c1-13ec18b10244\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zmv42" Feb 18 09:28:05 crc kubenswrapper[4556]: I0218 09:28:05.923236 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xhlp\" (UniqueName: \"kubernetes.io/projected/d1d95573-7bea-4ee6-b0c1-13ec18b10244-kube-api-access-4xhlp\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-zmv42\" (UID: \"d1d95573-7bea-4ee6-b0c1-13ec18b10244\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zmv42" Feb 18 09:28:05 crc kubenswrapper[4556]: I0218 09:28:05.923390 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d1d95573-7bea-4ee6-b0c1-13ec18b10244-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-zmv42\" (UID: \"d1d95573-7bea-4ee6-b0c1-13ec18b10244\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zmv42" Feb 18 09:28:05 crc kubenswrapper[4556]: I0218 09:28:05.923435 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d1d95573-7bea-4ee6-b0c1-13ec18b10244-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-zmv42\" (UID: \"d1d95573-7bea-4ee6-b0c1-13ec18b10244\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zmv42" Feb 18 09:28:05 crc kubenswrapper[4556]: I0218 09:28:05.930025 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d1d95573-7bea-4ee6-b0c1-13ec18b10244-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-zmv42\" (UID: \"d1d95573-7bea-4ee6-b0c1-13ec18b10244\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zmv42" Feb 18 09:28:05 crc kubenswrapper[4556]: I0218 09:28:05.930026 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d1d95573-7bea-4ee6-b0c1-13ec18b10244-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-zmv42\" (UID: \"d1d95573-7bea-4ee6-b0c1-13ec18b10244\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zmv42" Feb 18 09:28:05 crc kubenswrapper[4556]: I0218 09:28:05.939186 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xhlp\" (UniqueName: \"kubernetes.io/projected/d1d95573-7bea-4ee6-b0c1-13ec18b10244-kube-api-access-4xhlp\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-zmv42\" (UID: \"d1d95573-7bea-4ee6-b0c1-13ec18b10244\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zmv42" Feb 18 09:28:06 crc kubenswrapper[4556]: I0218 09:28:06.070242 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zmv42" Feb 18 09:28:06 crc kubenswrapper[4556]: I0218 09:28:06.514945 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-zmv42"] Feb 18 09:28:06 crc kubenswrapper[4556]: I0218 09:28:06.714357 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zmv42" event={"ID":"d1d95573-7bea-4ee6-b0c1-13ec18b10244","Type":"ContainerStarted","Data":"ea2e778944157eab6fe1c45e57dde92565ede415c31b54d37eecc89ee657051d"} Feb 18 09:28:07 crc kubenswrapper[4556]: I0218 09:28:07.723706 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zmv42" event={"ID":"d1d95573-7bea-4ee6-b0c1-13ec18b10244","Type":"ContainerStarted","Data":"f3f830d49b471bc289e8690bf16575cbc969a532603c4e2e909b92983bc84754"} Feb 18 09:28:07 crc kubenswrapper[4556]: I0218 09:28:07.735209 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zmv42" podStartSLOduration=2.223175524 podStartE2EDuration="2.735180964s" podCreationTimestamp="2026-02-18 09:28:05 +0000 UTC" firstStartedPulling="2026-02-18 09:28:06.525407129 +0000 UTC m=+1443.542368110" lastFinishedPulling="2026-02-18 09:28:07.037412569 +0000 UTC m=+1444.054373550" observedRunningTime="2026-02-18 09:28:07.734749009 +0000 UTC m=+1444.751709989" watchObservedRunningTime="2026-02-18 09:28:07.735180964 +0000 UTC m=+1444.752141944" Feb 18 09:28:09 crc kubenswrapper[4556]: I0218 09:28:09.743311 4556 scope.go:117] "RemoveContainer" containerID="4d34e95a0625ce58e4c834cf1c350945e7d68e2b836629b5f4fbd36bb21af0c5" Feb 18 09:28:09 crc kubenswrapper[4556]: I0218 09:28:09.763255 4556 scope.go:117] "RemoveContainer" containerID="0b7821217191a992f6f9836f655f75444413ac1a3fcb61632d8577074da62496" Feb 18 09:28:09 crc kubenswrapper[4556]: I0218 09:28:09.799687 4556 scope.go:117] "RemoveContainer" containerID="60fc6854fa8cb8ee9bf35bd5ed24e067c93a7990408bc10a44f23a6a41ec457f" Feb 18 09:28:09 crc kubenswrapper[4556]: I0218 09:28:09.828840 4556 scope.go:117] "RemoveContainer" containerID="d714dd60903d22c75de48292c1cf000bdfad19de9da8e937a7413be5a38e4e5e" Feb 18 09:28:09 crc kubenswrapper[4556]: I0218 09:28:09.858134 4556 scope.go:117] "RemoveContainer" containerID="90c9dc724cf9d95bf753513453202e2c030fad7cd2d4ef3b508497249f9a4a7e" Feb 18 09:28:09 crc kubenswrapper[4556]: I0218 09:28:09.890197 4556 scope.go:117] "RemoveContainer" containerID="6f8121f2008bdbb40e850334b987b119fb32bea427009c272cd281fb6b8a4735" Feb 18 09:28:09 crc kubenswrapper[4556]: I0218 09:28:09.925398 4556 scope.go:117] "RemoveContainer" containerID="98c5176b13bae0c1183c67a8105d8142bcb611f8fd37c453ef15b184633083cd" Feb 18 09:28:09 crc kubenswrapper[4556]: I0218 09:28:09.940589 4556 scope.go:117] "RemoveContainer" containerID="c429fb238171b24a587bcafa3d8a82c21f54094c5652e03e587e69b41a0106c0" Feb 18 09:28:09 crc kubenswrapper[4556]: I0218 09:28:09.961345 4556 scope.go:117] "RemoveContainer" containerID="63868b1e8917bd5b78ec577b9090971cf503941ce12fce83e038cadfaa8c7368" Feb 18 09:28:15 crc kubenswrapper[4556]: I0218 09:28:15.033879 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-brdr4"] Feb 18 09:28:15 crc kubenswrapper[4556]: I0218 09:28:15.040117 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-brdr4"] Feb 18 09:28:15 crc kubenswrapper[4556]: I0218 09:28:15.290779 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6347657-1897-458a-9a02-36d8655525f2" path="/var/lib/kubelet/pods/a6347657-1897-458a-9a02-36d8655525f2/volumes" Feb 18 09:28:21 crc kubenswrapper[4556]: I0218 09:28:21.026895 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-xhbg8"] Feb 18 09:28:21 crc kubenswrapper[4556]: I0218 09:28:21.033198 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-xhbg8"] Feb 18 09:28:21 crc kubenswrapper[4556]: I0218 09:28:21.291850 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b111dd2-11a6-47cc-ba9e-2886186c676b" path="/var/lib/kubelet/pods/2b111dd2-11a6-47cc-ba9e-2886186c676b/volumes" Feb 18 09:28:25 crc kubenswrapper[4556]: I0218 09:28:25.026015 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-r4h4k"] Feb 18 09:28:25 crc kubenswrapper[4556]: I0218 09:28:25.032192 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-r4h4k"] Feb 18 09:28:25 crc kubenswrapper[4556]: I0218 09:28:25.290014 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c46304b-1b60-493b-918f-678451f13057" path="/var/lib/kubelet/pods/8c46304b-1b60-493b-918f-678451f13057/volumes" Feb 18 09:28:31 crc kubenswrapper[4556]: I0218 09:28:31.727273 4556 patch_prober.go:28] interesting pod/machine-config-daemon-f76hs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 09:28:31 crc kubenswrapper[4556]: I0218 09:28:31.727771 4556 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 09:28:32 crc kubenswrapper[4556]: I0218 09:28:32.027831 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-dtq7t"] Feb 18 09:28:32 crc kubenswrapper[4556]: I0218 09:28:32.034515 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-dtq7t"] Feb 18 09:28:33 crc kubenswrapper[4556]: I0218 09:28:33.294710 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb11270a-c754-44b2-b0b1-937cff515ddd" path="/var/lib/kubelet/pods/cb11270a-c754-44b2-b0b1-937cff515ddd/volumes" Feb 18 09:28:33 crc kubenswrapper[4556]: I0218 09:28:33.959788 4556 generic.go:334] "Generic (PLEG): container finished" podID="d1d95573-7bea-4ee6-b0c1-13ec18b10244" containerID="f3f830d49b471bc289e8690bf16575cbc969a532603c4e2e909b92983bc84754" exitCode=0 Feb 18 09:28:33 crc kubenswrapper[4556]: I0218 09:28:33.959829 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zmv42" event={"ID":"d1d95573-7bea-4ee6-b0c1-13ec18b10244","Type":"ContainerDied","Data":"f3f830d49b471bc289e8690bf16575cbc969a532603c4e2e909b92983bc84754"} Feb 18 09:28:35 crc kubenswrapper[4556]: I0218 09:28:35.334011 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zmv42" Feb 18 09:28:35 crc kubenswrapper[4556]: I0218 09:28:35.449004 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d1d95573-7bea-4ee6-b0c1-13ec18b10244-inventory\") pod \"d1d95573-7bea-4ee6-b0c1-13ec18b10244\" (UID: \"d1d95573-7bea-4ee6-b0c1-13ec18b10244\") " Feb 18 09:28:35 crc kubenswrapper[4556]: I0218 09:28:35.449078 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d1d95573-7bea-4ee6-b0c1-13ec18b10244-ssh-key-openstack-edpm-ipam\") pod \"d1d95573-7bea-4ee6-b0c1-13ec18b10244\" (UID: \"d1d95573-7bea-4ee6-b0c1-13ec18b10244\") " Feb 18 09:28:35 crc kubenswrapper[4556]: I0218 09:28:35.449167 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4xhlp\" (UniqueName: \"kubernetes.io/projected/d1d95573-7bea-4ee6-b0c1-13ec18b10244-kube-api-access-4xhlp\") pod \"d1d95573-7bea-4ee6-b0c1-13ec18b10244\" (UID: \"d1d95573-7bea-4ee6-b0c1-13ec18b10244\") " Feb 18 09:28:35 crc kubenswrapper[4556]: I0218 09:28:35.457083 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1d95573-7bea-4ee6-b0c1-13ec18b10244-kube-api-access-4xhlp" (OuterVolumeSpecName: "kube-api-access-4xhlp") pod "d1d95573-7bea-4ee6-b0c1-13ec18b10244" (UID: "d1d95573-7bea-4ee6-b0c1-13ec18b10244"). InnerVolumeSpecName "kube-api-access-4xhlp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:28:35 crc kubenswrapper[4556]: I0218 09:28:35.476019 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1d95573-7bea-4ee6-b0c1-13ec18b10244-inventory" (OuterVolumeSpecName: "inventory") pod "d1d95573-7bea-4ee6-b0c1-13ec18b10244" (UID: "d1d95573-7bea-4ee6-b0c1-13ec18b10244"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:28:35 crc kubenswrapper[4556]: I0218 09:28:35.476367 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1d95573-7bea-4ee6-b0c1-13ec18b10244-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "d1d95573-7bea-4ee6-b0c1-13ec18b10244" (UID: "d1d95573-7bea-4ee6-b0c1-13ec18b10244"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:28:35 crc kubenswrapper[4556]: I0218 09:28:35.552140 4556 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d1d95573-7bea-4ee6-b0c1-13ec18b10244-inventory\") on node \"crc\" DevicePath \"\"" Feb 18 09:28:35 crc kubenswrapper[4556]: I0218 09:28:35.552424 4556 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d1d95573-7bea-4ee6-b0c1-13ec18b10244-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 18 09:28:35 crc kubenswrapper[4556]: I0218 09:28:35.552438 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4xhlp\" (UniqueName: \"kubernetes.io/projected/d1d95573-7bea-4ee6-b0c1-13ec18b10244-kube-api-access-4xhlp\") on node \"crc\" DevicePath \"\"" Feb 18 09:28:35 crc kubenswrapper[4556]: I0218 09:28:35.978353 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zmv42" event={"ID":"d1d95573-7bea-4ee6-b0c1-13ec18b10244","Type":"ContainerDied","Data":"ea2e778944157eab6fe1c45e57dde92565ede415c31b54d37eecc89ee657051d"} Feb 18 09:28:35 crc kubenswrapper[4556]: I0218 09:28:35.978400 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ea2e778944157eab6fe1c45e57dde92565ede415c31b54d37eecc89ee657051d" Feb 18 09:28:35 crc kubenswrapper[4556]: I0218 09:28:35.978519 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zmv42" Feb 18 09:28:36 crc kubenswrapper[4556]: I0218 09:28:36.050097 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nfwmp"] Feb 18 09:28:36 crc kubenswrapper[4556]: E0218 09:28:36.050549 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1d95573-7bea-4ee6-b0c1-13ec18b10244" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Feb 18 09:28:36 crc kubenswrapper[4556]: I0218 09:28:36.050570 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1d95573-7bea-4ee6-b0c1-13ec18b10244" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Feb 18 09:28:36 crc kubenswrapper[4556]: I0218 09:28:36.050771 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1d95573-7bea-4ee6-b0c1-13ec18b10244" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Feb 18 09:28:36 crc kubenswrapper[4556]: I0218 09:28:36.051423 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nfwmp" Feb 18 09:28:36 crc kubenswrapper[4556]: I0218 09:28:36.053213 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 18 09:28:36 crc kubenswrapper[4556]: I0218 09:28:36.053424 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-54z7h" Feb 18 09:28:36 crc kubenswrapper[4556]: I0218 09:28:36.053580 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 18 09:28:36 crc kubenswrapper[4556]: I0218 09:28:36.055335 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 18 09:28:36 crc kubenswrapper[4556]: I0218 09:28:36.060257 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nfwmp"] Feb 18 09:28:36 crc kubenswrapper[4556]: I0218 09:28:36.167959 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lm684\" (UniqueName: \"kubernetes.io/projected/c71f2826-1f14-462a-8096-cace6bd934d3-kube-api-access-lm684\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-nfwmp\" (UID: \"c71f2826-1f14-462a-8096-cace6bd934d3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nfwmp" Feb 18 09:28:36 crc kubenswrapper[4556]: I0218 09:28:36.168295 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c71f2826-1f14-462a-8096-cace6bd934d3-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-nfwmp\" (UID: \"c71f2826-1f14-462a-8096-cace6bd934d3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nfwmp" Feb 18 09:28:36 crc kubenswrapper[4556]: I0218 09:28:36.168504 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c71f2826-1f14-462a-8096-cace6bd934d3-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-nfwmp\" (UID: \"c71f2826-1f14-462a-8096-cace6bd934d3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nfwmp" Feb 18 09:28:36 crc kubenswrapper[4556]: I0218 09:28:36.270869 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c71f2826-1f14-462a-8096-cace6bd934d3-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-nfwmp\" (UID: \"c71f2826-1f14-462a-8096-cace6bd934d3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nfwmp" Feb 18 09:28:36 crc kubenswrapper[4556]: I0218 09:28:36.270986 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c71f2826-1f14-462a-8096-cace6bd934d3-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-nfwmp\" (UID: \"c71f2826-1f14-462a-8096-cace6bd934d3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nfwmp" Feb 18 09:28:36 crc kubenswrapper[4556]: I0218 09:28:36.271117 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lm684\" (UniqueName: \"kubernetes.io/projected/c71f2826-1f14-462a-8096-cace6bd934d3-kube-api-access-lm684\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-nfwmp\" (UID: \"c71f2826-1f14-462a-8096-cace6bd934d3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nfwmp" Feb 18 09:28:36 crc kubenswrapper[4556]: I0218 09:28:36.275483 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c71f2826-1f14-462a-8096-cace6bd934d3-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-nfwmp\" (UID: \"c71f2826-1f14-462a-8096-cace6bd934d3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nfwmp" Feb 18 09:28:36 crc kubenswrapper[4556]: I0218 09:28:36.276147 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c71f2826-1f14-462a-8096-cace6bd934d3-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-nfwmp\" (UID: \"c71f2826-1f14-462a-8096-cace6bd934d3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nfwmp" Feb 18 09:28:36 crc kubenswrapper[4556]: I0218 09:28:36.287795 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lm684\" (UniqueName: \"kubernetes.io/projected/c71f2826-1f14-462a-8096-cace6bd934d3-kube-api-access-lm684\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-nfwmp\" (UID: \"c71f2826-1f14-462a-8096-cace6bd934d3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nfwmp" Feb 18 09:28:36 crc kubenswrapper[4556]: I0218 09:28:36.370969 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nfwmp" Feb 18 09:28:36 crc kubenswrapper[4556]: I0218 09:28:36.858670 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nfwmp"] Feb 18 09:28:36 crc kubenswrapper[4556]: I0218 09:28:36.991773 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nfwmp" event={"ID":"c71f2826-1f14-462a-8096-cace6bd934d3","Type":"ContainerStarted","Data":"ad3736c4e1c14ea55322556d08b8a1c6221cf3f99b3950625e87b2a44d2aaedc"} Feb 18 09:28:38 crc kubenswrapper[4556]: I0218 09:28:38.011224 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nfwmp" event={"ID":"c71f2826-1f14-462a-8096-cace6bd934d3","Type":"ContainerStarted","Data":"884aa5391489cc689e18ca41ef8c3e82c037017f0a26eb183dbcc2a61de7bbca"} Feb 18 09:28:38 crc kubenswrapper[4556]: I0218 09:28:38.034543 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nfwmp" podStartSLOduration=1.537341746 podStartE2EDuration="2.034519886s" podCreationTimestamp="2026-02-18 09:28:36 +0000 UTC" firstStartedPulling="2026-02-18 09:28:36.866165165 +0000 UTC m=+1473.883126145" lastFinishedPulling="2026-02-18 09:28:37.363343305 +0000 UTC m=+1474.380304285" observedRunningTime="2026-02-18 09:28:38.026706576 +0000 UTC m=+1475.043667557" watchObservedRunningTime="2026-02-18 09:28:38.034519886 +0000 UTC m=+1475.051480855" Feb 18 09:29:01 crc kubenswrapper[4556]: I0218 09:29:01.727222 4556 patch_prober.go:28] interesting pod/machine-config-daemon-f76hs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 09:29:01 crc kubenswrapper[4556]: I0218 09:29:01.727746 4556 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 09:29:01 crc kubenswrapper[4556]: I0218 09:29:01.727788 4556 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" Feb 18 09:29:01 crc kubenswrapper[4556]: I0218 09:29:01.728368 4556 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"507a9134b6862bc8be1ac03b7993c890ef64f16c21975e00aa96c53fe413415d"} pod="openshift-machine-config-operator/machine-config-daemon-f76hs" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 18 09:29:01 crc kubenswrapper[4556]: I0218 09:29:01.728420 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerName="machine-config-daemon" containerID="cri-o://507a9134b6862bc8be1ac03b7993c890ef64f16c21975e00aa96c53fe413415d" gracePeriod=600 Feb 18 09:29:01 crc kubenswrapper[4556]: E0218 09:29:01.847404 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:29:02 crc kubenswrapper[4556]: I0218 09:29:02.241579 4556 generic.go:334] "Generic (PLEG): container finished" podID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerID="507a9134b6862bc8be1ac03b7993c890ef64f16c21975e00aa96c53fe413415d" exitCode=0 Feb 18 09:29:02 crc kubenswrapper[4556]: I0218 09:29:02.241644 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" event={"ID":"8dac7f27-d3d1-4778-9e54-f273035a1d37","Type":"ContainerDied","Data":"507a9134b6862bc8be1ac03b7993c890ef64f16c21975e00aa96c53fe413415d"} Feb 18 09:29:02 crc kubenswrapper[4556]: I0218 09:29:02.242181 4556 scope.go:117] "RemoveContainer" containerID="06650a93598feaae08bc8fc051647e71397e8b9f4be0a24b1b089259c0bc5843" Feb 18 09:29:02 crc kubenswrapper[4556]: I0218 09:29:02.242847 4556 scope.go:117] "RemoveContainer" containerID="507a9134b6862bc8be1ac03b7993c890ef64f16c21975e00aa96c53fe413415d" Feb 18 09:29:02 crc kubenswrapper[4556]: E0218 09:29:02.243200 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:29:10 crc kubenswrapper[4556]: I0218 09:29:10.099984 4556 scope.go:117] "RemoveContainer" containerID="cb6f76579994554cb029ab9574593a5fd90e6979565cbbfb1eacfea7989f8262" Feb 18 09:29:10 crc kubenswrapper[4556]: I0218 09:29:10.149256 4556 scope.go:117] "RemoveContainer" containerID="647910ecbb2afa2329d622e6150e7766438dd9b5e0ec136356c23c9ef4cfc82a" Feb 18 09:29:10 crc kubenswrapper[4556]: I0218 09:29:10.192870 4556 scope.go:117] "RemoveContainer" containerID="51fa6da89b840614647763e3ed2be148353737332fbcbef9091dbbbf265415f3" Feb 18 09:29:10 crc kubenswrapper[4556]: I0218 09:29:10.225340 4556 scope.go:117] "RemoveContainer" containerID="627b08c9854a12fcffab1aed34bc89bff2f1f8dfd17c4c63a28713fd77d9847b" Feb 18 09:29:13 crc kubenswrapper[4556]: I0218 09:29:13.360405 4556 generic.go:334] "Generic (PLEG): container finished" podID="c71f2826-1f14-462a-8096-cace6bd934d3" containerID="884aa5391489cc689e18ca41ef8c3e82c037017f0a26eb183dbcc2a61de7bbca" exitCode=0 Feb 18 09:29:13 crc kubenswrapper[4556]: I0218 09:29:13.360510 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nfwmp" event={"ID":"c71f2826-1f14-462a-8096-cace6bd934d3","Type":"ContainerDied","Data":"884aa5391489cc689e18ca41ef8c3e82c037017f0a26eb183dbcc2a61de7bbca"} Feb 18 09:29:14 crc kubenswrapper[4556]: I0218 09:29:14.735949 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nfwmp" Feb 18 09:29:14 crc kubenswrapper[4556]: I0218 09:29:14.820847 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lm684\" (UniqueName: \"kubernetes.io/projected/c71f2826-1f14-462a-8096-cace6bd934d3-kube-api-access-lm684\") pod \"c71f2826-1f14-462a-8096-cace6bd934d3\" (UID: \"c71f2826-1f14-462a-8096-cace6bd934d3\") " Feb 18 09:29:14 crc kubenswrapper[4556]: I0218 09:29:14.821211 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c71f2826-1f14-462a-8096-cace6bd934d3-ssh-key-openstack-edpm-ipam\") pod \"c71f2826-1f14-462a-8096-cace6bd934d3\" (UID: \"c71f2826-1f14-462a-8096-cace6bd934d3\") " Feb 18 09:29:14 crc kubenswrapper[4556]: I0218 09:29:14.821376 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c71f2826-1f14-462a-8096-cace6bd934d3-inventory\") pod \"c71f2826-1f14-462a-8096-cace6bd934d3\" (UID: \"c71f2826-1f14-462a-8096-cace6bd934d3\") " Feb 18 09:29:14 crc kubenswrapper[4556]: I0218 09:29:14.827957 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c71f2826-1f14-462a-8096-cace6bd934d3-kube-api-access-lm684" (OuterVolumeSpecName: "kube-api-access-lm684") pod "c71f2826-1f14-462a-8096-cace6bd934d3" (UID: "c71f2826-1f14-462a-8096-cace6bd934d3"). InnerVolumeSpecName "kube-api-access-lm684". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:29:14 crc kubenswrapper[4556]: I0218 09:29:14.846271 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c71f2826-1f14-462a-8096-cace6bd934d3-inventory" (OuterVolumeSpecName: "inventory") pod "c71f2826-1f14-462a-8096-cace6bd934d3" (UID: "c71f2826-1f14-462a-8096-cace6bd934d3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:29:14 crc kubenswrapper[4556]: I0218 09:29:14.847446 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c71f2826-1f14-462a-8096-cace6bd934d3-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "c71f2826-1f14-462a-8096-cace6bd934d3" (UID: "c71f2826-1f14-462a-8096-cace6bd934d3"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:29:14 crc kubenswrapper[4556]: I0218 09:29:14.924319 4556 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c71f2826-1f14-462a-8096-cace6bd934d3-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 18 09:29:14 crc kubenswrapper[4556]: I0218 09:29:14.924353 4556 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c71f2826-1f14-462a-8096-cace6bd934d3-inventory\") on node \"crc\" DevicePath \"\"" Feb 18 09:29:14 crc kubenswrapper[4556]: I0218 09:29:14.924364 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lm684\" (UniqueName: \"kubernetes.io/projected/c71f2826-1f14-462a-8096-cace6bd934d3-kube-api-access-lm684\") on node \"crc\" DevicePath \"\"" Feb 18 09:29:15 crc kubenswrapper[4556]: I0218 09:29:15.378637 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nfwmp" event={"ID":"c71f2826-1f14-462a-8096-cace6bd934d3","Type":"ContainerDied","Data":"ad3736c4e1c14ea55322556d08b8a1c6221cf3f99b3950625e87b2a44d2aaedc"} Feb 18 09:29:15 crc kubenswrapper[4556]: I0218 09:29:15.378967 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ad3736c4e1c14ea55322556d08b8a1c6221cf3f99b3950625e87b2a44d2aaedc" Feb 18 09:29:15 crc kubenswrapper[4556]: I0218 09:29:15.378688 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nfwmp" Feb 18 09:29:15 crc kubenswrapper[4556]: I0218 09:29:15.540224 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-92c7r"] Feb 18 09:29:15 crc kubenswrapper[4556]: E0218 09:29:15.540585 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c71f2826-1f14-462a-8096-cace6bd934d3" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Feb 18 09:29:15 crc kubenswrapper[4556]: I0218 09:29:15.540607 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="c71f2826-1f14-462a-8096-cace6bd934d3" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Feb 18 09:29:15 crc kubenswrapper[4556]: I0218 09:29:15.540804 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="c71f2826-1f14-462a-8096-cace6bd934d3" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Feb 18 09:29:15 crc kubenswrapper[4556]: I0218 09:29:15.541642 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-92c7r" Feb 18 09:29:15 crc kubenswrapper[4556]: I0218 09:29:15.543323 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 18 09:29:15 crc kubenswrapper[4556]: I0218 09:29:15.544650 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-54z7h" Feb 18 09:29:15 crc kubenswrapper[4556]: I0218 09:29:15.544730 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 18 09:29:15 crc kubenswrapper[4556]: I0218 09:29:15.550550 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-92c7r"] Feb 18 09:29:15 crc kubenswrapper[4556]: I0218 09:29:15.551888 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 18 09:29:15 crc kubenswrapper[4556]: I0218 09:29:15.635975 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6cq7\" (UniqueName: \"kubernetes.io/projected/11adbf49-97c4-4878-96c0-4fa453f7a819-kube-api-access-h6cq7\") pod \"ssh-known-hosts-edpm-deployment-92c7r\" (UID: \"11adbf49-97c4-4878-96c0-4fa453f7a819\") " pod="openstack/ssh-known-hosts-edpm-deployment-92c7r" Feb 18 09:29:15 crc kubenswrapper[4556]: I0218 09:29:15.636303 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/11adbf49-97c4-4878-96c0-4fa453f7a819-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-92c7r\" (UID: \"11adbf49-97c4-4878-96c0-4fa453f7a819\") " pod="openstack/ssh-known-hosts-edpm-deployment-92c7r" Feb 18 09:29:15 crc kubenswrapper[4556]: I0218 09:29:15.636510 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/11adbf49-97c4-4878-96c0-4fa453f7a819-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-92c7r\" (UID: \"11adbf49-97c4-4878-96c0-4fa453f7a819\") " pod="openstack/ssh-known-hosts-edpm-deployment-92c7r" Feb 18 09:29:15 crc kubenswrapper[4556]: I0218 09:29:15.737769 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6cq7\" (UniqueName: \"kubernetes.io/projected/11adbf49-97c4-4878-96c0-4fa453f7a819-kube-api-access-h6cq7\") pod \"ssh-known-hosts-edpm-deployment-92c7r\" (UID: \"11adbf49-97c4-4878-96c0-4fa453f7a819\") " pod="openstack/ssh-known-hosts-edpm-deployment-92c7r" Feb 18 09:29:15 crc kubenswrapper[4556]: I0218 09:29:15.737871 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/11adbf49-97c4-4878-96c0-4fa453f7a819-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-92c7r\" (UID: \"11adbf49-97c4-4878-96c0-4fa453f7a819\") " pod="openstack/ssh-known-hosts-edpm-deployment-92c7r" Feb 18 09:29:15 crc kubenswrapper[4556]: I0218 09:29:15.737989 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/11adbf49-97c4-4878-96c0-4fa453f7a819-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-92c7r\" (UID: \"11adbf49-97c4-4878-96c0-4fa453f7a819\") " pod="openstack/ssh-known-hosts-edpm-deployment-92c7r" Feb 18 09:29:15 crc kubenswrapper[4556]: I0218 09:29:15.741680 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/11adbf49-97c4-4878-96c0-4fa453f7a819-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-92c7r\" (UID: \"11adbf49-97c4-4878-96c0-4fa453f7a819\") " pod="openstack/ssh-known-hosts-edpm-deployment-92c7r" Feb 18 09:29:15 crc kubenswrapper[4556]: I0218 09:29:15.741883 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/11adbf49-97c4-4878-96c0-4fa453f7a819-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-92c7r\" (UID: \"11adbf49-97c4-4878-96c0-4fa453f7a819\") " pod="openstack/ssh-known-hosts-edpm-deployment-92c7r" Feb 18 09:29:15 crc kubenswrapper[4556]: I0218 09:29:15.753103 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6cq7\" (UniqueName: \"kubernetes.io/projected/11adbf49-97c4-4878-96c0-4fa453f7a819-kube-api-access-h6cq7\") pod \"ssh-known-hosts-edpm-deployment-92c7r\" (UID: \"11adbf49-97c4-4878-96c0-4fa453f7a819\") " pod="openstack/ssh-known-hosts-edpm-deployment-92c7r" Feb 18 09:29:15 crc kubenswrapper[4556]: I0218 09:29:15.864178 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-92c7r" Feb 18 09:29:16 crc kubenswrapper[4556]: I0218 09:29:16.049369 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-6c31-account-create-update-2wxx2"] Feb 18 09:29:16 crc kubenswrapper[4556]: I0218 09:29:16.058690 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-f7xdx"] Feb 18 09:29:16 crc kubenswrapper[4556]: I0218 09:29:16.070200 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-2c5c-account-create-update-7z6nx"] Feb 18 09:29:16 crc kubenswrapper[4556]: I0218 09:29:16.079411 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-5bef-account-create-update-g2mzr"] Feb 18 09:29:16 crc kubenswrapper[4556]: I0218 09:29:16.087844 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-nq759"] Feb 18 09:29:16 crc kubenswrapper[4556]: I0218 09:29:16.095488 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-ln2s2"] Feb 18 09:29:16 crc kubenswrapper[4556]: I0218 09:29:16.101253 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-2c5c-account-create-update-7z6nx"] Feb 18 09:29:16 crc kubenswrapper[4556]: I0218 09:29:16.107682 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-6c31-account-create-update-2wxx2"] Feb 18 09:29:16 crc kubenswrapper[4556]: I0218 09:29:16.115295 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-f7xdx"] Feb 18 09:29:16 crc kubenswrapper[4556]: I0218 09:29:16.124295 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-ln2s2"] Feb 18 09:29:16 crc kubenswrapper[4556]: I0218 09:29:16.143604 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-nq759"] Feb 18 09:29:16 crc kubenswrapper[4556]: I0218 09:29:16.151022 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-5bef-account-create-update-g2mzr"] Feb 18 09:29:16 crc kubenswrapper[4556]: I0218 09:29:16.282778 4556 scope.go:117] "RemoveContainer" containerID="507a9134b6862bc8be1ac03b7993c890ef64f16c21975e00aa96c53fe413415d" Feb 18 09:29:16 crc kubenswrapper[4556]: E0218 09:29:16.283249 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:29:16 crc kubenswrapper[4556]: I0218 09:29:16.363030 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-92c7r"] Feb 18 09:29:16 crc kubenswrapper[4556]: I0218 09:29:16.387898 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-92c7r" event={"ID":"11adbf49-97c4-4878-96c0-4fa453f7a819","Type":"ContainerStarted","Data":"39520e9ed27e44d4551488c596b1edf856b9d1a1580031f209bb9b467fad8f95"} Feb 18 09:29:17 crc kubenswrapper[4556]: I0218 09:29:17.290950 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65bf8803-f4f0-4366-8f1f-d894f2f407e2" path="/var/lib/kubelet/pods/65bf8803-f4f0-4366-8f1f-d894f2f407e2/volumes" Feb 18 09:29:17 crc kubenswrapper[4556]: I0218 09:29:17.291825 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7da9e29c-e9f2-49bf-a3b3-6e1b764c2b78" path="/var/lib/kubelet/pods/7da9e29c-e9f2-49bf-a3b3-6e1b764c2b78/volumes" Feb 18 09:29:17 crc kubenswrapper[4556]: I0218 09:29:17.292366 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96a4b91d-0008-4af7-8522-6842a8c627a8" path="/var/lib/kubelet/pods/96a4b91d-0008-4af7-8522-6842a8c627a8/volumes" Feb 18 09:29:17 crc kubenswrapper[4556]: I0218 09:29:17.292891 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d482227-d966-4319-bcfd-10944085417b" path="/var/lib/kubelet/pods/9d482227-d966-4319-bcfd-10944085417b/volumes" Feb 18 09:29:17 crc kubenswrapper[4556]: I0218 09:29:17.293808 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3bab749-15ef-4b1d-a9ab-f51acb445a96" path="/var/lib/kubelet/pods/b3bab749-15ef-4b1d-a9ab-f51acb445a96/volumes" Feb 18 09:29:17 crc kubenswrapper[4556]: I0218 09:29:17.294303 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d89c1ba7-214f-45af-8a78-4f52474b16f7" path="/var/lib/kubelet/pods/d89c1ba7-214f-45af-8a78-4f52474b16f7/volumes" Feb 18 09:29:17 crc kubenswrapper[4556]: I0218 09:29:17.397629 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-92c7r" event={"ID":"11adbf49-97c4-4878-96c0-4fa453f7a819","Type":"ContainerStarted","Data":"44fd49ff6060fd64096ef64206e5157c2ef5f722b989de7f2797a84611187e72"} Feb 18 09:29:17 crc kubenswrapper[4556]: I0218 09:29:17.422745 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-92c7r" podStartSLOduration=1.869773834 podStartE2EDuration="2.422721266s" podCreationTimestamp="2026-02-18 09:29:15 +0000 UTC" firstStartedPulling="2026-02-18 09:29:16.370608764 +0000 UTC m=+1513.387569745" lastFinishedPulling="2026-02-18 09:29:16.923556208 +0000 UTC m=+1513.940517177" observedRunningTime="2026-02-18 09:29:17.411168312 +0000 UTC m=+1514.428129292" watchObservedRunningTime="2026-02-18 09:29:17.422721266 +0000 UTC m=+1514.439682245" Feb 18 09:29:22 crc kubenswrapper[4556]: I0218 09:29:22.443593 4556 generic.go:334] "Generic (PLEG): container finished" podID="11adbf49-97c4-4878-96c0-4fa453f7a819" containerID="44fd49ff6060fd64096ef64206e5157c2ef5f722b989de7f2797a84611187e72" exitCode=0 Feb 18 09:29:22 crc kubenswrapper[4556]: I0218 09:29:22.443708 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-92c7r" event={"ID":"11adbf49-97c4-4878-96c0-4fa453f7a819","Type":"ContainerDied","Data":"44fd49ff6060fd64096ef64206e5157c2ef5f722b989de7f2797a84611187e72"} Feb 18 09:29:23 crc kubenswrapper[4556]: I0218 09:29:23.792694 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-92c7r" Feb 18 09:29:23 crc kubenswrapper[4556]: I0218 09:29:23.815788 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/11adbf49-97c4-4878-96c0-4fa453f7a819-inventory-0\") pod \"11adbf49-97c4-4878-96c0-4fa453f7a819\" (UID: \"11adbf49-97c4-4878-96c0-4fa453f7a819\") " Feb 18 09:29:23 crc kubenswrapper[4556]: I0218 09:29:23.815881 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/11adbf49-97c4-4878-96c0-4fa453f7a819-ssh-key-openstack-edpm-ipam\") pod \"11adbf49-97c4-4878-96c0-4fa453f7a819\" (UID: \"11adbf49-97c4-4878-96c0-4fa453f7a819\") " Feb 18 09:29:23 crc kubenswrapper[4556]: I0218 09:29:23.815944 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h6cq7\" (UniqueName: \"kubernetes.io/projected/11adbf49-97c4-4878-96c0-4fa453f7a819-kube-api-access-h6cq7\") pod \"11adbf49-97c4-4878-96c0-4fa453f7a819\" (UID: \"11adbf49-97c4-4878-96c0-4fa453f7a819\") " Feb 18 09:29:23 crc kubenswrapper[4556]: I0218 09:29:23.827335 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11adbf49-97c4-4878-96c0-4fa453f7a819-kube-api-access-h6cq7" (OuterVolumeSpecName: "kube-api-access-h6cq7") pod "11adbf49-97c4-4878-96c0-4fa453f7a819" (UID: "11adbf49-97c4-4878-96c0-4fa453f7a819"). InnerVolumeSpecName "kube-api-access-h6cq7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:29:23 crc kubenswrapper[4556]: I0218 09:29:23.841172 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11adbf49-97c4-4878-96c0-4fa453f7a819-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "11adbf49-97c4-4878-96c0-4fa453f7a819" (UID: "11adbf49-97c4-4878-96c0-4fa453f7a819"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:29:23 crc kubenswrapper[4556]: I0218 09:29:23.846507 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11adbf49-97c4-4878-96c0-4fa453f7a819-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "11adbf49-97c4-4878-96c0-4fa453f7a819" (UID: "11adbf49-97c4-4878-96c0-4fa453f7a819"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:29:23 crc kubenswrapper[4556]: I0218 09:29:23.918400 4556 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/11adbf49-97c4-4878-96c0-4fa453f7a819-inventory-0\") on node \"crc\" DevicePath \"\"" Feb 18 09:29:23 crc kubenswrapper[4556]: I0218 09:29:23.918431 4556 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/11adbf49-97c4-4878-96c0-4fa453f7a819-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 18 09:29:23 crc kubenswrapper[4556]: I0218 09:29:23.918442 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h6cq7\" (UniqueName: \"kubernetes.io/projected/11adbf49-97c4-4878-96c0-4fa453f7a819-kube-api-access-h6cq7\") on node \"crc\" DevicePath \"\"" Feb 18 09:29:24 crc kubenswrapper[4556]: I0218 09:29:24.465396 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-92c7r" event={"ID":"11adbf49-97c4-4878-96c0-4fa453f7a819","Type":"ContainerDied","Data":"39520e9ed27e44d4551488c596b1edf856b9d1a1580031f209bb9b467fad8f95"} Feb 18 09:29:24 crc kubenswrapper[4556]: I0218 09:29:24.465444 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="39520e9ed27e44d4551488c596b1edf856b9d1a1580031f209bb9b467fad8f95" Feb 18 09:29:24 crc kubenswrapper[4556]: I0218 09:29:24.465471 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-92c7r" Feb 18 09:29:24 crc kubenswrapper[4556]: I0218 09:29:24.545766 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-2j6nf"] Feb 18 09:29:24 crc kubenswrapper[4556]: E0218 09:29:24.546300 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11adbf49-97c4-4878-96c0-4fa453f7a819" containerName="ssh-known-hosts-edpm-deployment" Feb 18 09:29:24 crc kubenswrapper[4556]: I0218 09:29:24.546322 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="11adbf49-97c4-4878-96c0-4fa453f7a819" containerName="ssh-known-hosts-edpm-deployment" Feb 18 09:29:24 crc kubenswrapper[4556]: I0218 09:29:24.546563 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="11adbf49-97c4-4878-96c0-4fa453f7a819" containerName="ssh-known-hosts-edpm-deployment" Feb 18 09:29:24 crc kubenswrapper[4556]: I0218 09:29:24.548789 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2j6nf" Feb 18 09:29:24 crc kubenswrapper[4556]: I0218 09:29:24.553572 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 18 09:29:24 crc kubenswrapper[4556]: I0218 09:29:24.553797 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 18 09:29:24 crc kubenswrapper[4556]: I0218 09:29:24.553954 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-54z7h" Feb 18 09:29:24 crc kubenswrapper[4556]: I0218 09:29:24.554911 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 18 09:29:24 crc kubenswrapper[4556]: I0218 09:29:24.562502 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-2j6nf"] Feb 18 09:29:24 crc kubenswrapper[4556]: I0218 09:29:24.633430 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgb2w\" (UniqueName: \"kubernetes.io/projected/1d8fa453-3277-4c76-9098-0e7838fd8d44-kube-api-access-tgb2w\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-2j6nf\" (UID: \"1d8fa453-3277-4c76-9098-0e7838fd8d44\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2j6nf" Feb 18 09:29:24 crc kubenswrapper[4556]: I0218 09:29:24.633775 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1d8fa453-3277-4c76-9098-0e7838fd8d44-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-2j6nf\" (UID: \"1d8fa453-3277-4c76-9098-0e7838fd8d44\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2j6nf" Feb 18 09:29:24 crc kubenswrapper[4556]: I0218 09:29:24.633855 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1d8fa453-3277-4c76-9098-0e7838fd8d44-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-2j6nf\" (UID: \"1d8fa453-3277-4c76-9098-0e7838fd8d44\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2j6nf" Feb 18 09:29:24 crc kubenswrapper[4556]: I0218 09:29:24.736335 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1d8fa453-3277-4c76-9098-0e7838fd8d44-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-2j6nf\" (UID: \"1d8fa453-3277-4c76-9098-0e7838fd8d44\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2j6nf" Feb 18 09:29:24 crc kubenswrapper[4556]: I0218 09:29:24.736468 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1d8fa453-3277-4c76-9098-0e7838fd8d44-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-2j6nf\" (UID: \"1d8fa453-3277-4c76-9098-0e7838fd8d44\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2j6nf" Feb 18 09:29:24 crc kubenswrapper[4556]: I0218 09:29:24.736947 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgb2w\" (UniqueName: \"kubernetes.io/projected/1d8fa453-3277-4c76-9098-0e7838fd8d44-kube-api-access-tgb2w\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-2j6nf\" (UID: \"1d8fa453-3277-4c76-9098-0e7838fd8d44\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2j6nf" Feb 18 09:29:24 crc kubenswrapper[4556]: I0218 09:29:24.741818 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1d8fa453-3277-4c76-9098-0e7838fd8d44-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-2j6nf\" (UID: \"1d8fa453-3277-4c76-9098-0e7838fd8d44\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2j6nf" Feb 18 09:29:24 crc kubenswrapper[4556]: I0218 09:29:24.742879 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1d8fa453-3277-4c76-9098-0e7838fd8d44-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-2j6nf\" (UID: \"1d8fa453-3277-4c76-9098-0e7838fd8d44\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2j6nf" Feb 18 09:29:24 crc kubenswrapper[4556]: I0218 09:29:24.756838 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgb2w\" (UniqueName: \"kubernetes.io/projected/1d8fa453-3277-4c76-9098-0e7838fd8d44-kube-api-access-tgb2w\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-2j6nf\" (UID: \"1d8fa453-3277-4c76-9098-0e7838fd8d44\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2j6nf" Feb 18 09:29:24 crc kubenswrapper[4556]: I0218 09:29:24.864592 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2j6nf" Feb 18 09:29:25 crc kubenswrapper[4556]: I0218 09:29:25.333875 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-2j6nf"] Feb 18 09:29:25 crc kubenswrapper[4556]: I0218 09:29:25.475238 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2j6nf" event={"ID":"1d8fa453-3277-4c76-9098-0e7838fd8d44","Type":"ContainerStarted","Data":"ed75b70df33722c698fb27e873a9805286b5ec65cfb651475f2fe2c68b4a3419"} Feb 18 09:29:26 crc kubenswrapper[4556]: I0218 09:29:26.488503 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2j6nf" event={"ID":"1d8fa453-3277-4c76-9098-0e7838fd8d44","Type":"ContainerStarted","Data":"b86a60f1e3ef4665e569bce6811b271010fb7914ba59c416682d443b0e7e97e2"} Feb 18 09:29:26 crc kubenswrapper[4556]: I0218 09:29:26.513018 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2j6nf" podStartSLOduration=2.01297846 podStartE2EDuration="2.513002177s" podCreationTimestamp="2026-02-18 09:29:24 +0000 UTC" firstStartedPulling="2026-02-18 09:29:25.346311234 +0000 UTC m=+1522.363272215" lastFinishedPulling="2026-02-18 09:29:25.846334952 +0000 UTC m=+1522.863295932" observedRunningTime="2026-02-18 09:29:26.506005779 +0000 UTC m=+1523.522966758" watchObservedRunningTime="2026-02-18 09:29:26.513002177 +0000 UTC m=+1523.529963156" Feb 18 09:29:30 crc kubenswrapper[4556]: I0218 09:29:30.284299 4556 scope.go:117] "RemoveContainer" containerID="507a9134b6862bc8be1ac03b7993c890ef64f16c21975e00aa96c53fe413415d" Feb 18 09:29:30 crc kubenswrapper[4556]: E0218 09:29:30.285717 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:29:32 crc kubenswrapper[4556]: I0218 09:29:32.546483 4556 generic.go:334] "Generic (PLEG): container finished" podID="1d8fa453-3277-4c76-9098-0e7838fd8d44" containerID="b86a60f1e3ef4665e569bce6811b271010fb7914ba59c416682d443b0e7e97e2" exitCode=0 Feb 18 09:29:32 crc kubenswrapper[4556]: I0218 09:29:32.546562 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2j6nf" event={"ID":"1d8fa453-3277-4c76-9098-0e7838fd8d44","Type":"ContainerDied","Data":"b86a60f1e3ef4665e569bce6811b271010fb7914ba59c416682d443b0e7e97e2"} Feb 18 09:29:33 crc kubenswrapper[4556]: I0218 09:29:33.930132 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2j6nf" Feb 18 09:29:33 crc kubenswrapper[4556]: I0218 09:29:33.960190 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tgb2w\" (UniqueName: \"kubernetes.io/projected/1d8fa453-3277-4c76-9098-0e7838fd8d44-kube-api-access-tgb2w\") pod \"1d8fa453-3277-4c76-9098-0e7838fd8d44\" (UID: \"1d8fa453-3277-4c76-9098-0e7838fd8d44\") " Feb 18 09:29:33 crc kubenswrapper[4556]: I0218 09:29:33.960294 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1d8fa453-3277-4c76-9098-0e7838fd8d44-ssh-key-openstack-edpm-ipam\") pod \"1d8fa453-3277-4c76-9098-0e7838fd8d44\" (UID: \"1d8fa453-3277-4c76-9098-0e7838fd8d44\") " Feb 18 09:29:33 crc kubenswrapper[4556]: I0218 09:29:33.960402 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1d8fa453-3277-4c76-9098-0e7838fd8d44-inventory\") pod \"1d8fa453-3277-4c76-9098-0e7838fd8d44\" (UID: \"1d8fa453-3277-4c76-9098-0e7838fd8d44\") " Feb 18 09:29:33 crc kubenswrapper[4556]: I0218 09:29:33.967119 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d8fa453-3277-4c76-9098-0e7838fd8d44-kube-api-access-tgb2w" (OuterVolumeSpecName: "kube-api-access-tgb2w") pod "1d8fa453-3277-4c76-9098-0e7838fd8d44" (UID: "1d8fa453-3277-4c76-9098-0e7838fd8d44"). InnerVolumeSpecName "kube-api-access-tgb2w". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:29:33 crc kubenswrapper[4556]: I0218 09:29:33.992903 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d8fa453-3277-4c76-9098-0e7838fd8d44-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "1d8fa453-3277-4c76-9098-0e7838fd8d44" (UID: "1d8fa453-3277-4c76-9098-0e7838fd8d44"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:29:33 crc kubenswrapper[4556]: I0218 09:29:33.996364 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d8fa453-3277-4c76-9098-0e7838fd8d44-inventory" (OuterVolumeSpecName: "inventory") pod "1d8fa453-3277-4c76-9098-0e7838fd8d44" (UID: "1d8fa453-3277-4c76-9098-0e7838fd8d44"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:29:34 crc kubenswrapper[4556]: I0218 09:29:34.062071 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tgb2w\" (UniqueName: \"kubernetes.io/projected/1d8fa453-3277-4c76-9098-0e7838fd8d44-kube-api-access-tgb2w\") on node \"crc\" DevicePath \"\"" Feb 18 09:29:34 crc kubenswrapper[4556]: I0218 09:29:34.062108 4556 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1d8fa453-3277-4c76-9098-0e7838fd8d44-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 18 09:29:34 crc kubenswrapper[4556]: I0218 09:29:34.062120 4556 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1d8fa453-3277-4c76-9098-0e7838fd8d44-inventory\") on node \"crc\" DevicePath \"\"" Feb 18 09:29:34 crc kubenswrapper[4556]: I0218 09:29:34.569618 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2j6nf" event={"ID":"1d8fa453-3277-4c76-9098-0e7838fd8d44","Type":"ContainerDied","Data":"ed75b70df33722c698fb27e873a9805286b5ec65cfb651475f2fe2c68b4a3419"} Feb 18 09:29:34 crc kubenswrapper[4556]: I0218 09:29:34.569696 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ed75b70df33722c698fb27e873a9805286b5ec65cfb651475f2fe2c68b4a3419" Feb 18 09:29:34 crc kubenswrapper[4556]: I0218 09:29:34.570147 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2j6nf" Feb 18 09:29:34 crc kubenswrapper[4556]: I0218 09:29:34.625999 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jmgv7"] Feb 18 09:29:34 crc kubenswrapper[4556]: E0218 09:29:34.626463 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d8fa453-3277-4c76-9098-0e7838fd8d44" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Feb 18 09:29:34 crc kubenswrapper[4556]: I0218 09:29:34.626483 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d8fa453-3277-4c76-9098-0e7838fd8d44" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Feb 18 09:29:34 crc kubenswrapper[4556]: I0218 09:29:34.626665 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d8fa453-3277-4c76-9098-0e7838fd8d44" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Feb 18 09:29:34 crc kubenswrapper[4556]: I0218 09:29:34.627265 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jmgv7" Feb 18 09:29:34 crc kubenswrapper[4556]: I0218 09:29:34.629213 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 18 09:29:34 crc kubenswrapper[4556]: I0218 09:29:34.629413 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 18 09:29:34 crc kubenswrapper[4556]: I0218 09:29:34.631316 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 18 09:29:34 crc kubenswrapper[4556]: I0218 09:29:34.631586 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-54z7h" Feb 18 09:29:34 crc kubenswrapper[4556]: I0218 09:29:34.638568 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jmgv7"] Feb 18 09:29:34 crc kubenswrapper[4556]: I0218 09:29:34.672537 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d0b159c4-dbac-4ddb-af28-523bb86639e5-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-jmgv7\" (UID: \"d0b159c4-dbac-4ddb-af28-523bb86639e5\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jmgv7" Feb 18 09:29:34 crc kubenswrapper[4556]: I0218 09:29:34.672621 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blg8l\" (UniqueName: \"kubernetes.io/projected/d0b159c4-dbac-4ddb-af28-523bb86639e5-kube-api-access-blg8l\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-jmgv7\" (UID: \"d0b159c4-dbac-4ddb-af28-523bb86639e5\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jmgv7" Feb 18 09:29:34 crc kubenswrapper[4556]: I0218 09:29:34.672861 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d0b159c4-dbac-4ddb-af28-523bb86639e5-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-jmgv7\" (UID: \"d0b159c4-dbac-4ddb-af28-523bb86639e5\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jmgv7" Feb 18 09:29:34 crc kubenswrapper[4556]: I0218 09:29:34.775565 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d0b159c4-dbac-4ddb-af28-523bb86639e5-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-jmgv7\" (UID: \"d0b159c4-dbac-4ddb-af28-523bb86639e5\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jmgv7" Feb 18 09:29:34 crc kubenswrapper[4556]: I0218 09:29:34.775663 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blg8l\" (UniqueName: \"kubernetes.io/projected/d0b159c4-dbac-4ddb-af28-523bb86639e5-kube-api-access-blg8l\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-jmgv7\" (UID: \"d0b159c4-dbac-4ddb-af28-523bb86639e5\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jmgv7" Feb 18 09:29:34 crc kubenswrapper[4556]: I0218 09:29:34.775815 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d0b159c4-dbac-4ddb-af28-523bb86639e5-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-jmgv7\" (UID: \"d0b159c4-dbac-4ddb-af28-523bb86639e5\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jmgv7" Feb 18 09:29:34 crc kubenswrapper[4556]: I0218 09:29:34.782643 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d0b159c4-dbac-4ddb-af28-523bb86639e5-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-jmgv7\" (UID: \"d0b159c4-dbac-4ddb-af28-523bb86639e5\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jmgv7" Feb 18 09:29:34 crc kubenswrapper[4556]: I0218 09:29:34.782752 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d0b159c4-dbac-4ddb-af28-523bb86639e5-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-jmgv7\" (UID: \"d0b159c4-dbac-4ddb-af28-523bb86639e5\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jmgv7" Feb 18 09:29:34 crc kubenswrapper[4556]: I0218 09:29:34.792134 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blg8l\" (UniqueName: \"kubernetes.io/projected/d0b159c4-dbac-4ddb-af28-523bb86639e5-kube-api-access-blg8l\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-jmgv7\" (UID: \"d0b159c4-dbac-4ddb-af28-523bb86639e5\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jmgv7" Feb 18 09:29:34 crc kubenswrapper[4556]: I0218 09:29:34.942278 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jmgv7" Feb 18 09:29:35 crc kubenswrapper[4556]: I0218 09:29:35.428824 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jmgv7"] Feb 18 09:29:35 crc kubenswrapper[4556]: I0218 09:29:35.581457 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jmgv7" event={"ID":"d0b159c4-dbac-4ddb-af28-523bb86639e5","Type":"ContainerStarted","Data":"adb72f4cba6285a145419dd252b0721b1cf6d3090bd0b60f751635c2d5adaa4d"} Feb 18 09:29:36 crc kubenswrapper[4556]: I0218 09:29:36.028495 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-wxltl"] Feb 18 09:29:36 crc kubenswrapper[4556]: I0218 09:29:36.037448 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-wxltl"] Feb 18 09:29:36 crc kubenswrapper[4556]: I0218 09:29:36.596429 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jmgv7" event={"ID":"d0b159c4-dbac-4ddb-af28-523bb86639e5","Type":"ContainerStarted","Data":"afa0dcbb4dd62440fc5095301267c3aa5cd5a119e4330ed8dde6c4bdedcb86b7"} Feb 18 09:29:36 crc kubenswrapper[4556]: I0218 09:29:36.854647 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jmgv7" podStartSLOduration=2.3186644579999998 podStartE2EDuration="2.854622742s" podCreationTimestamp="2026-02-18 09:29:34 +0000 UTC" firstStartedPulling="2026-02-18 09:29:35.431349773 +0000 UTC m=+1532.448310753" lastFinishedPulling="2026-02-18 09:29:35.967308057 +0000 UTC m=+1532.984269037" observedRunningTime="2026-02-18 09:29:36.624685829 +0000 UTC m=+1533.641646808" watchObservedRunningTime="2026-02-18 09:29:36.854622742 +0000 UTC m=+1533.871583722" Feb 18 09:29:36 crc kubenswrapper[4556]: I0218 09:29:36.858809 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-8dvj9"] Feb 18 09:29:36 crc kubenswrapper[4556]: I0218 09:29:36.860794 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8dvj9" Feb 18 09:29:36 crc kubenswrapper[4556]: I0218 09:29:36.889531 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8dvj9"] Feb 18 09:29:36 crc kubenswrapper[4556]: I0218 09:29:36.932907 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5f11486-6f3d-43ce-9805-e65bea6634bb-catalog-content\") pod \"redhat-operators-8dvj9\" (UID: \"e5f11486-6f3d-43ce-9805-e65bea6634bb\") " pod="openshift-marketplace/redhat-operators-8dvj9" Feb 18 09:29:36 crc kubenswrapper[4556]: I0218 09:29:36.933057 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vz45\" (UniqueName: \"kubernetes.io/projected/e5f11486-6f3d-43ce-9805-e65bea6634bb-kube-api-access-7vz45\") pod \"redhat-operators-8dvj9\" (UID: \"e5f11486-6f3d-43ce-9805-e65bea6634bb\") " pod="openshift-marketplace/redhat-operators-8dvj9" Feb 18 09:29:36 crc kubenswrapper[4556]: I0218 09:29:36.933130 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5f11486-6f3d-43ce-9805-e65bea6634bb-utilities\") pod \"redhat-operators-8dvj9\" (UID: \"e5f11486-6f3d-43ce-9805-e65bea6634bb\") " pod="openshift-marketplace/redhat-operators-8dvj9" Feb 18 09:29:37 crc kubenswrapper[4556]: I0218 09:29:37.034310 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5f11486-6f3d-43ce-9805-e65bea6634bb-catalog-content\") pod \"redhat-operators-8dvj9\" (UID: \"e5f11486-6f3d-43ce-9805-e65bea6634bb\") " pod="openshift-marketplace/redhat-operators-8dvj9" Feb 18 09:29:37 crc kubenswrapper[4556]: I0218 09:29:37.034379 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vz45\" (UniqueName: \"kubernetes.io/projected/e5f11486-6f3d-43ce-9805-e65bea6634bb-kube-api-access-7vz45\") pod \"redhat-operators-8dvj9\" (UID: \"e5f11486-6f3d-43ce-9805-e65bea6634bb\") " pod="openshift-marketplace/redhat-operators-8dvj9" Feb 18 09:29:37 crc kubenswrapper[4556]: I0218 09:29:37.034414 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5f11486-6f3d-43ce-9805-e65bea6634bb-utilities\") pod \"redhat-operators-8dvj9\" (UID: \"e5f11486-6f3d-43ce-9805-e65bea6634bb\") " pod="openshift-marketplace/redhat-operators-8dvj9" Feb 18 09:29:37 crc kubenswrapper[4556]: I0218 09:29:37.034807 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5f11486-6f3d-43ce-9805-e65bea6634bb-catalog-content\") pod \"redhat-operators-8dvj9\" (UID: \"e5f11486-6f3d-43ce-9805-e65bea6634bb\") " pod="openshift-marketplace/redhat-operators-8dvj9" Feb 18 09:29:37 crc kubenswrapper[4556]: I0218 09:29:37.034843 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5f11486-6f3d-43ce-9805-e65bea6634bb-utilities\") pod \"redhat-operators-8dvj9\" (UID: \"e5f11486-6f3d-43ce-9805-e65bea6634bb\") " pod="openshift-marketplace/redhat-operators-8dvj9" Feb 18 09:29:37 crc kubenswrapper[4556]: I0218 09:29:37.051668 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vz45\" (UniqueName: \"kubernetes.io/projected/e5f11486-6f3d-43ce-9805-e65bea6634bb-kube-api-access-7vz45\") pod \"redhat-operators-8dvj9\" (UID: \"e5f11486-6f3d-43ce-9805-e65bea6634bb\") " pod="openshift-marketplace/redhat-operators-8dvj9" Feb 18 09:29:37 crc kubenswrapper[4556]: I0218 09:29:37.174963 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8dvj9" Feb 18 09:29:37 crc kubenswrapper[4556]: I0218 09:29:37.293209 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4a792c8-97e0-46f2-8a40-747a7cae58f5" path="/var/lib/kubelet/pods/f4a792c8-97e0-46f2-8a40-747a7cae58f5/volumes" Feb 18 09:29:37 crc kubenswrapper[4556]: I0218 09:29:37.609790 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8dvj9"] Feb 18 09:29:38 crc kubenswrapper[4556]: I0218 09:29:38.615187 4556 generic.go:334] "Generic (PLEG): container finished" podID="e5f11486-6f3d-43ce-9805-e65bea6634bb" containerID="09f5ab1825fa6eedb079cd43f10bec6ac49306fc63c24ea12d7172de69f13471" exitCode=0 Feb 18 09:29:38 crc kubenswrapper[4556]: I0218 09:29:38.615285 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8dvj9" event={"ID":"e5f11486-6f3d-43ce-9805-e65bea6634bb","Type":"ContainerDied","Data":"09f5ab1825fa6eedb079cd43f10bec6ac49306fc63c24ea12d7172de69f13471"} Feb 18 09:29:38 crc kubenswrapper[4556]: I0218 09:29:38.615737 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8dvj9" event={"ID":"e5f11486-6f3d-43ce-9805-e65bea6634bb","Type":"ContainerStarted","Data":"299c58ff38fbb31ce185988180b51edb135b0af4e9a1575df39a93ac9082c9d4"} Feb 18 09:29:39 crc kubenswrapper[4556]: I0218 09:29:39.625278 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8dvj9" event={"ID":"e5f11486-6f3d-43ce-9805-e65bea6634bb","Type":"ContainerStarted","Data":"4a7dbc86a2e040801298f30b41a97a254c0bc28d91af4498a273329371f922a1"} Feb 18 09:29:41 crc kubenswrapper[4556]: I0218 09:29:41.282926 4556 scope.go:117] "RemoveContainer" containerID="507a9134b6862bc8be1ac03b7993c890ef64f16c21975e00aa96c53fe413415d" Feb 18 09:29:41 crc kubenswrapper[4556]: E0218 09:29:41.283566 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:29:41 crc kubenswrapper[4556]: I0218 09:29:41.646474 4556 generic.go:334] "Generic (PLEG): container finished" podID="e5f11486-6f3d-43ce-9805-e65bea6634bb" containerID="4a7dbc86a2e040801298f30b41a97a254c0bc28d91af4498a273329371f922a1" exitCode=0 Feb 18 09:29:41 crc kubenswrapper[4556]: I0218 09:29:41.646542 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8dvj9" event={"ID":"e5f11486-6f3d-43ce-9805-e65bea6634bb","Type":"ContainerDied","Data":"4a7dbc86a2e040801298f30b41a97a254c0bc28d91af4498a273329371f922a1"} Feb 18 09:29:42 crc kubenswrapper[4556]: I0218 09:29:42.674994 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8dvj9" event={"ID":"e5f11486-6f3d-43ce-9805-e65bea6634bb","Type":"ContainerStarted","Data":"c7fab49f889539198845af7bf36eb77d6cd4831d9e151cf4908715a852292cdb"} Feb 18 09:29:42 crc kubenswrapper[4556]: I0218 09:29:42.699775 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-8dvj9" podStartSLOduration=3.026148278 podStartE2EDuration="6.699755079s" podCreationTimestamp="2026-02-18 09:29:36 +0000 UTC" firstStartedPulling="2026-02-18 09:29:38.61764714 +0000 UTC m=+1535.634608120" lastFinishedPulling="2026-02-18 09:29:42.29125394 +0000 UTC m=+1539.308214921" observedRunningTime="2026-02-18 09:29:42.691752472 +0000 UTC m=+1539.708713453" watchObservedRunningTime="2026-02-18 09:29:42.699755079 +0000 UTC m=+1539.716716060" Feb 18 09:29:43 crc kubenswrapper[4556]: I0218 09:29:43.689948 4556 generic.go:334] "Generic (PLEG): container finished" podID="d0b159c4-dbac-4ddb-af28-523bb86639e5" containerID="afa0dcbb4dd62440fc5095301267c3aa5cd5a119e4330ed8dde6c4bdedcb86b7" exitCode=0 Feb 18 09:29:43 crc kubenswrapper[4556]: I0218 09:29:43.690173 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jmgv7" event={"ID":"d0b159c4-dbac-4ddb-af28-523bb86639e5","Type":"ContainerDied","Data":"afa0dcbb4dd62440fc5095301267c3aa5cd5a119e4330ed8dde6c4bdedcb86b7"} Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.041950 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jmgv7" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.138839 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d0b159c4-dbac-4ddb-af28-523bb86639e5-ssh-key-openstack-edpm-ipam\") pod \"d0b159c4-dbac-4ddb-af28-523bb86639e5\" (UID: \"d0b159c4-dbac-4ddb-af28-523bb86639e5\") " Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.138893 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-blg8l\" (UniqueName: \"kubernetes.io/projected/d0b159c4-dbac-4ddb-af28-523bb86639e5-kube-api-access-blg8l\") pod \"d0b159c4-dbac-4ddb-af28-523bb86639e5\" (UID: \"d0b159c4-dbac-4ddb-af28-523bb86639e5\") " Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.138989 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d0b159c4-dbac-4ddb-af28-523bb86639e5-inventory\") pod \"d0b159c4-dbac-4ddb-af28-523bb86639e5\" (UID: \"d0b159c4-dbac-4ddb-af28-523bb86639e5\") " Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.144171 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0b159c4-dbac-4ddb-af28-523bb86639e5-kube-api-access-blg8l" (OuterVolumeSpecName: "kube-api-access-blg8l") pod "d0b159c4-dbac-4ddb-af28-523bb86639e5" (UID: "d0b159c4-dbac-4ddb-af28-523bb86639e5"). InnerVolumeSpecName "kube-api-access-blg8l". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.163056 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0b159c4-dbac-4ddb-af28-523bb86639e5-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "d0b159c4-dbac-4ddb-af28-523bb86639e5" (UID: "d0b159c4-dbac-4ddb-af28-523bb86639e5"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.165519 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0b159c4-dbac-4ddb-af28-523bb86639e5-inventory" (OuterVolumeSpecName: "inventory") pod "d0b159c4-dbac-4ddb-af28-523bb86639e5" (UID: "d0b159c4-dbac-4ddb-af28-523bb86639e5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.242441 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-blg8l\" (UniqueName: \"kubernetes.io/projected/d0b159c4-dbac-4ddb-af28-523bb86639e5-kube-api-access-blg8l\") on node \"crc\" DevicePath \"\"" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.242479 4556 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d0b159c4-dbac-4ddb-af28-523bb86639e5-inventory\") on node \"crc\" DevicePath \"\"" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.242490 4556 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d0b159c4-dbac-4ddb-af28-523bb86639e5-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.711615 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jmgv7" event={"ID":"d0b159c4-dbac-4ddb-af28-523bb86639e5","Type":"ContainerDied","Data":"adb72f4cba6285a145419dd252b0721b1cf6d3090bd0b60f751635c2d5adaa4d"} Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.711658 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="adb72f4cba6285a145419dd252b0721b1cf6d3090bd0b60f751635c2d5adaa4d" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.711681 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jmgv7" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.778035 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9trr"] Feb 18 09:29:45 crc kubenswrapper[4556]: E0218 09:29:45.778689 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0b159c4-dbac-4ddb-af28-523bb86639e5" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.778711 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0b159c4-dbac-4ddb-af28-523bb86639e5" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.778906 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0b159c4-dbac-4ddb-af28-523bb86639e5" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.779479 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9trr" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.782316 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.782325 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.782495 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.782785 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.782865 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.782794 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.783098 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-54z7h" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.787742 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.792917 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9trr"] Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.858438 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9trr\" (UID: \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9trr" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.858559 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9trr\" (UID: \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9trr" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.858693 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prm9w\" (UniqueName: \"kubernetes.io/projected/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-kube-api-access-prm9w\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9trr\" (UID: \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9trr" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.858803 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9trr\" (UID: \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9trr" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.858874 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9trr\" (UID: \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9trr" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.858952 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9trr\" (UID: \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9trr" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.858990 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9trr\" (UID: \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9trr" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.859055 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9trr\" (UID: \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9trr" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.859101 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9trr\" (UID: \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9trr" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.859273 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9trr\" (UID: \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9trr" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.859327 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9trr\" (UID: \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9trr" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.859407 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9trr\" (UID: \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9trr" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.859526 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9trr\" (UID: \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9trr" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.859713 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9trr\" (UID: \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9trr" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.960693 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9trr\" (UID: \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9trr" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.960763 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9trr\" (UID: \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9trr" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.960804 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9trr\" (UID: \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9trr" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.960858 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9trr\" (UID: \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9trr" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.960887 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9trr\" (UID: \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9trr" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.960942 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9trr\" (UID: \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9trr" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.960965 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prm9w\" (UniqueName: \"kubernetes.io/projected/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-kube-api-access-prm9w\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9trr\" (UID: \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9trr" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.960992 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9trr\" (UID: \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9trr" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.961025 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9trr\" (UID: \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9trr" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.961052 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9trr\" (UID: \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9trr" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.961094 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9trr\" (UID: \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9trr" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.961125 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9trr\" (UID: \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9trr" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.961175 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9trr\" (UID: \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9trr" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.961218 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9trr\" (UID: \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9trr" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.968017 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9trr\" (UID: \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9trr" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.968342 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9trr\" (UID: \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9trr" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.968591 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9trr\" (UID: \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9trr" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.969075 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9trr\" (UID: \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9trr" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.969921 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9trr\" (UID: \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9trr" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.970245 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9trr\" (UID: \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9trr" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.970613 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9trr\" (UID: \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9trr" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.972020 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9trr\" (UID: \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9trr" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.972026 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9trr\" (UID: \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9trr" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.972524 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9trr\" (UID: \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9trr" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.973109 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9trr\" (UID: \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9trr" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.973292 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9trr\" (UID: \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9trr" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.973403 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9trr\" (UID: \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9trr" Feb 18 09:29:45 crc kubenswrapper[4556]: I0218 09:29:45.979268 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prm9w\" (UniqueName: \"kubernetes.io/projected/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-kube-api-access-prm9w\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9trr\" (UID: \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9trr" Feb 18 09:29:46 crc kubenswrapper[4556]: I0218 09:29:46.095010 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9trr" Feb 18 09:29:46 crc kubenswrapper[4556]: I0218 09:29:46.587119 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9trr"] Feb 18 09:29:46 crc kubenswrapper[4556]: I0218 09:29:46.723438 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9trr" event={"ID":"d249a625-7b1c-4a2f-a92d-c18d7fbb6142","Type":"ContainerStarted","Data":"2cb94cb3ce78a2552320a2566f80b218d4b9d2b03e8028253ed3463ce10490b4"} Feb 18 09:29:47 crc kubenswrapper[4556]: I0218 09:29:47.176555 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-8dvj9" Feb 18 09:29:47 crc kubenswrapper[4556]: I0218 09:29:47.176847 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-8dvj9" Feb 18 09:29:47 crc kubenswrapper[4556]: I0218 09:29:47.731341 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9trr" event={"ID":"d249a625-7b1c-4a2f-a92d-c18d7fbb6142","Type":"ContainerStarted","Data":"336874da07be51d4b7bb0a939adfd8f360a7e23c7a6def27e6b3897ad2a9a979"} Feb 18 09:29:47 crc kubenswrapper[4556]: I0218 09:29:47.763146 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9trr" podStartSLOduration=2.240902241 podStartE2EDuration="2.763124264s" podCreationTimestamp="2026-02-18 09:29:45 +0000 UTC" firstStartedPulling="2026-02-18 09:29:46.595495053 +0000 UTC m=+1543.612456032" lastFinishedPulling="2026-02-18 09:29:47.117717074 +0000 UTC m=+1544.134678055" observedRunningTime="2026-02-18 09:29:47.751508442 +0000 UTC m=+1544.768469421" watchObservedRunningTime="2026-02-18 09:29:47.763124264 +0000 UTC m=+1544.780085244" Feb 18 09:29:48 crc kubenswrapper[4556]: I0218 09:29:48.218690 4556 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8dvj9" podUID="e5f11486-6f3d-43ce-9805-e65bea6634bb" containerName="registry-server" probeResult="failure" output=< Feb 18 09:29:48 crc kubenswrapper[4556]: timeout: failed to connect service ":50051" within 1s Feb 18 09:29:48 crc kubenswrapper[4556]: > Feb 18 09:29:53 crc kubenswrapper[4556]: I0218 09:29:53.289030 4556 scope.go:117] "RemoveContainer" containerID="507a9134b6862bc8be1ac03b7993c890ef64f16c21975e00aa96c53fe413415d" Feb 18 09:29:53 crc kubenswrapper[4556]: E0218 09:29:53.289545 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:29:56 crc kubenswrapper[4556]: I0218 09:29:56.041627 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-fj6lv"] Feb 18 09:29:56 crc kubenswrapper[4556]: I0218 09:29:56.050566 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-fj6lv"] Feb 18 09:29:57 crc kubenswrapper[4556]: I0218 09:29:57.027370 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-dgmmz"] Feb 18 09:29:57 crc kubenswrapper[4556]: I0218 09:29:57.033935 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-dgmmz"] Feb 18 09:29:57 crc kubenswrapper[4556]: I0218 09:29:57.222024 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-8dvj9" Feb 18 09:29:57 crc kubenswrapper[4556]: I0218 09:29:57.261141 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-8dvj9" Feb 18 09:29:57 crc kubenswrapper[4556]: I0218 09:29:57.292550 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24dad0a1-29ed-49a7-be65-d024a02fcff4" path="/var/lib/kubelet/pods/24dad0a1-29ed-49a7-be65-d024a02fcff4/volumes" Feb 18 09:29:57 crc kubenswrapper[4556]: I0218 09:29:57.293345 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef39cb1a-74bf-48cf-8166-5cf97485bd88" path="/var/lib/kubelet/pods/ef39cb1a-74bf-48cf-8166-5cf97485bd88/volumes" Feb 18 09:29:57 crc kubenswrapper[4556]: I0218 09:29:57.468305 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8dvj9"] Feb 18 09:29:58 crc kubenswrapper[4556]: I0218 09:29:58.829044 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-8dvj9" podUID="e5f11486-6f3d-43ce-9805-e65bea6634bb" containerName="registry-server" containerID="cri-o://c7fab49f889539198845af7bf36eb77d6cd4831d9e151cf4908715a852292cdb" gracePeriod=2 Feb 18 09:29:59 crc kubenswrapper[4556]: I0218 09:29:59.257195 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8dvj9" Feb 18 09:29:59 crc kubenswrapper[4556]: I0218 09:29:59.399172 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5f11486-6f3d-43ce-9805-e65bea6634bb-catalog-content\") pod \"e5f11486-6f3d-43ce-9805-e65bea6634bb\" (UID: \"e5f11486-6f3d-43ce-9805-e65bea6634bb\") " Feb 18 09:29:59 crc kubenswrapper[4556]: I0218 09:29:59.399578 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5f11486-6f3d-43ce-9805-e65bea6634bb-utilities\") pod \"e5f11486-6f3d-43ce-9805-e65bea6634bb\" (UID: \"e5f11486-6f3d-43ce-9805-e65bea6634bb\") " Feb 18 09:29:59 crc kubenswrapper[4556]: I0218 09:29:59.399909 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7vz45\" (UniqueName: \"kubernetes.io/projected/e5f11486-6f3d-43ce-9805-e65bea6634bb-kube-api-access-7vz45\") pod \"e5f11486-6f3d-43ce-9805-e65bea6634bb\" (UID: \"e5f11486-6f3d-43ce-9805-e65bea6634bb\") " Feb 18 09:29:59 crc kubenswrapper[4556]: I0218 09:29:59.402466 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5f11486-6f3d-43ce-9805-e65bea6634bb-utilities" (OuterVolumeSpecName: "utilities") pod "e5f11486-6f3d-43ce-9805-e65bea6634bb" (UID: "e5f11486-6f3d-43ce-9805-e65bea6634bb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:29:59 crc kubenswrapper[4556]: I0218 09:29:59.408168 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5f11486-6f3d-43ce-9805-e65bea6634bb-kube-api-access-7vz45" (OuterVolumeSpecName: "kube-api-access-7vz45") pod "e5f11486-6f3d-43ce-9805-e65bea6634bb" (UID: "e5f11486-6f3d-43ce-9805-e65bea6634bb"). InnerVolumeSpecName "kube-api-access-7vz45". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:29:59 crc kubenswrapper[4556]: I0218 09:29:59.495617 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5f11486-6f3d-43ce-9805-e65bea6634bb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e5f11486-6f3d-43ce-9805-e65bea6634bb" (UID: "e5f11486-6f3d-43ce-9805-e65bea6634bb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:29:59 crc kubenswrapper[4556]: I0218 09:29:59.504170 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7vz45\" (UniqueName: \"kubernetes.io/projected/e5f11486-6f3d-43ce-9805-e65bea6634bb-kube-api-access-7vz45\") on node \"crc\" DevicePath \"\"" Feb 18 09:29:59 crc kubenswrapper[4556]: I0218 09:29:59.504217 4556 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5f11486-6f3d-43ce-9805-e65bea6634bb-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 09:29:59 crc kubenswrapper[4556]: I0218 09:29:59.504230 4556 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5f11486-6f3d-43ce-9805-e65bea6634bb-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 09:29:59 crc kubenswrapper[4556]: I0218 09:29:59.838555 4556 generic.go:334] "Generic (PLEG): container finished" podID="e5f11486-6f3d-43ce-9805-e65bea6634bb" containerID="c7fab49f889539198845af7bf36eb77d6cd4831d9e151cf4908715a852292cdb" exitCode=0 Feb 18 09:29:59 crc kubenswrapper[4556]: I0218 09:29:59.838606 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8dvj9" event={"ID":"e5f11486-6f3d-43ce-9805-e65bea6634bb","Type":"ContainerDied","Data":"c7fab49f889539198845af7bf36eb77d6cd4831d9e151cf4908715a852292cdb"} Feb 18 09:29:59 crc kubenswrapper[4556]: I0218 09:29:59.838630 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8dvj9" Feb 18 09:29:59 crc kubenswrapper[4556]: I0218 09:29:59.838655 4556 scope.go:117] "RemoveContainer" containerID="c7fab49f889539198845af7bf36eb77d6cd4831d9e151cf4908715a852292cdb" Feb 18 09:29:59 crc kubenswrapper[4556]: I0218 09:29:59.838641 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8dvj9" event={"ID":"e5f11486-6f3d-43ce-9805-e65bea6634bb","Type":"ContainerDied","Data":"299c58ff38fbb31ce185988180b51edb135b0af4e9a1575df39a93ac9082c9d4"} Feb 18 09:29:59 crc kubenswrapper[4556]: I0218 09:29:59.866377 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8dvj9"] Feb 18 09:29:59 crc kubenswrapper[4556]: I0218 09:29:59.866740 4556 scope.go:117] "RemoveContainer" containerID="4a7dbc86a2e040801298f30b41a97a254c0bc28d91af4498a273329371f922a1" Feb 18 09:29:59 crc kubenswrapper[4556]: I0218 09:29:59.884741 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-8dvj9"] Feb 18 09:29:59 crc kubenswrapper[4556]: I0218 09:29:59.892201 4556 scope.go:117] "RemoveContainer" containerID="09f5ab1825fa6eedb079cd43f10bec6ac49306fc63c24ea12d7172de69f13471" Feb 18 09:29:59 crc kubenswrapper[4556]: I0218 09:29:59.921662 4556 scope.go:117] "RemoveContainer" containerID="c7fab49f889539198845af7bf36eb77d6cd4831d9e151cf4908715a852292cdb" Feb 18 09:29:59 crc kubenswrapper[4556]: E0218 09:29:59.922578 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7fab49f889539198845af7bf36eb77d6cd4831d9e151cf4908715a852292cdb\": container with ID starting with c7fab49f889539198845af7bf36eb77d6cd4831d9e151cf4908715a852292cdb not found: ID does not exist" containerID="c7fab49f889539198845af7bf36eb77d6cd4831d9e151cf4908715a852292cdb" Feb 18 09:29:59 crc kubenswrapper[4556]: I0218 09:29:59.922644 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7fab49f889539198845af7bf36eb77d6cd4831d9e151cf4908715a852292cdb"} err="failed to get container status \"c7fab49f889539198845af7bf36eb77d6cd4831d9e151cf4908715a852292cdb\": rpc error: code = NotFound desc = could not find container \"c7fab49f889539198845af7bf36eb77d6cd4831d9e151cf4908715a852292cdb\": container with ID starting with c7fab49f889539198845af7bf36eb77d6cd4831d9e151cf4908715a852292cdb not found: ID does not exist" Feb 18 09:29:59 crc kubenswrapper[4556]: I0218 09:29:59.922692 4556 scope.go:117] "RemoveContainer" containerID="4a7dbc86a2e040801298f30b41a97a254c0bc28d91af4498a273329371f922a1" Feb 18 09:29:59 crc kubenswrapper[4556]: E0218 09:29:59.923134 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a7dbc86a2e040801298f30b41a97a254c0bc28d91af4498a273329371f922a1\": container with ID starting with 4a7dbc86a2e040801298f30b41a97a254c0bc28d91af4498a273329371f922a1 not found: ID does not exist" containerID="4a7dbc86a2e040801298f30b41a97a254c0bc28d91af4498a273329371f922a1" Feb 18 09:29:59 crc kubenswrapper[4556]: I0218 09:29:59.923222 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a7dbc86a2e040801298f30b41a97a254c0bc28d91af4498a273329371f922a1"} err="failed to get container status \"4a7dbc86a2e040801298f30b41a97a254c0bc28d91af4498a273329371f922a1\": rpc error: code = NotFound desc = could not find container \"4a7dbc86a2e040801298f30b41a97a254c0bc28d91af4498a273329371f922a1\": container with ID starting with 4a7dbc86a2e040801298f30b41a97a254c0bc28d91af4498a273329371f922a1 not found: ID does not exist" Feb 18 09:29:59 crc kubenswrapper[4556]: I0218 09:29:59.923259 4556 scope.go:117] "RemoveContainer" containerID="09f5ab1825fa6eedb079cd43f10bec6ac49306fc63c24ea12d7172de69f13471" Feb 18 09:29:59 crc kubenswrapper[4556]: E0218 09:29:59.923735 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09f5ab1825fa6eedb079cd43f10bec6ac49306fc63c24ea12d7172de69f13471\": container with ID starting with 09f5ab1825fa6eedb079cd43f10bec6ac49306fc63c24ea12d7172de69f13471 not found: ID does not exist" containerID="09f5ab1825fa6eedb079cd43f10bec6ac49306fc63c24ea12d7172de69f13471" Feb 18 09:29:59 crc kubenswrapper[4556]: I0218 09:29:59.923915 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09f5ab1825fa6eedb079cd43f10bec6ac49306fc63c24ea12d7172de69f13471"} err="failed to get container status \"09f5ab1825fa6eedb079cd43f10bec6ac49306fc63c24ea12d7172de69f13471\": rpc error: code = NotFound desc = could not find container \"09f5ab1825fa6eedb079cd43f10bec6ac49306fc63c24ea12d7172de69f13471\": container with ID starting with 09f5ab1825fa6eedb079cd43f10bec6ac49306fc63c24ea12d7172de69f13471 not found: ID does not exist" Feb 18 09:30:00 crc kubenswrapper[4556]: I0218 09:30:00.167267 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29523450-q8cqz"] Feb 18 09:30:00 crc kubenswrapper[4556]: E0218 09:30:00.168369 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5f11486-6f3d-43ce-9805-e65bea6634bb" containerName="extract-utilities" Feb 18 09:30:00 crc kubenswrapper[4556]: I0218 09:30:00.168391 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5f11486-6f3d-43ce-9805-e65bea6634bb" containerName="extract-utilities" Feb 18 09:30:00 crc kubenswrapper[4556]: E0218 09:30:00.168439 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5f11486-6f3d-43ce-9805-e65bea6634bb" containerName="registry-server" Feb 18 09:30:00 crc kubenswrapper[4556]: I0218 09:30:00.168450 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5f11486-6f3d-43ce-9805-e65bea6634bb" containerName="registry-server" Feb 18 09:30:00 crc kubenswrapper[4556]: E0218 09:30:00.168477 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5f11486-6f3d-43ce-9805-e65bea6634bb" containerName="extract-content" Feb 18 09:30:00 crc kubenswrapper[4556]: I0218 09:30:00.168485 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5f11486-6f3d-43ce-9805-e65bea6634bb" containerName="extract-content" Feb 18 09:30:00 crc kubenswrapper[4556]: I0218 09:30:00.168986 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5f11486-6f3d-43ce-9805-e65bea6634bb" containerName="registry-server" Feb 18 09:30:00 crc kubenswrapper[4556]: I0218 09:30:00.172411 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29523450-q8cqz" Feb 18 09:30:00 crc kubenswrapper[4556]: I0218 09:30:00.176010 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 18 09:30:00 crc kubenswrapper[4556]: I0218 09:30:00.176556 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 18 09:30:00 crc kubenswrapper[4556]: I0218 09:30:00.185885 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29523450-q8cqz"] Feb 18 09:30:00 crc kubenswrapper[4556]: I0218 09:30:00.323708 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/01705f70-cfff-4383-b99b-8b18e46add95-secret-volume\") pod \"collect-profiles-29523450-q8cqz\" (UID: \"01705f70-cfff-4383-b99b-8b18e46add95\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523450-q8cqz" Feb 18 09:30:00 crc kubenswrapper[4556]: I0218 09:30:00.323881 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47lqv\" (UniqueName: \"kubernetes.io/projected/01705f70-cfff-4383-b99b-8b18e46add95-kube-api-access-47lqv\") pod \"collect-profiles-29523450-q8cqz\" (UID: \"01705f70-cfff-4383-b99b-8b18e46add95\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523450-q8cqz" Feb 18 09:30:00 crc kubenswrapper[4556]: I0218 09:30:00.324893 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/01705f70-cfff-4383-b99b-8b18e46add95-config-volume\") pod \"collect-profiles-29523450-q8cqz\" (UID: \"01705f70-cfff-4383-b99b-8b18e46add95\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523450-q8cqz" Feb 18 09:30:00 crc kubenswrapper[4556]: I0218 09:30:00.426561 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/01705f70-cfff-4383-b99b-8b18e46add95-config-volume\") pod \"collect-profiles-29523450-q8cqz\" (UID: \"01705f70-cfff-4383-b99b-8b18e46add95\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523450-q8cqz" Feb 18 09:30:00 crc kubenswrapper[4556]: I0218 09:30:00.426708 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/01705f70-cfff-4383-b99b-8b18e46add95-secret-volume\") pod \"collect-profiles-29523450-q8cqz\" (UID: \"01705f70-cfff-4383-b99b-8b18e46add95\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523450-q8cqz" Feb 18 09:30:00 crc kubenswrapper[4556]: I0218 09:30:00.426783 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47lqv\" (UniqueName: \"kubernetes.io/projected/01705f70-cfff-4383-b99b-8b18e46add95-kube-api-access-47lqv\") pod \"collect-profiles-29523450-q8cqz\" (UID: \"01705f70-cfff-4383-b99b-8b18e46add95\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523450-q8cqz" Feb 18 09:30:00 crc kubenswrapper[4556]: I0218 09:30:00.427954 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/01705f70-cfff-4383-b99b-8b18e46add95-config-volume\") pod \"collect-profiles-29523450-q8cqz\" (UID: \"01705f70-cfff-4383-b99b-8b18e46add95\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523450-q8cqz" Feb 18 09:30:00 crc kubenswrapper[4556]: I0218 09:30:00.430385 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/01705f70-cfff-4383-b99b-8b18e46add95-secret-volume\") pod \"collect-profiles-29523450-q8cqz\" (UID: \"01705f70-cfff-4383-b99b-8b18e46add95\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523450-q8cqz" Feb 18 09:30:00 crc kubenswrapper[4556]: I0218 09:30:00.442340 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47lqv\" (UniqueName: \"kubernetes.io/projected/01705f70-cfff-4383-b99b-8b18e46add95-kube-api-access-47lqv\") pod \"collect-profiles-29523450-q8cqz\" (UID: \"01705f70-cfff-4383-b99b-8b18e46add95\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523450-q8cqz" Feb 18 09:30:00 crc kubenswrapper[4556]: I0218 09:30:00.492601 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29523450-q8cqz" Feb 18 09:30:00 crc kubenswrapper[4556]: I0218 09:30:00.922500 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29523450-q8cqz"] Feb 18 09:30:00 crc kubenswrapper[4556]: W0218 09:30:00.923958 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod01705f70_cfff_4383_b99b_8b18e46add95.slice/crio-dcf23d7a81790a7464add3ac8eb028610835bf2142c9e60f86643c7b5ec354f4 WatchSource:0}: Error finding container dcf23d7a81790a7464add3ac8eb028610835bf2142c9e60f86643c7b5ec354f4: Status 404 returned error can't find the container with id dcf23d7a81790a7464add3ac8eb028610835bf2142c9e60f86643c7b5ec354f4 Feb 18 09:30:01 crc kubenswrapper[4556]: I0218 09:30:01.294385 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5f11486-6f3d-43ce-9805-e65bea6634bb" path="/var/lib/kubelet/pods/e5f11486-6f3d-43ce-9805-e65bea6634bb/volumes" Feb 18 09:30:01 crc kubenswrapper[4556]: I0218 09:30:01.861439 4556 generic.go:334] "Generic (PLEG): container finished" podID="01705f70-cfff-4383-b99b-8b18e46add95" containerID="43a1f2f9f282545d045d528ba6a8052663230d2049e9983a459fef5091a0fd12" exitCode=0 Feb 18 09:30:01 crc kubenswrapper[4556]: I0218 09:30:01.861491 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29523450-q8cqz" event={"ID":"01705f70-cfff-4383-b99b-8b18e46add95","Type":"ContainerDied","Data":"43a1f2f9f282545d045d528ba6a8052663230d2049e9983a459fef5091a0fd12"} Feb 18 09:30:01 crc kubenswrapper[4556]: I0218 09:30:01.861549 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29523450-q8cqz" event={"ID":"01705f70-cfff-4383-b99b-8b18e46add95","Type":"ContainerStarted","Data":"dcf23d7a81790a7464add3ac8eb028610835bf2142c9e60f86643c7b5ec354f4"} Feb 18 09:30:02 crc kubenswrapper[4556]: I0218 09:30:02.875874 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jbdqg"] Feb 18 09:30:02 crc kubenswrapper[4556]: I0218 09:30:02.879182 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jbdqg" Feb 18 09:30:02 crc kubenswrapper[4556]: I0218 09:30:02.899603 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jbdqg"] Feb 18 09:30:02 crc kubenswrapper[4556]: I0218 09:30:02.989286 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15f8a4d8-f6b9-4a28-9751-4e7c1b377dd9-catalog-content\") pod \"redhat-marketplace-jbdqg\" (UID: \"15f8a4d8-f6b9-4a28-9751-4e7c1b377dd9\") " pod="openshift-marketplace/redhat-marketplace-jbdqg" Feb 18 09:30:02 crc kubenswrapper[4556]: I0218 09:30:02.989531 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15f8a4d8-f6b9-4a28-9751-4e7c1b377dd9-utilities\") pod \"redhat-marketplace-jbdqg\" (UID: \"15f8a4d8-f6b9-4a28-9751-4e7c1b377dd9\") " pod="openshift-marketplace/redhat-marketplace-jbdqg" Feb 18 09:30:02 crc kubenswrapper[4556]: I0218 09:30:02.989620 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ff5b6\" (UniqueName: \"kubernetes.io/projected/15f8a4d8-f6b9-4a28-9751-4e7c1b377dd9-kube-api-access-ff5b6\") pod \"redhat-marketplace-jbdqg\" (UID: \"15f8a4d8-f6b9-4a28-9751-4e7c1b377dd9\") " pod="openshift-marketplace/redhat-marketplace-jbdqg" Feb 18 09:30:03 crc kubenswrapper[4556]: I0218 09:30:03.091943 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15f8a4d8-f6b9-4a28-9751-4e7c1b377dd9-utilities\") pod \"redhat-marketplace-jbdqg\" (UID: \"15f8a4d8-f6b9-4a28-9751-4e7c1b377dd9\") " pod="openshift-marketplace/redhat-marketplace-jbdqg" Feb 18 09:30:03 crc kubenswrapper[4556]: I0218 09:30:03.092072 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ff5b6\" (UniqueName: \"kubernetes.io/projected/15f8a4d8-f6b9-4a28-9751-4e7c1b377dd9-kube-api-access-ff5b6\") pod \"redhat-marketplace-jbdqg\" (UID: \"15f8a4d8-f6b9-4a28-9751-4e7c1b377dd9\") " pod="openshift-marketplace/redhat-marketplace-jbdqg" Feb 18 09:30:03 crc kubenswrapper[4556]: I0218 09:30:03.092273 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15f8a4d8-f6b9-4a28-9751-4e7c1b377dd9-catalog-content\") pod \"redhat-marketplace-jbdqg\" (UID: \"15f8a4d8-f6b9-4a28-9751-4e7c1b377dd9\") " pod="openshift-marketplace/redhat-marketplace-jbdqg" Feb 18 09:30:03 crc kubenswrapper[4556]: I0218 09:30:03.092418 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15f8a4d8-f6b9-4a28-9751-4e7c1b377dd9-utilities\") pod \"redhat-marketplace-jbdqg\" (UID: \"15f8a4d8-f6b9-4a28-9751-4e7c1b377dd9\") " pod="openshift-marketplace/redhat-marketplace-jbdqg" Feb 18 09:30:03 crc kubenswrapper[4556]: I0218 09:30:03.092708 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15f8a4d8-f6b9-4a28-9751-4e7c1b377dd9-catalog-content\") pod \"redhat-marketplace-jbdqg\" (UID: \"15f8a4d8-f6b9-4a28-9751-4e7c1b377dd9\") " pod="openshift-marketplace/redhat-marketplace-jbdqg" Feb 18 09:30:03 crc kubenswrapper[4556]: I0218 09:30:03.109533 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ff5b6\" (UniqueName: \"kubernetes.io/projected/15f8a4d8-f6b9-4a28-9751-4e7c1b377dd9-kube-api-access-ff5b6\") pod \"redhat-marketplace-jbdqg\" (UID: \"15f8a4d8-f6b9-4a28-9751-4e7c1b377dd9\") " pod="openshift-marketplace/redhat-marketplace-jbdqg" Feb 18 09:30:03 crc kubenswrapper[4556]: I0218 09:30:03.176555 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29523450-q8cqz" Feb 18 09:30:03 crc kubenswrapper[4556]: I0218 09:30:03.207508 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jbdqg" Feb 18 09:30:03 crc kubenswrapper[4556]: I0218 09:30:03.295830 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/01705f70-cfff-4383-b99b-8b18e46add95-secret-volume\") pod \"01705f70-cfff-4383-b99b-8b18e46add95\" (UID: \"01705f70-cfff-4383-b99b-8b18e46add95\") " Feb 18 09:30:03 crc kubenswrapper[4556]: I0218 09:30:03.296086 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/01705f70-cfff-4383-b99b-8b18e46add95-config-volume\") pod \"01705f70-cfff-4383-b99b-8b18e46add95\" (UID: \"01705f70-cfff-4383-b99b-8b18e46add95\") " Feb 18 09:30:03 crc kubenswrapper[4556]: I0218 09:30:03.296295 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-47lqv\" (UniqueName: \"kubernetes.io/projected/01705f70-cfff-4383-b99b-8b18e46add95-kube-api-access-47lqv\") pod \"01705f70-cfff-4383-b99b-8b18e46add95\" (UID: \"01705f70-cfff-4383-b99b-8b18e46add95\") " Feb 18 09:30:03 crc kubenswrapper[4556]: I0218 09:30:03.301465 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01705f70-cfff-4383-b99b-8b18e46add95-config-volume" (OuterVolumeSpecName: "config-volume") pod "01705f70-cfff-4383-b99b-8b18e46add95" (UID: "01705f70-cfff-4383-b99b-8b18e46add95"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:30:03 crc kubenswrapper[4556]: I0218 09:30:03.302744 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01705f70-cfff-4383-b99b-8b18e46add95-kube-api-access-47lqv" (OuterVolumeSpecName: "kube-api-access-47lqv") pod "01705f70-cfff-4383-b99b-8b18e46add95" (UID: "01705f70-cfff-4383-b99b-8b18e46add95"). InnerVolumeSpecName "kube-api-access-47lqv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:30:03 crc kubenswrapper[4556]: I0218 09:30:03.307050 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01705f70-cfff-4383-b99b-8b18e46add95-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "01705f70-cfff-4383-b99b-8b18e46add95" (UID: "01705f70-cfff-4383-b99b-8b18e46add95"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:30:03 crc kubenswrapper[4556]: I0218 09:30:03.399385 4556 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/01705f70-cfff-4383-b99b-8b18e46add95-config-volume\") on node \"crc\" DevicePath \"\"" Feb 18 09:30:03 crc kubenswrapper[4556]: I0218 09:30:03.399422 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-47lqv\" (UniqueName: \"kubernetes.io/projected/01705f70-cfff-4383-b99b-8b18e46add95-kube-api-access-47lqv\") on node \"crc\" DevicePath \"\"" Feb 18 09:30:03 crc kubenswrapper[4556]: I0218 09:30:03.399433 4556 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/01705f70-cfff-4383-b99b-8b18e46add95-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 18 09:30:03 crc kubenswrapper[4556]: I0218 09:30:03.640765 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jbdqg"] Feb 18 09:30:03 crc kubenswrapper[4556]: W0218 09:30:03.645350 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod15f8a4d8_f6b9_4a28_9751_4e7c1b377dd9.slice/crio-bac262a509b229e5636433e25fcd3f49b007b95adafb63b937a423d93d928f99 WatchSource:0}: Error finding container bac262a509b229e5636433e25fcd3f49b007b95adafb63b937a423d93d928f99: Status 404 returned error can't find the container with id bac262a509b229e5636433e25fcd3f49b007b95adafb63b937a423d93d928f99 Feb 18 09:30:03 crc kubenswrapper[4556]: I0218 09:30:03.884667 4556 generic.go:334] "Generic (PLEG): container finished" podID="15f8a4d8-f6b9-4a28-9751-4e7c1b377dd9" containerID="699e9cb852329bc14c7e6aa99f1a2f0d76f3461e70786c665b0d7c3bb2a7f510" exitCode=0 Feb 18 09:30:03 crc kubenswrapper[4556]: I0218 09:30:03.884804 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jbdqg" event={"ID":"15f8a4d8-f6b9-4a28-9751-4e7c1b377dd9","Type":"ContainerDied","Data":"699e9cb852329bc14c7e6aa99f1a2f0d76f3461e70786c665b0d7c3bb2a7f510"} Feb 18 09:30:03 crc kubenswrapper[4556]: I0218 09:30:03.884855 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jbdqg" event={"ID":"15f8a4d8-f6b9-4a28-9751-4e7c1b377dd9","Type":"ContainerStarted","Data":"bac262a509b229e5636433e25fcd3f49b007b95adafb63b937a423d93d928f99"} Feb 18 09:30:03 crc kubenswrapper[4556]: I0218 09:30:03.887702 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29523450-q8cqz" event={"ID":"01705f70-cfff-4383-b99b-8b18e46add95","Type":"ContainerDied","Data":"dcf23d7a81790a7464add3ac8eb028610835bf2142c9e60f86643c7b5ec354f4"} Feb 18 09:30:03 crc kubenswrapper[4556]: I0218 09:30:03.887858 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dcf23d7a81790a7464add3ac8eb028610835bf2142c9e60f86643c7b5ec354f4" Feb 18 09:30:03 crc kubenswrapper[4556]: I0218 09:30:03.887761 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29523450-q8cqz" Feb 18 09:30:05 crc kubenswrapper[4556]: I0218 09:30:05.908684 4556 generic.go:334] "Generic (PLEG): container finished" podID="15f8a4d8-f6b9-4a28-9751-4e7c1b377dd9" containerID="db4c1e12ae7c48b71f848c38b1d05971552d979a35ab5611efa734534940646d" exitCode=0 Feb 18 09:30:05 crc kubenswrapper[4556]: I0218 09:30:05.908807 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jbdqg" event={"ID":"15f8a4d8-f6b9-4a28-9751-4e7c1b377dd9","Type":"ContainerDied","Data":"db4c1e12ae7c48b71f848c38b1d05971552d979a35ab5611efa734534940646d"} Feb 18 09:30:06 crc kubenswrapper[4556]: I0218 09:30:06.283384 4556 scope.go:117] "RemoveContainer" containerID="507a9134b6862bc8be1ac03b7993c890ef64f16c21975e00aa96c53fe413415d" Feb 18 09:30:06 crc kubenswrapper[4556]: E0218 09:30:06.284973 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:30:06 crc kubenswrapper[4556]: I0218 09:30:06.921772 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jbdqg" event={"ID":"15f8a4d8-f6b9-4a28-9751-4e7c1b377dd9","Type":"ContainerStarted","Data":"9840bbd44d07a8834529e51524bcdf31509ad453b3d405fd620314c2bc2fc5d1"} Feb 18 09:30:06 crc kubenswrapper[4556]: I0218 09:30:06.941669 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jbdqg" podStartSLOduration=2.437753814 podStartE2EDuration="4.941649872s" podCreationTimestamp="2026-02-18 09:30:02 +0000 UTC" firstStartedPulling="2026-02-18 09:30:03.887757398 +0000 UTC m=+1560.904718378" lastFinishedPulling="2026-02-18 09:30:06.391653456 +0000 UTC m=+1563.408614436" observedRunningTime="2026-02-18 09:30:06.938302888 +0000 UTC m=+1563.955263868" watchObservedRunningTime="2026-02-18 09:30:06.941649872 +0000 UTC m=+1563.958610852" Feb 18 09:30:10 crc kubenswrapper[4556]: I0218 09:30:10.325225 4556 scope.go:117] "RemoveContainer" containerID="f8c3715c17081b2315cdcf8d51278635a41a1190d1e5b321467b1cf3d6a471b6" Feb 18 09:30:10 crc kubenswrapper[4556]: I0218 09:30:10.354682 4556 scope.go:117] "RemoveContainer" containerID="eed458cdccc5b601f881f1ea893deb67f0c201b5b99dc68921127d9a44b49b65" Feb 18 09:30:10 crc kubenswrapper[4556]: I0218 09:30:10.404984 4556 scope.go:117] "RemoveContainer" containerID="fa0e4cd24d874dcbc0d61004056eba071ea491b925058c7e338ea6c08eba387d" Feb 18 09:30:10 crc kubenswrapper[4556]: I0218 09:30:10.439175 4556 scope.go:117] "RemoveContainer" containerID="69ace43bee98401345011fa1708b2b02040a330a77aef476145808875a5f4267" Feb 18 09:30:10 crc kubenswrapper[4556]: I0218 09:30:10.471902 4556 scope.go:117] "RemoveContainer" containerID="c94cc4c9927e86110534aaf1c83ff7cd2ee6e8e5117aa4499d7c535f0ad637de" Feb 18 09:30:10 crc kubenswrapper[4556]: I0218 09:30:10.519400 4556 scope.go:117] "RemoveContainer" containerID="d53f239a31c608a32853374e8528b40c92c47ff4879c4c4ffac72086bcf2cbc9" Feb 18 09:30:10 crc kubenswrapper[4556]: I0218 09:30:10.553310 4556 scope.go:117] "RemoveContainer" containerID="21545e68d19b79db777bb501d6ab6fae34c8d8b1e4b0d8e7049d0803a0b1d066" Feb 18 09:30:10 crc kubenswrapper[4556]: I0218 09:30:10.580836 4556 scope.go:117] "RemoveContainer" containerID="11a37ae750e25db51acdd71f8efb9605029d7f4853cb1d8aed87ad1132925aac" Feb 18 09:30:10 crc kubenswrapper[4556]: I0218 09:30:10.596501 4556 scope.go:117] "RemoveContainer" containerID="69e41aca003b0c33c159e3090331559d090f11ef66400ffc6495e1d73d0945a8" Feb 18 09:30:13 crc kubenswrapper[4556]: I0218 09:30:13.208791 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jbdqg" Feb 18 09:30:13 crc kubenswrapper[4556]: I0218 09:30:13.209220 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jbdqg" Feb 18 09:30:13 crc kubenswrapper[4556]: I0218 09:30:13.245090 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jbdqg" Feb 18 09:30:14 crc kubenswrapper[4556]: I0218 09:30:14.017601 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jbdqg" Feb 18 09:30:14 crc kubenswrapper[4556]: I0218 09:30:14.060508 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jbdqg"] Feb 18 09:30:14 crc kubenswrapper[4556]: I0218 09:30:14.994032 4556 generic.go:334] "Generic (PLEG): container finished" podID="d249a625-7b1c-4a2f-a92d-c18d7fbb6142" containerID="336874da07be51d4b7bb0a939adfd8f360a7e23c7a6def27e6b3897ad2a9a979" exitCode=0 Feb 18 09:30:14 crc kubenswrapper[4556]: I0218 09:30:14.994097 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9trr" event={"ID":"d249a625-7b1c-4a2f-a92d-c18d7fbb6142","Type":"ContainerDied","Data":"336874da07be51d4b7bb0a939adfd8f360a7e23c7a6def27e6b3897ad2a9a979"} Feb 18 09:30:16 crc kubenswrapper[4556]: I0218 09:30:16.002398 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jbdqg" podUID="15f8a4d8-f6b9-4a28-9751-4e7c1b377dd9" containerName="registry-server" containerID="cri-o://9840bbd44d07a8834529e51524bcdf31509ad453b3d405fd620314c2bc2fc5d1" gracePeriod=2 Feb 18 09:30:16 crc kubenswrapper[4556]: I0218 09:30:16.345687 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9trr" Feb 18 09:30:16 crc kubenswrapper[4556]: I0218 09:30:16.408439 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-neutron-metadata-combined-ca-bundle\") pod \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\" (UID: \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\") " Feb 18 09:30:16 crc kubenswrapper[4556]: I0218 09:30:16.408563 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-nova-combined-ca-bundle\") pod \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\" (UID: \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\") " Feb 18 09:30:16 crc kubenswrapper[4556]: I0218 09:30:16.408590 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-telemetry-combined-ca-bundle\") pod \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\" (UID: \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\") " Feb 18 09:30:16 crc kubenswrapper[4556]: I0218 09:30:16.408613 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-openstack-edpm-ipam-ovn-default-certs-0\") pod \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\" (UID: \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\") " Feb 18 09:30:16 crc kubenswrapper[4556]: I0218 09:30:16.408758 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\" (UID: \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\") " Feb 18 09:30:16 crc kubenswrapper[4556]: I0218 09:30:16.408912 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-ssh-key-openstack-edpm-ipam\") pod \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\" (UID: \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\") " Feb 18 09:30:16 crc kubenswrapper[4556]: I0218 09:30:16.408940 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-libvirt-combined-ca-bundle\") pod \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\" (UID: \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\") " Feb 18 09:30:16 crc kubenswrapper[4556]: I0218 09:30:16.408966 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\" (UID: \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\") " Feb 18 09:30:16 crc kubenswrapper[4556]: I0218 09:30:16.408983 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-ovn-combined-ca-bundle\") pod \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\" (UID: \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\") " Feb 18 09:30:16 crc kubenswrapper[4556]: I0218 09:30:16.409022 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-inventory\") pod \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\" (UID: \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\") " Feb 18 09:30:16 crc kubenswrapper[4556]: I0218 09:30:16.409043 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-bootstrap-combined-ca-bundle\") pod \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\" (UID: \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\") " Feb 18 09:30:16 crc kubenswrapper[4556]: I0218 09:30:16.409065 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-repo-setup-combined-ca-bundle\") pod \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\" (UID: \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\") " Feb 18 09:30:16 crc kubenswrapper[4556]: I0218 09:30:16.409101 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-prm9w\" (UniqueName: \"kubernetes.io/projected/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-kube-api-access-prm9w\") pod \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\" (UID: \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\") " Feb 18 09:30:16 crc kubenswrapper[4556]: I0218 09:30:16.409144 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\" (UID: \"d249a625-7b1c-4a2f-a92d-c18d7fbb6142\") " Feb 18 09:30:16 crc kubenswrapper[4556]: I0218 09:30:16.414748 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "d249a625-7b1c-4a2f-a92d-c18d7fbb6142" (UID: "d249a625-7b1c-4a2f-a92d-c18d7fbb6142"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:30:16 crc kubenswrapper[4556]: I0218 09:30:16.414783 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "d249a625-7b1c-4a2f-a92d-c18d7fbb6142" (UID: "d249a625-7b1c-4a2f-a92d-c18d7fbb6142"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:30:16 crc kubenswrapper[4556]: I0218 09:30:16.415252 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "d249a625-7b1c-4a2f-a92d-c18d7fbb6142" (UID: "d249a625-7b1c-4a2f-a92d-c18d7fbb6142"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:30:16 crc kubenswrapper[4556]: I0218 09:30:16.415340 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "d249a625-7b1c-4a2f-a92d-c18d7fbb6142" (UID: "d249a625-7b1c-4a2f-a92d-c18d7fbb6142"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:30:16 crc kubenswrapper[4556]: I0218 09:30:16.415597 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "d249a625-7b1c-4a2f-a92d-c18d7fbb6142" (UID: "d249a625-7b1c-4a2f-a92d-c18d7fbb6142"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:30:16 crc kubenswrapper[4556]: I0218 09:30:16.415639 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "d249a625-7b1c-4a2f-a92d-c18d7fbb6142" (UID: "d249a625-7b1c-4a2f-a92d-c18d7fbb6142"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:30:16 crc kubenswrapper[4556]: I0218 09:30:16.416207 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "d249a625-7b1c-4a2f-a92d-c18d7fbb6142" (UID: "d249a625-7b1c-4a2f-a92d-c18d7fbb6142"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:30:16 crc kubenswrapper[4556]: I0218 09:30:16.416534 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "d249a625-7b1c-4a2f-a92d-c18d7fbb6142" (UID: "d249a625-7b1c-4a2f-a92d-c18d7fbb6142"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:30:16 crc kubenswrapper[4556]: I0218 09:30:16.416586 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "d249a625-7b1c-4a2f-a92d-c18d7fbb6142" (UID: "d249a625-7b1c-4a2f-a92d-c18d7fbb6142"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:30:16 crc kubenswrapper[4556]: I0218 09:30:16.416926 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "d249a625-7b1c-4a2f-a92d-c18d7fbb6142" (UID: "d249a625-7b1c-4a2f-a92d-c18d7fbb6142"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:30:16 crc kubenswrapper[4556]: I0218 09:30:16.417551 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-kube-api-access-prm9w" (OuterVolumeSpecName: "kube-api-access-prm9w") pod "d249a625-7b1c-4a2f-a92d-c18d7fbb6142" (UID: "d249a625-7b1c-4a2f-a92d-c18d7fbb6142"). InnerVolumeSpecName "kube-api-access-prm9w". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:30:16 crc kubenswrapper[4556]: I0218 09:30:16.417965 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "d249a625-7b1c-4a2f-a92d-c18d7fbb6142" (UID: "d249a625-7b1c-4a2f-a92d-c18d7fbb6142"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:30:16 crc kubenswrapper[4556]: I0218 09:30:16.432459 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jbdqg" Feb 18 09:30:16 crc kubenswrapper[4556]: I0218 09:30:16.434302 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "d249a625-7b1c-4a2f-a92d-c18d7fbb6142" (UID: "d249a625-7b1c-4a2f-a92d-c18d7fbb6142"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:30:16 crc kubenswrapper[4556]: I0218 09:30:16.436243 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-inventory" (OuterVolumeSpecName: "inventory") pod "d249a625-7b1c-4a2f-a92d-c18d7fbb6142" (UID: "d249a625-7b1c-4a2f-a92d-c18d7fbb6142"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:30:16 crc kubenswrapper[4556]: I0218 09:30:16.510776 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15f8a4d8-f6b9-4a28-9751-4e7c1b377dd9-catalog-content\") pod \"15f8a4d8-f6b9-4a28-9751-4e7c1b377dd9\" (UID: \"15f8a4d8-f6b9-4a28-9751-4e7c1b377dd9\") " Feb 18 09:30:16 crc kubenswrapper[4556]: I0218 09:30:16.510853 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15f8a4d8-f6b9-4a28-9751-4e7c1b377dd9-utilities\") pod \"15f8a4d8-f6b9-4a28-9751-4e7c1b377dd9\" (UID: \"15f8a4d8-f6b9-4a28-9751-4e7c1b377dd9\") " Feb 18 09:30:16 crc kubenswrapper[4556]: I0218 09:30:16.511049 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ff5b6\" (UniqueName: \"kubernetes.io/projected/15f8a4d8-f6b9-4a28-9751-4e7c1b377dd9-kube-api-access-ff5b6\") pod \"15f8a4d8-f6b9-4a28-9751-4e7c1b377dd9\" (UID: \"15f8a4d8-f6b9-4a28-9751-4e7c1b377dd9\") " Feb 18 09:30:16 crc kubenswrapper[4556]: I0218 09:30:16.511541 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15f8a4d8-f6b9-4a28-9751-4e7c1b377dd9-utilities" (OuterVolumeSpecName: "utilities") pod "15f8a4d8-f6b9-4a28-9751-4e7c1b377dd9" (UID: "15f8a4d8-f6b9-4a28-9751-4e7c1b377dd9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:30:16 crc kubenswrapper[4556]: I0218 09:30:16.512314 4556 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 18 09:30:16 crc kubenswrapper[4556]: I0218 09:30:16.512341 4556 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15f8a4d8-f6b9-4a28-9751-4e7c1b377dd9-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 09:30:16 crc kubenswrapper[4556]: I0218 09:30:16.512357 4556 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 18 09:30:16 crc kubenswrapper[4556]: I0218 09:30:16.512372 4556 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:30:16 crc kubenswrapper[4556]: I0218 09:30:16.512386 4556 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 18 09:30:16 crc kubenswrapper[4556]: I0218 09:30:16.512401 4556 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:30:16 crc kubenswrapper[4556]: I0218 09:30:16.512414 4556 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-inventory\") on node \"crc\" DevicePath \"\"" Feb 18 09:30:16 crc kubenswrapper[4556]: I0218 09:30:16.512427 4556 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:30:16 crc kubenswrapper[4556]: I0218 09:30:16.512439 4556 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:30:16 crc kubenswrapper[4556]: I0218 09:30:16.512452 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-prm9w\" (UniqueName: \"kubernetes.io/projected/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-kube-api-access-prm9w\") on node \"crc\" DevicePath \"\"" Feb 18 09:30:16 crc kubenswrapper[4556]: I0218 09:30:16.512465 4556 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 18 09:30:16 crc kubenswrapper[4556]: I0218 09:30:16.512477 4556 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:30:16 crc kubenswrapper[4556]: I0218 09:30:16.512489 4556 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:30:16 crc kubenswrapper[4556]: I0218 09:30:16.512499 4556 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:30:16 crc kubenswrapper[4556]: I0218 09:30:16.512508 4556 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d249a625-7b1c-4a2f-a92d-c18d7fbb6142-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 18 09:30:16 crc kubenswrapper[4556]: I0218 09:30:16.514669 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15f8a4d8-f6b9-4a28-9751-4e7c1b377dd9-kube-api-access-ff5b6" (OuterVolumeSpecName: "kube-api-access-ff5b6") pod "15f8a4d8-f6b9-4a28-9751-4e7c1b377dd9" (UID: "15f8a4d8-f6b9-4a28-9751-4e7c1b377dd9"). InnerVolumeSpecName "kube-api-access-ff5b6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:30:16 crc kubenswrapper[4556]: I0218 09:30:16.529865 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15f8a4d8-f6b9-4a28-9751-4e7c1b377dd9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "15f8a4d8-f6b9-4a28-9751-4e7c1b377dd9" (UID: "15f8a4d8-f6b9-4a28-9751-4e7c1b377dd9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:30:16 crc kubenswrapper[4556]: I0218 09:30:16.614309 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ff5b6\" (UniqueName: \"kubernetes.io/projected/15f8a4d8-f6b9-4a28-9751-4e7c1b377dd9-kube-api-access-ff5b6\") on node \"crc\" DevicePath \"\"" Feb 18 09:30:16 crc kubenswrapper[4556]: I0218 09:30:16.614345 4556 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15f8a4d8-f6b9-4a28-9751-4e7c1b377dd9-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 09:30:17 crc kubenswrapper[4556]: I0218 09:30:17.014139 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9trr" Feb 18 09:30:17 crc kubenswrapper[4556]: I0218 09:30:17.014123 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9trr" event={"ID":"d249a625-7b1c-4a2f-a92d-c18d7fbb6142","Type":"ContainerDied","Data":"2cb94cb3ce78a2552320a2566f80b218d4b9d2b03e8028253ed3463ce10490b4"} Feb 18 09:30:17 crc kubenswrapper[4556]: I0218 09:30:17.014950 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2cb94cb3ce78a2552320a2566f80b218d4b9d2b03e8028253ed3463ce10490b4" Feb 18 09:30:17 crc kubenswrapper[4556]: I0218 09:30:17.017312 4556 generic.go:334] "Generic (PLEG): container finished" podID="15f8a4d8-f6b9-4a28-9751-4e7c1b377dd9" containerID="9840bbd44d07a8834529e51524bcdf31509ad453b3d405fd620314c2bc2fc5d1" exitCode=0 Feb 18 09:30:17 crc kubenswrapper[4556]: I0218 09:30:17.017368 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jbdqg" event={"ID":"15f8a4d8-f6b9-4a28-9751-4e7c1b377dd9","Type":"ContainerDied","Data":"9840bbd44d07a8834529e51524bcdf31509ad453b3d405fd620314c2bc2fc5d1"} Feb 18 09:30:17 crc kubenswrapper[4556]: I0218 09:30:17.017411 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jbdqg" Feb 18 09:30:17 crc kubenswrapper[4556]: I0218 09:30:17.017444 4556 scope.go:117] "RemoveContainer" containerID="9840bbd44d07a8834529e51524bcdf31509ad453b3d405fd620314c2bc2fc5d1" Feb 18 09:30:17 crc kubenswrapper[4556]: I0218 09:30:17.017418 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jbdqg" event={"ID":"15f8a4d8-f6b9-4a28-9751-4e7c1b377dd9","Type":"ContainerDied","Data":"bac262a509b229e5636433e25fcd3f49b007b95adafb63b937a423d93d928f99"} Feb 18 09:30:17 crc kubenswrapper[4556]: I0218 09:30:17.042499 4556 scope.go:117] "RemoveContainer" containerID="db4c1e12ae7c48b71f848c38b1d05971552d979a35ab5611efa734534940646d" Feb 18 09:30:17 crc kubenswrapper[4556]: I0218 09:30:17.069805 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jbdqg"] Feb 18 09:30:17 crc kubenswrapper[4556]: I0218 09:30:17.070269 4556 scope.go:117] "RemoveContainer" containerID="699e9cb852329bc14c7e6aa99f1a2f0d76f3461e70786c665b0d7c3bb2a7f510" Feb 18 09:30:17 crc kubenswrapper[4556]: I0218 09:30:17.075682 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jbdqg"] Feb 18 09:30:17 crc kubenswrapper[4556]: I0218 09:30:17.084548 4556 scope.go:117] "RemoveContainer" containerID="9840bbd44d07a8834529e51524bcdf31509ad453b3d405fd620314c2bc2fc5d1" Feb 18 09:30:17 crc kubenswrapper[4556]: E0218 09:30:17.084829 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9840bbd44d07a8834529e51524bcdf31509ad453b3d405fd620314c2bc2fc5d1\": container with ID starting with 9840bbd44d07a8834529e51524bcdf31509ad453b3d405fd620314c2bc2fc5d1 not found: ID does not exist" containerID="9840bbd44d07a8834529e51524bcdf31509ad453b3d405fd620314c2bc2fc5d1" Feb 18 09:30:17 crc kubenswrapper[4556]: I0218 09:30:17.084860 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9840bbd44d07a8834529e51524bcdf31509ad453b3d405fd620314c2bc2fc5d1"} err="failed to get container status \"9840bbd44d07a8834529e51524bcdf31509ad453b3d405fd620314c2bc2fc5d1\": rpc error: code = NotFound desc = could not find container \"9840bbd44d07a8834529e51524bcdf31509ad453b3d405fd620314c2bc2fc5d1\": container with ID starting with 9840bbd44d07a8834529e51524bcdf31509ad453b3d405fd620314c2bc2fc5d1 not found: ID does not exist" Feb 18 09:30:17 crc kubenswrapper[4556]: I0218 09:30:17.084883 4556 scope.go:117] "RemoveContainer" containerID="db4c1e12ae7c48b71f848c38b1d05971552d979a35ab5611efa734534940646d" Feb 18 09:30:17 crc kubenswrapper[4556]: E0218 09:30:17.085100 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db4c1e12ae7c48b71f848c38b1d05971552d979a35ab5611efa734534940646d\": container with ID starting with db4c1e12ae7c48b71f848c38b1d05971552d979a35ab5611efa734534940646d not found: ID does not exist" containerID="db4c1e12ae7c48b71f848c38b1d05971552d979a35ab5611efa734534940646d" Feb 18 09:30:17 crc kubenswrapper[4556]: I0218 09:30:17.085126 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db4c1e12ae7c48b71f848c38b1d05971552d979a35ab5611efa734534940646d"} err="failed to get container status \"db4c1e12ae7c48b71f848c38b1d05971552d979a35ab5611efa734534940646d\": rpc error: code = NotFound desc = could not find container \"db4c1e12ae7c48b71f848c38b1d05971552d979a35ab5611efa734534940646d\": container with ID starting with db4c1e12ae7c48b71f848c38b1d05971552d979a35ab5611efa734534940646d not found: ID does not exist" Feb 18 09:30:17 crc kubenswrapper[4556]: I0218 09:30:17.085147 4556 scope.go:117] "RemoveContainer" containerID="699e9cb852329bc14c7e6aa99f1a2f0d76f3461e70786c665b0d7c3bb2a7f510" Feb 18 09:30:17 crc kubenswrapper[4556]: E0218 09:30:17.085352 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"699e9cb852329bc14c7e6aa99f1a2f0d76f3461e70786c665b0d7c3bb2a7f510\": container with ID starting with 699e9cb852329bc14c7e6aa99f1a2f0d76f3461e70786c665b0d7c3bb2a7f510 not found: ID does not exist" containerID="699e9cb852329bc14c7e6aa99f1a2f0d76f3461e70786c665b0d7c3bb2a7f510" Feb 18 09:30:17 crc kubenswrapper[4556]: I0218 09:30:17.085377 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"699e9cb852329bc14c7e6aa99f1a2f0d76f3461e70786c665b0d7c3bb2a7f510"} err="failed to get container status \"699e9cb852329bc14c7e6aa99f1a2f0d76f3461e70786c665b0d7c3bb2a7f510\": rpc error: code = NotFound desc = could not find container \"699e9cb852329bc14c7e6aa99f1a2f0d76f3461e70786c665b0d7c3bb2a7f510\": container with ID starting with 699e9cb852329bc14c7e6aa99f1a2f0d76f3461e70786c665b0d7c3bb2a7f510 not found: ID does not exist" Feb 18 09:30:17 crc kubenswrapper[4556]: I0218 09:30:17.134850 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-vgw62"] Feb 18 09:30:17 crc kubenswrapper[4556]: E0218 09:30:17.135212 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15f8a4d8-f6b9-4a28-9751-4e7c1b377dd9" containerName="registry-server" Feb 18 09:30:17 crc kubenswrapper[4556]: I0218 09:30:17.135233 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="15f8a4d8-f6b9-4a28-9751-4e7c1b377dd9" containerName="registry-server" Feb 18 09:30:17 crc kubenswrapper[4556]: E0218 09:30:17.135243 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15f8a4d8-f6b9-4a28-9751-4e7c1b377dd9" containerName="extract-content" Feb 18 09:30:17 crc kubenswrapper[4556]: I0218 09:30:17.135250 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="15f8a4d8-f6b9-4a28-9751-4e7c1b377dd9" containerName="extract-content" Feb 18 09:30:17 crc kubenswrapper[4556]: E0218 09:30:17.135275 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15f8a4d8-f6b9-4a28-9751-4e7c1b377dd9" containerName="extract-utilities" Feb 18 09:30:17 crc kubenswrapper[4556]: I0218 09:30:17.135281 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="15f8a4d8-f6b9-4a28-9751-4e7c1b377dd9" containerName="extract-utilities" Feb 18 09:30:17 crc kubenswrapper[4556]: E0218 09:30:17.135288 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d249a625-7b1c-4a2f-a92d-c18d7fbb6142" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Feb 18 09:30:17 crc kubenswrapper[4556]: I0218 09:30:17.135294 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="d249a625-7b1c-4a2f-a92d-c18d7fbb6142" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Feb 18 09:30:17 crc kubenswrapper[4556]: E0218 09:30:17.135307 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01705f70-cfff-4383-b99b-8b18e46add95" containerName="collect-profiles" Feb 18 09:30:17 crc kubenswrapper[4556]: I0218 09:30:17.135313 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="01705f70-cfff-4383-b99b-8b18e46add95" containerName="collect-profiles" Feb 18 09:30:17 crc kubenswrapper[4556]: I0218 09:30:17.135490 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="d249a625-7b1c-4a2f-a92d-c18d7fbb6142" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Feb 18 09:30:17 crc kubenswrapper[4556]: I0218 09:30:17.135509 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="01705f70-cfff-4383-b99b-8b18e46add95" containerName="collect-profiles" Feb 18 09:30:17 crc kubenswrapper[4556]: I0218 09:30:17.135518 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="15f8a4d8-f6b9-4a28-9751-4e7c1b377dd9" containerName="registry-server" Feb 18 09:30:17 crc kubenswrapper[4556]: I0218 09:30:17.136036 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vgw62" Feb 18 09:30:17 crc kubenswrapper[4556]: I0218 09:30:17.137651 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 18 09:30:17 crc kubenswrapper[4556]: I0218 09:30:17.138018 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 18 09:30:17 crc kubenswrapper[4556]: I0218 09:30:17.138396 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 18 09:30:17 crc kubenswrapper[4556]: I0218 09:30:17.138917 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-54z7h" Feb 18 09:30:17 crc kubenswrapper[4556]: I0218 09:30:17.139779 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Feb 18 09:30:17 crc kubenswrapper[4556]: I0218 09:30:17.145304 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-vgw62"] Feb 18 09:30:17 crc kubenswrapper[4556]: I0218 09:30:17.227540 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c63f939-1090-4fae-b45b-9d55f09f489d-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-vgw62\" (UID: \"7c63f939-1090-4fae-b45b-9d55f09f489d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vgw62" Feb 18 09:30:17 crc kubenswrapper[4556]: I0218 09:30:17.227655 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndmcv\" (UniqueName: \"kubernetes.io/projected/7c63f939-1090-4fae-b45b-9d55f09f489d-kube-api-access-ndmcv\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-vgw62\" (UID: \"7c63f939-1090-4fae-b45b-9d55f09f489d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vgw62" Feb 18 09:30:17 crc kubenswrapper[4556]: I0218 09:30:17.227775 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7c63f939-1090-4fae-b45b-9d55f09f489d-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-vgw62\" (UID: \"7c63f939-1090-4fae-b45b-9d55f09f489d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vgw62" Feb 18 09:30:17 crc kubenswrapper[4556]: I0218 09:30:17.227901 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7c63f939-1090-4fae-b45b-9d55f09f489d-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-vgw62\" (UID: \"7c63f939-1090-4fae-b45b-9d55f09f489d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vgw62" Feb 18 09:30:17 crc kubenswrapper[4556]: I0218 09:30:17.227953 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/7c63f939-1090-4fae-b45b-9d55f09f489d-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-vgw62\" (UID: \"7c63f939-1090-4fae-b45b-9d55f09f489d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vgw62" Feb 18 09:30:17 crc kubenswrapper[4556]: I0218 09:30:17.291828 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15f8a4d8-f6b9-4a28-9751-4e7c1b377dd9" path="/var/lib/kubelet/pods/15f8a4d8-f6b9-4a28-9751-4e7c1b377dd9/volumes" Feb 18 09:30:17 crc kubenswrapper[4556]: I0218 09:30:17.329862 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7c63f939-1090-4fae-b45b-9d55f09f489d-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-vgw62\" (UID: \"7c63f939-1090-4fae-b45b-9d55f09f489d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vgw62" Feb 18 09:30:17 crc kubenswrapper[4556]: I0218 09:30:17.329986 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7c63f939-1090-4fae-b45b-9d55f09f489d-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-vgw62\" (UID: \"7c63f939-1090-4fae-b45b-9d55f09f489d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vgw62" Feb 18 09:30:17 crc kubenswrapper[4556]: I0218 09:30:17.330026 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/7c63f939-1090-4fae-b45b-9d55f09f489d-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-vgw62\" (UID: \"7c63f939-1090-4fae-b45b-9d55f09f489d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vgw62" Feb 18 09:30:17 crc kubenswrapper[4556]: I0218 09:30:17.330198 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c63f939-1090-4fae-b45b-9d55f09f489d-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-vgw62\" (UID: \"7c63f939-1090-4fae-b45b-9d55f09f489d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vgw62" Feb 18 09:30:17 crc kubenswrapper[4556]: I0218 09:30:17.330257 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndmcv\" (UniqueName: \"kubernetes.io/projected/7c63f939-1090-4fae-b45b-9d55f09f489d-kube-api-access-ndmcv\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-vgw62\" (UID: \"7c63f939-1090-4fae-b45b-9d55f09f489d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vgw62" Feb 18 09:30:17 crc kubenswrapper[4556]: I0218 09:30:17.332427 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/7c63f939-1090-4fae-b45b-9d55f09f489d-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-vgw62\" (UID: \"7c63f939-1090-4fae-b45b-9d55f09f489d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vgw62" Feb 18 09:30:17 crc kubenswrapper[4556]: I0218 09:30:17.350046 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7c63f939-1090-4fae-b45b-9d55f09f489d-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-vgw62\" (UID: \"7c63f939-1090-4fae-b45b-9d55f09f489d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vgw62" Feb 18 09:30:17 crc kubenswrapper[4556]: I0218 09:30:17.350086 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7c63f939-1090-4fae-b45b-9d55f09f489d-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-vgw62\" (UID: \"7c63f939-1090-4fae-b45b-9d55f09f489d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vgw62" Feb 18 09:30:17 crc kubenswrapper[4556]: I0218 09:30:17.354834 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c63f939-1090-4fae-b45b-9d55f09f489d-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-vgw62\" (UID: \"7c63f939-1090-4fae-b45b-9d55f09f489d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vgw62" Feb 18 09:30:17 crc kubenswrapper[4556]: I0218 09:30:17.356484 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndmcv\" (UniqueName: \"kubernetes.io/projected/7c63f939-1090-4fae-b45b-9d55f09f489d-kube-api-access-ndmcv\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-vgw62\" (UID: \"7c63f939-1090-4fae-b45b-9d55f09f489d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vgw62" Feb 18 09:30:17 crc kubenswrapper[4556]: I0218 09:30:17.478447 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vgw62" Feb 18 09:30:17 crc kubenswrapper[4556]: I0218 09:30:17.942950 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-vgw62"] Feb 18 09:30:18 crc kubenswrapper[4556]: I0218 09:30:18.025398 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vgw62" event={"ID":"7c63f939-1090-4fae-b45b-9d55f09f489d","Type":"ContainerStarted","Data":"98beddc14af2cc3504ec58406030a06f887e27669b65f3d7875fbf034d3a631e"} Feb 18 09:30:19 crc kubenswrapper[4556]: I0218 09:30:19.039326 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vgw62" event={"ID":"7c63f939-1090-4fae-b45b-9d55f09f489d","Type":"ContainerStarted","Data":"62a21056e25119741d39d0f09efa1b5c3fc81a6c2fd19015180907b7c3c76439"} Feb 18 09:30:19 crc kubenswrapper[4556]: I0218 09:30:19.057949 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vgw62" podStartSLOduration=1.507427151 podStartE2EDuration="2.057933338s" podCreationTimestamp="2026-02-18 09:30:17 +0000 UTC" firstStartedPulling="2026-02-18 09:30:17.946247762 +0000 UTC m=+1574.963208742" lastFinishedPulling="2026-02-18 09:30:18.496753949 +0000 UTC m=+1575.513714929" observedRunningTime="2026-02-18 09:30:19.054999274 +0000 UTC m=+1576.071960254" watchObservedRunningTime="2026-02-18 09:30:19.057933338 +0000 UTC m=+1576.074894319" Feb 18 09:30:21 crc kubenswrapper[4556]: I0218 09:30:21.282965 4556 scope.go:117] "RemoveContainer" containerID="507a9134b6862bc8be1ac03b7993c890ef64f16c21975e00aa96c53fe413415d" Feb 18 09:30:21 crc kubenswrapper[4556]: E0218 09:30:21.283605 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:30:32 crc kubenswrapper[4556]: I0218 09:30:32.282986 4556 scope.go:117] "RemoveContainer" containerID="507a9134b6862bc8be1ac03b7993c890ef64f16c21975e00aa96c53fe413415d" Feb 18 09:30:32 crc kubenswrapper[4556]: E0218 09:30:32.283842 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:30:42 crc kubenswrapper[4556]: I0218 09:30:42.031209 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-x62wg"] Feb 18 09:30:42 crc kubenswrapper[4556]: I0218 09:30:42.038644 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-x62wg"] Feb 18 09:30:43 crc kubenswrapper[4556]: I0218 09:30:43.239232 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nrbln"] Feb 18 09:30:43 crc kubenswrapper[4556]: I0218 09:30:43.241389 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nrbln" Feb 18 09:30:43 crc kubenswrapper[4556]: I0218 09:30:43.250048 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nrbln"] Feb 18 09:30:43 crc kubenswrapper[4556]: I0218 09:30:43.287277 4556 scope.go:117] "RemoveContainer" containerID="507a9134b6862bc8be1ac03b7993c890ef64f16c21975e00aa96c53fe413415d" Feb 18 09:30:43 crc kubenswrapper[4556]: E0218 09:30:43.287509 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:30:43 crc kubenswrapper[4556]: I0218 09:30:43.291963 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0c20df4-63b1-46f0-9aca-d533c8ed6b8c" path="/var/lib/kubelet/pods/d0c20df4-63b1-46f0-9aca-d533c8ed6b8c/volumes" Feb 18 09:30:43 crc kubenswrapper[4556]: I0218 09:30:43.383661 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba68071e-eb60-4707-9ddf-7fe7072e9c25-utilities\") pod \"community-operators-nrbln\" (UID: \"ba68071e-eb60-4707-9ddf-7fe7072e9c25\") " pod="openshift-marketplace/community-operators-nrbln" Feb 18 09:30:43 crc kubenswrapper[4556]: I0218 09:30:43.383822 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmv77\" (UniqueName: \"kubernetes.io/projected/ba68071e-eb60-4707-9ddf-7fe7072e9c25-kube-api-access-jmv77\") pod \"community-operators-nrbln\" (UID: \"ba68071e-eb60-4707-9ddf-7fe7072e9c25\") " pod="openshift-marketplace/community-operators-nrbln" Feb 18 09:30:43 crc kubenswrapper[4556]: I0218 09:30:43.383907 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba68071e-eb60-4707-9ddf-7fe7072e9c25-catalog-content\") pod \"community-operators-nrbln\" (UID: \"ba68071e-eb60-4707-9ddf-7fe7072e9c25\") " pod="openshift-marketplace/community-operators-nrbln" Feb 18 09:30:43 crc kubenswrapper[4556]: I0218 09:30:43.486193 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba68071e-eb60-4707-9ddf-7fe7072e9c25-utilities\") pod \"community-operators-nrbln\" (UID: \"ba68071e-eb60-4707-9ddf-7fe7072e9c25\") " pod="openshift-marketplace/community-operators-nrbln" Feb 18 09:30:43 crc kubenswrapper[4556]: I0218 09:30:43.486351 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmv77\" (UniqueName: \"kubernetes.io/projected/ba68071e-eb60-4707-9ddf-7fe7072e9c25-kube-api-access-jmv77\") pod \"community-operators-nrbln\" (UID: \"ba68071e-eb60-4707-9ddf-7fe7072e9c25\") " pod="openshift-marketplace/community-operators-nrbln" Feb 18 09:30:43 crc kubenswrapper[4556]: I0218 09:30:43.486444 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba68071e-eb60-4707-9ddf-7fe7072e9c25-catalog-content\") pod \"community-operators-nrbln\" (UID: \"ba68071e-eb60-4707-9ddf-7fe7072e9c25\") " pod="openshift-marketplace/community-operators-nrbln" Feb 18 09:30:43 crc kubenswrapper[4556]: I0218 09:30:43.487253 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba68071e-eb60-4707-9ddf-7fe7072e9c25-catalog-content\") pod \"community-operators-nrbln\" (UID: \"ba68071e-eb60-4707-9ddf-7fe7072e9c25\") " pod="openshift-marketplace/community-operators-nrbln" Feb 18 09:30:43 crc kubenswrapper[4556]: I0218 09:30:43.487425 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba68071e-eb60-4707-9ddf-7fe7072e9c25-utilities\") pod \"community-operators-nrbln\" (UID: \"ba68071e-eb60-4707-9ddf-7fe7072e9c25\") " pod="openshift-marketplace/community-operators-nrbln" Feb 18 09:30:43 crc kubenswrapper[4556]: I0218 09:30:43.508928 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmv77\" (UniqueName: \"kubernetes.io/projected/ba68071e-eb60-4707-9ddf-7fe7072e9c25-kube-api-access-jmv77\") pod \"community-operators-nrbln\" (UID: \"ba68071e-eb60-4707-9ddf-7fe7072e9c25\") " pod="openshift-marketplace/community-operators-nrbln" Feb 18 09:30:43 crc kubenswrapper[4556]: I0218 09:30:43.561236 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nrbln" Feb 18 09:30:43 crc kubenswrapper[4556]: I0218 09:30:43.988027 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nrbln"] Feb 18 09:30:44 crc kubenswrapper[4556]: I0218 09:30:44.205309 4556 generic.go:334] "Generic (PLEG): container finished" podID="ba68071e-eb60-4707-9ddf-7fe7072e9c25" containerID="6cee362fe4d3b46a330e33163145641e8b24b00eee4f5d0f7182147dd655f1b4" exitCode=0 Feb 18 09:30:44 crc kubenswrapper[4556]: I0218 09:30:44.205394 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nrbln" event={"ID":"ba68071e-eb60-4707-9ddf-7fe7072e9c25","Type":"ContainerDied","Data":"6cee362fe4d3b46a330e33163145641e8b24b00eee4f5d0f7182147dd655f1b4"} Feb 18 09:30:44 crc kubenswrapper[4556]: I0218 09:30:44.205697 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nrbln" event={"ID":"ba68071e-eb60-4707-9ddf-7fe7072e9c25","Type":"ContainerStarted","Data":"7aa90e5c9e230dca2b6d6e158b198697895dcd8ed862deadcad7834b852090ab"} Feb 18 09:30:45 crc kubenswrapper[4556]: I0218 09:30:45.216483 4556 generic.go:334] "Generic (PLEG): container finished" podID="ba68071e-eb60-4707-9ddf-7fe7072e9c25" containerID="806db858101f1995dfa54c4d566edd0eb3a32a10615352cb700bda8be750276a" exitCode=0 Feb 18 09:30:45 crc kubenswrapper[4556]: I0218 09:30:45.216553 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nrbln" event={"ID":"ba68071e-eb60-4707-9ddf-7fe7072e9c25","Type":"ContainerDied","Data":"806db858101f1995dfa54c4d566edd0eb3a32a10615352cb700bda8be750276a"} Feb 18 09:30:46 crc kubenswrapper[4556]: I0218 09:30:46.225611 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nrbln" event={"ID":"ba68071e-eb60-4707-9ddf-7fe7072e9c25","Type":"ContainerStarted","Data":"c346bfbe55437245c037e7d8056a9b8769fd3b34fdc180bfd270f3b56fda534d"} Feb 18 09:30:46 crc kubenswrapper[4556]: I0218 09:30:46.244299 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nrbln" podStartSLOduration=1.7593364839999999 podStartE2EDuration="3.244286122s" podCreationTimestamp="2026-02-18 09:30:43 +0000 UTC" firstStartedPulling="2026-02-18 09:30:44.206520831 +0000 UTC m=+1601.223481811" lastFinishedPulling="2026-02-18 09:30:45.69147047 +0000 UTC m=+1602.708431449" observedRunningTime="2026-02-18 09:30:46.238179654 +0000 UTC m=+1603.255140644" watchObservedRunningTime="2026-02-18 09:30:46.244286122 +0000 UTC m=+1603.261247103" Feb 18 09:30:53 crc kubenswrapper[4556]: I0218 09:30:53.561812 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nrbln" Feb 18 09:30:53 crc kubenswrapper[4556]: I0218 09:30:53.562214 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nrbln" Feb 18 09:30:53 crc kubenswrapper[4556]: I0218 09:30:53.593477 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nrbln" Feb 18 09:30:54 crc kubenswrapper[4556]: I0218 09:30:54.308776 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nrbln" Feb 18 09:30:54 crc kubenswrapper[4556]: I0218 09:30:54.343594 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nrbln"] Feb 18 09:30:55 crc kubenswrapper[4556]: I0218 09:30:55.282144 4556 scope.go:117] "RemoveContainer" containerID="507a9134b6862bc8be1ac03b7993c890ef64f16c21975e00aa96c53fe413415d" Feb 18 09:30:55 crc kubenswrapper[4556]: E0218 09:30:55.282496 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:30:56 crc kubenswrapper[4556]: I0218 09:30:56.289839 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-nrbln" podUID="ba68071e-eb60-4707-9ddf-7fe7072e9c25" containerName="registry-server" containerID="cri-o://c346bfbe55437245c037e7d8056a9b8769fd3b34fdc180bfd270f3b56fda534d" gracePeriod=2 Feb 18 09:30:56 crc kubenswrapper[4556]: I0218 09:30:56.639252 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nrbln" Feb 18 09:30:56 crc kubenswrapper[4556]: I0218 09:30:56.721579 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jmv77\" (UniqueName: \"kubernetes.io/projected/ba68071e-eb60-4707-9ddf-7fe7072e9c25-kube-api-access-jmv77\") pod \"ba68071e-eb60-4707-9ddf-7fe7072e9c25\" (UID: \"ba68071e-eb60-4707-9ddf-7fe7072e9c25\") " Feb 18 09:30:56 crc kubenswrapper[4556]: I0218 09:30:56.721703 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba68071e-eb60-4707-9ddf-7fe7072e9c25-utilities\") pod \"ba68071e-eb60-4707-9ddf-7fe7072e9c25\" (UID: \"ba68071e-eb60-4707-9ddf-7fe7072e9c25\") " Feb 18 09:30:56 crc kubenswrapper[4556]: I0218 09:30:56.721767 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba68071e-eb60-4707-9ddf-7fe7072e9c25-catalog-content\") pod \"ba68071e-eb60-4707-9ddf-7fe7072e9c25\" (UID: \"ba68071e-eb60-4707-9ddf-7fe7072e9c25\") " Feb 18 09:30:56 crc kubenswrapper[4556]: I0218 09:30:56.722785 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba68071e-eb60-4707-9ddf-7fe7072e9c25-utilities" (OuterVolumeSpecName: "utilities") pod "ba68071e-eb60-4707-9ddf-7fe7072e9c25" (UID: "ba68071e-eb60-4707-9ddf-7fe7072e9c25"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:30:56 crc kubenswrapper[4556]: I0218 09:30:56.726346 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba68071e-eb60-4707-9ddf-7fe7072e9c25-kube-api-access-jmv77" (OuterVolumeSpecName: "kube-api-access-jmv77") pod "ba68071e-eb60-4707-9ddf-7fe7072e9c25" (UID: "ba68071e-eb60-4707-9ddf-7fe7072e9c25"). InnerVolumeSpecName "kube-api-access-jmv77". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:30:56 crc kubenswrapper[4556]: I0218 09:30:56.758367 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba68071e-eb60-4707-9ddf-7fe7072e9c25-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ba68071e-eb60-4707-9ddf-7fe7072e9c25" (UID: "ba68071e-eb60-4707-9ddf-7fe7072e9c25"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:30:56 crc kubenswrapper[4556]: I0218 09:30:56.824192 4556 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba68071e-eb60-4707-9ddf-7fe7072e9c25-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 09:30:56 crc kubenswrapper[4556]: I0218 09:30:56.824218 4556 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba68071e-eb60-4707-9ddf-7fe7072e9c25-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 09:30:56 crc kubenswrapper[4556]: I0218 09:30:56.824230 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jmv77\" (UniqueName: \"kubernetes.io/projected/ba68071e-eb60-4707-9ddf-7fe7072e9c25-kube-api-access-jmv77\") on node \"crc\" DevicePath \"\"" Feb 18 09:30:57 crc kubenswrapper[4556]: I0218 09:30:57.297172 4556 generic.go:334] "Generic (PLEG): container finished" podID="ba68071e-eb60-4707-9ddf-7fe7072e9c25" containerID="c346bfbe55437245c037e7d8056a9b8769fd3b34fdc180bfd270f3b56fda534d" exitCode=0 Feb 18 09:30:57 crc kubenswrapper[4556]: I0218 09:30:57.297211 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nrbln" event={"ID":"ba68071e-eb60-4707-9ddf-7fe7072e9c25","Type":"ContainerDied","Data":"c346bfbe55437245c037e7d8056a9b8769fd3b34fdc180bfd270f3b56fda534d"} Feb 18 09:30:57 crc kubenswrapper[4556]: I0218 09:30:57.297216 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nrbln" Feb 18 09:30:57 crc kubenswrapper[4556]: I0218 09:30:57.297242 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nrbln" event={"ID":"ba68071e-eb60-4707-9ddf-7fe7072e9c25","Type":"ContainerDied","Data":"7aa90e5c9e230dca2b6d6e158b198697895dcd8ed862deadcad7834b852090ab"} Feb 18 09:30:57 crc kubenswrapper[4556]: I0218 09:30:57.297262 4556 scope.go:117] "RemoveContainer" containerID="c346bfbe55437245c037e7d8056a9b8769fd3b34fdc180bfd270f3b56fda534d" Feb 18 09:30:57 crc kubenswrapper[4556]: I0218 09:30:57.313878 4556 scope.go:117] "RemoveContainer" containerID="806db858101f1995dfa54c4d566edd0eb3a32a10615352cb700bda8be750276a" Feb 18 09:30:57 crc kubenswrapper[4556]: I0218 09:30:57.326331 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nrbln"] Feb 18 09:30:57 crc kubenswrapper[4556]: I0218 09:30:57.331560 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-nrbln"] Feb 18 09:30:57 crc kubenswrapper[4556]: I0218 09:30:57.345584 4556 scope.go:117] "RemoveContainer" containerID="6cee362fe4d3b46a330e33163145641e8b24b00eee4f5d0f7182147dd655f1b4" Feb 18 09:30:57 crc kubenswrapper[4556]: I0218 09:30:57.363144 4556 scope.go:117] "RemoveContainer" containerID="c346bfbe55437245c037e7d8056a9b8769fd3b34fdc180bfd270f3b56fda534d" Feb 18 09:30:57 crc kubenswrapper[4556]: E0218 09:30:57.363464 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c346bfbe55437245c037e7d8056a9b8769fd3b34fdc180bfd270f3b56fda534d\": container with ID starting with c346bfbe55437245c037e7d8056a9b8769fd3b34fdc180bfd270f3b56fda534d not found: ID does not exist" containerID="c346bfbe55437245c037e7d8056a9b8769fd3b34fdc180bfd270f3b56fda534d" Feb 18 09:30:57 crc kubenswrapper[4556]: I0218 09:30:57.363579 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c346bfbe55437245c037e7d8056a9b8769fd3b34fdc180bfd270f3b56fda534d"} err="failed to get container status \"c346bfbe55437245c037e7d8056a9b8769fd3b34fdc180bfd270f3b56fda534d\": rpc error: code = NotFound desc = could not find container \"c346bfbe55437245c037e7d8056a9b8769fd3b34fdc180bfd270f3b56fda534d\": container with ID starting with c346bfbe55437245c037e7d8056a9b8769fd3b34fdc180bfd270f3b56fda534d not found: ID does not exist" Feb 18 09:30:57 crc kubenswrapper[4556]: I0218 09:30:57.363674 4556 scope.go:117] "RemoveContainer" containerID="806db858101f1995dfa54c4d566edd0eb3a32a10615352cb700bda8be750276a" Feb 18 09:30:57 crc kubenswrapper[4556]: E0218 09:30:57.363998 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"806db858101f1995dfa54c4d566edd0eb3a32a10615352cb700bda8be750276a\": container with ID starting with 806db858101f1995dfa54c4d566edd0eb3a32a10615352cb700bda8be750276a not found: ID does not exist" containerID="806db858101f1995dfa54c4d566edd0eb3a32a10615352cb700bda8be750276a" Feb 18 09:30:57 crc kubenswrapper[4556]: I0218 09:30:57.364072 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"806db858101f1995dfa54c4d566edd0eb3a32a10615352cb700bda8be750276a"} err="failed to get container status \"806db858101f1995dfa54c4d566edd0eb3a32a10615352cb700bda8be750276a\": rpc error: code = NotFound desc = could not find container \"806db858101f1995dfa54c4d566edd0eb3a32a10615352cb700bda8be750276a\": container with ID starting with 806db858101f1995dfa54c4d566edd0eb3a32a10615352cb700bda8be750276a not found: ID does not exist" Feb 18 09:30:57 crc kubenswrapper[4556]: I0218 09:30:57.364141 4556 scope.go:117] "RemoveContainer" containerID="6cee362fe4d3b46a330e33163145641e8b24b00eee4f5d0f7182147dd655f1b4" Feb 18 09:30:57 crc kubenswrapper[4556]: E0218 09:30:57.364428 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6cee362fe4d3b46a330e33163145641e8b24b00eee4f5d0f7182147dd655f1b4\": container with ID starting with 6cee362fe4d3b46a330e33163145641e8b24b00eee4f5d0f7182147dd655f1b4 not found: ID does not exist" containerID="6cee362fe4d3b46a330e33163145641e8b24b00eee4f5d0f7182147dd655f1b4" Feb 18 09:30:57 crc kubenswrapper[4556]: I0218 09:30:57.364507 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6cee362fe4d3b46a330e33163145641e8b24b00eee4f5d0f7182147dd655f1b4"} err="failed to get container status \"6cee362fe4d3b46a330e33163145641e8b24b00eee4f5d0f7182147dd655f1b4\": rpc error: code = NotFound desc = could not find container \"6cee362fe4d3b46a330e33163145641e8b24b00eee4f5d0f7182147dd655f1b4\": container with ID starting with 6cee362fe4d3b46a330e33163145641e8b24b00eee4f5d0f7182147dd655f1b4 not found: ID does not exist" Feb 18 09:30:59 crc kubenswrapper[4556]: I0218 09:30:59.291230 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba68071e-eb60-4707-9ddf-7fe7072e9c25" path="/var/lib/kubelet/pods/ba68071e-eb60-4707-9ddf-7fe7072e9c25/volumes" Feb 18 09:31:04 crc kubenswrapper[4556]: I0218 09:31:04.350701 4556 generic.go:334] "Generic (PLEG): container finished" podID="7c63f939-1090-4fae-b45b-9d55f09f489d" containerID="62a21056e25119741d39d0f09efa1b5c3fc81a6c2fd19015180907b7c3c76439" exitCode=0 Feb 18 09:31:04 crc kubenswrapper[4556]: I0218 09:31:04.350760 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vgw62" event={"ID":"7c63f939-1090-4fae-b45b-9d55f09f489d","Type":"ContainerDied","Data":"62a21056e25119741d39d0f09efa1b5c3fc81a6c2fd19015180907b7c3c76439"} Feb 18 09:31:05 crc kubenswrapper[4556]: I0218 09:31:05.659406 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vgw62" Feb 18 09:31:05 crc kubenswrapper[4556]: I0218 09:31:05.691287 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7c63f939-1090-4fae-b45b-9d55f09f489d-inventory\") pod \"7c63f939-1090-4fae-b45b-9d55f09f489d\" (UID: \"7c63f939-1090-4fae-b45b-9d55f09f489d\") " Feb 18 09:31:05 crc kubenswrapper[4556]: I0218 09:31:05.691333 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c63f939-1090-4fae-b45b-9d55f09f489d-ovn-combined-ca-bundle\") pod \"7c63f939-1090-4fae-b45b-9d55f09f489d\" (UID: \"7c63f939-1090-4fae-b45b-9d55f09f489d\") " Feb 18 09:31:05 crc kubenswrapper[4556]: I0218 09:31:05.691372 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/7c63f939-1090-4fae-b45b-9d55f09f489d-ovncontroller-config-0\") pod \"7c63f939-1090-4fae-b45b-9d55f09f489d\" (UID: \"7c63f939-1090-4fae-b45b-9d55f09f489d\") " Feb 18 09:31:05 crc kubenswrapper[4556]: I0218 09:31:05.691450 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ndmcv\" (UniqueName: \"kubernetes.io/projected/7c63f939-1090-4fae-b45b-9d55f09f489d-kube-api-access-ndmcv\") pod \"7c63f939-1090-4fae-b45b-9d55f09f489d\" (UID: \"7c63f939-1090-4fae-b45b-9d55f09f489d\") " Feb 18 09:31:05 crc kubenswrapper[4556]: I0218 09:31:05.691604 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7c63f939-1090-4fae-b45b-9d55f09f489d-ssh-key-openstack-edpm-ipam\") pod \"7c63f939-1090-4fae-b45b-9d55f09f489d\" (UID: \"7c63f939-1090-4fae-b45b-9d55f09f489d\") " Feb 18 09:31:05 crc kubenswrapper[4556]: I0218 09:31:05.696061 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c63f939-1090-4fae-b45b-9d55f09f489d-kube-api-access-ndmcv" (OuterVolumeSpecName: "kube-api-access-ndmcv") pod "7c63f939-1090-4fae-b45b-9d55f09f489d" (UID: "7c63f939-1090-4fae-b45b-9d55f09f489d"). InnerVolumeSpecName "kube-api-access-ndmcv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:31:05 crc kubenswrapper[4556]: I0218 09:31:05.696168 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c63f939-1090-4fae-b45b-9d55f09f489d-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "7c63f939-1090-4fae-b45b-9d55f09f489d" (UID: "7c63f939-1090-4fae-b45b-9d55f09f489d"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:31:05 crc kubenswrapper[4556]: I0218 09:31:05.710643 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c63f939-1090-4fae-b45b-9d55f09f489d-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "7c63f939-1090-4fae-b45b-9d55f09f489d" (UID: "7c63f939-1090-4fae-b45b-9d55f09f489d"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:31:05 crc kubenswrapper[4556]: I0218 09:31:05.711619 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c63f939-1090-4fae-b45b-9d55f09f489d-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "7c63f939-1090-4fae-b45b-9d55f09f489d" (UID: "7c63f939-1090-4fae-b45b-9d55f09f489d"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:31:05 crc kubenswrapper[4556]: I0218 09:31:05.712507 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c63f939-1090-4fae-b45b-9d55f09f489d-inventory" (OuterVolumeSpecName: "inventory") pod "7c63f939-1090-4fae-b45b-9d55f09f489d" (UID: "7c63f939-1090-4fae-b45b-9d55f09f489d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:31:05 crc kubenswrapper[4556]: I0218 09:31:05.793489 4556 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7c63f939-1090-4fae-b45b-9d55f09f489d-inventory\") on node \"crc\" DevicePath \"\"" Feb 18 09:31:05 crc kubenswrapper[4556]: I0218 09:31:05.793518 4556 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c63f939-1090-4fae-b45b-9d55f09f489d-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:31:05 crc kubenswrapper[4556]: I0218 09:31:05.793529 4556 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/7c63f939-1090-4fae-b45b-9d55f09f489d-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Feb 18 09:31:05 crc kubenswrapper[4556]: I0218 09:31:05.793537 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ndmcv\" (UniqueName: \"kubernetes.io/projected/7c63f939-1090-4fae-b45b-9d55f09f489d-kube-api-access-ndmcv\") on node \"crc\" DevicePath \"\"" Feb 18 09:31:05 crc kubenswrapper[4556]: I0218 09:31:05.793544 4556 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7c63f939-1090-4fae-b45b-9d55f09f489d-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 18 09:31:06 crc kubenswrapper[4556]: I0218 09:31:06.365410 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vgw62" event={"ID":"7c63f939-1090-4fae-b45b-9d55f09f489d","Type":"ContainerDied","Data":"98beddc14af2cc3504ec58406030a06f887e27669b65f3d7875fbf034d3a631e"} Feb 18 09:31:06 crc kubenswrapper[4556]: I0218 09:31:06.365441 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="98beddc14af2cc3504ec58406030a06f887e27669b65f3d7875fbf034d3a631e" Feb 18 09:31:06 crc kubenswrapper[4556]: I0218 09:31:06.365460 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vgw62" Feb 18 09:31:06 crc kubenswrapper[4556]: I0218 09:31:06.426921 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wcldl"] Feb 18 09:31:06 crc kubenswrapper[4556]: E0218 09:31:06.427298 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c63f939-1090-4fae-b45b-9d55f09f489d" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Feb 18 09:31:06 crc kubenswrapper[4556]: I0218 09:31:06.427314 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c63f939-1090-4fae-b45b-9d55f09f489d" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Feb 18 09:31:06 crc kubenswrapper[4556]: E0218 09:31:06.427329 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba68071e-eb60-4707-9ddf-7fe7072e9c25" containerName="extract-content" Feb 18 09:31:06 crc kubenswrapper[4556]: I0218 09:31:06.427335 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba68071e-eb60-4707-9ddf-7fe7072e9c25" containerName="extract-content" Feb 18 09:31:06 crc kubenswrapper[4556]: E0218 09:31:06.427353 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba68071e-eb60-4707-9ddf-7fe7072e9c25" containerName="extract-utilities" Feb 18 09:31:06 crc kubenswrapper[4556]: I0218 09:31:06.427360 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba68071e-eb60-4707-9ddf-7fe7072e9c25" containerName="extract-utilities" Feb 18 09:31:06 crc kubenswrapper[4556]: E0218 09:31:06.427369 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba68071e-eb60-4707-9ddf-7fe7072e9c25" containerName="registry-server" Feb 18 09:31:06 crc kubenswrapper[4556]: I0218 09:31:06.427375 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba68071e-eb60-4707-9ddf-7fe7072e9c25" containerName="registry-server" Feb 18 09:31:06 crc kubenswrapper[4556]: I0218 09:31:06.427558 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c63f939-1090-4fae-b45b-9d55f09f489d" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Feb 18 09:31:06 crc kubenswrapper[4556]: I0218 09:31:06.427577 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba68071e-eb60-4707-9ddf-7fe7072e9c25" containerName="registry-server" Feb 18 09:31:06 crc kubenswrapper[4556]: I0218 09:31:06.428081 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wcldl" Feb 18 09:31:06 crc kubenswrapper[4556]: I0218 09:31:06.429513 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Feb 18 09:31:06 crc kubenswrapper[4556]: I0218 09:31:06.429729 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 18 09:31:06 crc kubenswrapper[4556]: I0218 09:31:06.429874 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Feb 18 09:31:06 crc kubenswrapper[4556]: I0218 09:31:06.430123 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 18 09:31:06 crc kubenswrapper[4556]: I0218 09:31:06.430277 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-54z7h" Feb 18 09:31:06 crc kubenswrapper[4556]: I0218 09:31:06.430388 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 18 09:31:06 crc kubenswrapper[4556]: I0218 09:31:06.433953 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wcldl"] Feb 18 09:31:06 crc kubenswrapper[4556]: I0218 09:31:06.503883 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3343de49-153e-43bb-9d22-703a5f3a0b0b-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wcldl\" (UID: \"3343de49-153e-43bb-9d22-703a5f3a0b0b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wcldl" Feb 18 09:31:06 crc kubenswrapper[4556]: I0218 09:31:06.504101 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3343de49-153e-43bb-9d22-703a5f3a0b0b-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wcldl\" (UID: \"3343de49-153e-43bb-9d22-703a5f3a0b0b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wcldl" Feb 18 09:31:06 crc kubenswrapper[4556]: I0218 09:31:06.504227 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3343de49-153e-43bb-9d22-703a5f3a0b0b-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wcldl\" (UID: \"3343de49-153e-43bb-9d22-703a5f3a0b0b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wcldl" Feb 18 09:31:06 crc kubenswrapper[4556]: I0218 09:31:06.504307 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qm7kd\" (UniqueName: \"kubernetes.io/projected/3343de49-153e-43bb-9d22-703a5f3a0b0b-kube-api-access-qm7kd\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wcldl\" (UID: \"3343de49-153e-43bb-9d22-703a5f3a0b0b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wcldl" Feb 18 09:31:06 crc kubenswrapper[4556]: I0218 09:31:06.504335 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3343de49-153e-43bb-9d22-703a5f3a0b0b-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wcldl\" (UID: \"3343de49-153e-43bb-9d22-703a5f3a0b0b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wcldl" Feb 18 09:31:06 crc kubenswrapper[4556]: I0218 09:31:06.504507 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3343de49-153e-43bb-9d22-703a5f3a0b0b-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wcldl\" (UID: \"3343de49-153e-43bb-9d22-703a5f3a0b0b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wcldl" Feb 18 09:31:06 crc kubenswrapper[4556]: I0218 09:31:06.606406 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qm7kd\" (UniqueName: \"kubernetes.io/projected/3343de49-153e-43bb-9d22-703a5f3a0b0b-kube-api-access-qm7kd\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wcldl\" (UID: \"3343de49-153e-43bb-9d22-703a5f3a0b0b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wcldl" Feb 18 09:31:06 crc kubenswrapper[4556]: I0218 09:31:06.606469 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3343de49-153e-43bb-9d22-703a5f3a0b0b-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wcldl\" (UID: \"3343de49-153e-43bb-9d22-703a5f3a0b0b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wcldl" Feb 18 09:31:06 crc kubenswrapper[4556]: I0218 09:31:06.606598 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3343de49-153e-43bb-9d22-703a5f3a0b0b-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wcldl\" (UID: \"3343de49-153e-43bb-9d22-703a5f3a0b0b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wcldl" Feb 18 09:31:06 crc kubenswrapper[4556]: I0218 09:31:06.606660 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3343de49-153e-43bb-9d22-703a5f3a0b0b-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wcldl\" (UID: \"3343de49-153e-43bb-9d22-703a5f3a0b0b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wcldl" Feb 18 09:31:06 crc kubenswrapper[4556]: I0218 09:31:06.606696 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3343de49-153e-43bb-9d22-703a5f3a0b0b-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wcldl\" (UID: \"3343de49-153e-43bb-9d22-703a5f3a0b0b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wcldl" Feb 18 09:31:06 crc kubenswrapper[4556]: I0218 09:31:06.606784 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3343de49-153e-43bb-9d22-703a5f3a0b0b-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wcldl\" (UID: \"3343de49-153e-43bb-9d22-703a5f3a0b0b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wcldl" Feb 18 09:31:06 crc kubenswrapper[4556]: I0218 09:31:06.610755 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3343de49-153e-43bb-9d22-703a5f3a0b0b-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wcldl\" (UID: \"3343de49-153e-43bb-9d22-703a5f3a0b0b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wcldl" Feb 18 09:31:06 crc kubenswrapper[4556]: I0218 09:31:06.610953 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3343de49-153e-43bb-9d22-703a5f3a0b0b-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wcldl\" (UID: \"3343de49-153e-43bb-9d22-703a5f3a0b0b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wcldl" Feb 18 09:31:06 crc kubenswrapper[4556]: I0218 09:31:06.610958 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3343de49-153e-43bb-9d22-703a5f3a0b0b-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wcldl\" (UID: \"3343de49-153e-43bb-9d22-703a5f3a0b0b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wcldl" Feb 18 09:31:06 crc kubenswrapper[4556]: I0218 09:31:06.611287 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3343de49-153e-43bb-9d22-703a5f3a0b0b-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wcldl\" (UID: \"3343de49-153e-43bb-9d22-703a5f3a0b0b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wcldl" Feb 18 09:31:06 crc kubenswrapper[4556]: I0218 09:31:06.612647 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3343de49-153e-43bb-9d22-703a5f3a0b0b-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wcldl\" (UID: \"3343de49-153e-43bb-9d22-703a5f3a0b0b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wcldl" Feb 18 09:31:06 crc kubenswrapper[4556]: I0218 09:31:06.630719 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qm7kd\" (UniqueName: \"kubernetes.io/projected/3343de49-153e-43bb-9d22-703a5f3a0b0b-kube-api-access-qm7kd\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wcldl\" (UID: \"3343de49-153e-43bb-9d22-703a5f3a0b0b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wcldl" Feb 18 09:31:06 crc kubenswrapper[4556]: I0218 09:31:06.742770 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wcldl" Feb 18 09:31:07 crc kubenswrapper[4556]: I0218 09:31:07.164669 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wcldl"] Feb 18 09:31:07 crc kubenswrapper[4556]: I0218 09:31:07.373224 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wcldl" event={"ID":"3343de49-153e-43bb-9d22-703a5f3a0b0b","Type":"ContainerStarted","Data":"6116284a4ada294af3d29dd4f9d1f36d16f247e4e7ebdb3c2a973ee1c93f2e72"} Feb 18 09:31:08 crc kubenswrapper[4556]: I0218 09:31:08.381173 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wcldl" event={"ID":"3343de49-153e-43bb-9d22-703a5f3a0b0b","Type":"ContainerStarted","Data":"5699fc9dda965641fabedb95a40fcdbf4d31390fc79ff0d524247b0c5b1a3881"} Feb 18 09:31:08 crc kubenswrapper[4556]: I0218 09:31:08.396687 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wcldl" podStartSLOduration=1.90905084 podStartE2EDuration="2.396668589s" podCreationTimestamp="2026-02-18 09:31:06 +0000 UTC" firstStartedPulling="2026-02-18 09:31:07.168191945 +0000 UTC m=+1624.185152925" lastFinishedPulling="2026-02-18 09:31:07.655809694 +0000 UTC m=+1624.672770674" observedRunningTime="2026-02-18 09:31:08.392598502 +0000 UTC m=+1625.409559482" watchObservedRunningTime="2026-02-18 09:31:08.396668589 +0000 UTC m=+1625.413629569" Feb 18 09:31:09 crc kubenswrapper[4556]: I0218 09:31:09.282125 4556 scope.go:117] "RemoveContainer" containerID="507a9134b6862bc8be1ac03b7993c890ef64f16c21975e00aa96c53fe413415d" Feb 18 09:31:09 crc kubenswrapper[4556]: E0218 09:31:09.282538 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:31:10 crc kubenswrapper[4556]: I0218 09:31:10.801605 4556 scope.go:117] "RemoveContainer" containerID="d745e502212fd8d5ef69ed76fecf788d35f7191b8d1d71353d6a73a4736c0524" Feb 18 09:31:24 crc kubenswrapper[4556]: I0218 09:31:24.282206 4556 scope.go:117] "RemoveContainer" containerID="507a9134b6862bc8be1ac03b7993c890ef64f16c21975e00aa96c53fe413415d" Feb 18 09:31:24 crc kubenswrapper[4556]: E0218 09:31:24.282826 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:31:36 crc kubenswrapper[4556]: I0218 09:31:36.283310 4556 scope.go:117] "RemoveContainer" containerID="507a9134b6862bc8be1ac03b7993c890ef64f16c21975e00aa96c53fe413415d" Feb 18 09:31:36 crc kubenswrapper[4556]: E0218 09:31:36.284548 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:31:41 crc kubenswrapper[4556]: I0218 09:31:41.611878 4556 generic.go:334] "Generic (PLEG): container finished" podID="3343de49-153e-43bb-9d22-703a5f3a0b0b" containerID="5699fc9dda965641fabedb95a40fcdbf4d31390fc79ff0d524247b0c5b1a3881" exitCode=0 Feb 18 09:31:41 crc kubenswrapper[4556]: I0218 09:31:41.611963 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wcldl" event={"ID":"3343de49-153e-43bb-9d22-703a5f3a0b0b","Type":"ContainerDied","Data":"5699fc9dda965641fabedb95a40fcdbf4d31390fc79ff0d524247b0c5b1a3881"} Feb 18 09:31:42 crc kubenswrapper[4556]: I0218 09:31:42.984723 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wcldl" Feb 18 09:31:43 crc kubenswrapper[4556]: I0218 09:31:43.010717 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3343de49-153e-43bb-9d22-703a5f3a0b0b-inventory\") pod \"3343de49-153e-43bb-9d22-703a5f3a0b0b\" (UID: \"3343de49-153e-43bb-9d22-703a5f3a0b0b\") " Feb 18 09:31:43 crc kubenswrapper[4556]: I0218 09:31:43.010781 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3343de49-153e-43bb-9d22-703a5f3a0b0b-nova-metadata-neutron-config-0\") pod \"3343de49-153e-43bb-9d22-703a5f3a0b0b\" (UID: \"3343de49-153e-43bb-9d22-703a5f3a0b0b\") " Feb 18 09:31:43 crc kubenswrapper[4556]: I0218 09:31:43.010809 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qm7kd\" (UniqueName: \"kubernetes.io/projected/3343de49-153e-43bb-9d22-703a5f3a0b0b-kube-api-access-qm7kd\") pod \"3343de49-153e-43bb-9d22-703a5f3a0b0b\" (UID: \"3343de49-153e-43bb-9d22-703a5f3a0b0b\") " Feb 18 09:31:43 crc kubenswrapper[4556]: I0218 09:31:43.010834 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3343de49-153e-43bb-9d22-703a5f3a0b0b-neutron-metadata-combined-ca-bundle\") pod \"3343de49-153e-43bb-9d22-703a5f3a0b0b\" (UID: \"3343de49-153e-43bb-9d22-703a5f3a0b0b\") " Feb 18 09:31:43 crc kubenswrapper[4556]: I0218 09:31:43.010941 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3343de49-153e-43bb-9d22-703a5f3a0b0b-ssh-key-openstack-edpm-ipam\") pod \"3343de49-153e-43bb-9d22-703a5f3a0b0b\" (UID: \"3343de49-153e-43bb-9d22-703a5f3a0b0b\") " Feb 18 09:31:43 crc kubenswrapper[4556]: I0218 09:31:43.011081 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3343de49-153e-43bb-9d22-703a5f3a0b0b-neutron-ovn-metadata-agent-neutron-config-0\") pod \"3343de49-153e-43bb-9d22-703a5f3a0b0b\" (UID: \"3343de49-153e-43bb-9d22-703a5f3a0b0b\") " Feb 18 09:31:43 crc kubenswrapper[4556]: I0218 09:31:43.015973 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3343de49-153e-43bb-9d22-703a5f3a0b0b-kube-api-access-qm7kd" (OuterVolumeSpecName: "kube-api-access-qm7kd") pod "3343de49-153e-43bb-9d22-703a5f3a0b0b" (UID: "3343de49-153e-43bb-9d22-703a5f3a0b0b"). InnerVolumeSpecName "kube-api-access-qm7kd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:31:43 crc kubenswrapper[4556]: I0218 09:31:43.016086 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3343de49-153e-43bb-9d22-703a5f3a0b0b-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "3343de49-153e-43bb-9d22-703a5f3a0b0b" (UID: "3343de49-153e-43bb-9d22-703a5f3a0b0b"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:31:43 crc kubenswrapper[4556]: I0218 09:31:43.033076 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3343de49-153e-43bb-9d22-703a5f3a0b0b-inventory" (OuterVolumeSpecName: "inventory") pod "3343de49-153e-43bb-9d22-703a5f3a0b0b" (UID: "3343de49-153e-43bb-9d22-703a5f3a0b0b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:31:43 crc kubenswrapper[4556]: I0218 09:31:43.033819 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3343de49-153e-43bb-9d22-703a5f3a0b0b-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "3343de49-153e-43bb-9d22-703a5f3a0b0b" (UID: "3343de49-153e-43bb-9d22-703a5f3a0b0b"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:31:43 crc kubenswrapper[4556]: I0218 09:31:43.035362 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3343de49-153e-43bb-9d22-703a5f3a0b0b-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "3343de49-153e-43bb-9d22-703a5f3a0b0b" (UID: "3343de49-153e-43bb-9d22-703a5f3a0b0b"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:31:43 crc kubenswrapper[4556]: I0218 09:31:43.037822 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3343de49-153e-43bb-9d22-703a5f3a0b0b-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "3343de49-153e-43bb-9d22-703a5f3a0b0b" (UID: "3343de49-153e-43bb-9d22-703a5f3a0b0b"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:31:43 crc kubenswrapper[4556]: I0218 09:31:43.113111 4556 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3343de49-153e-43bb-9d22-703a5f3a0b0b-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 18 09:31:43 crc kubenswrapper[4556]: I0218 09:31:43.113136 4556 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3343de49-153e-43bb-9d22-703a5f3a0b0b-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Feb 18 09:31:43 crc kubenswrapper[4556]: I0218 09:31:43.113149 4556 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3343de49-153e-43bb-9d22-703a5f3a0b0b-inventory\") on node \"crc\" DevicePath \"\"" Feb 18 09:31:43 crc kubenswrapper[4556]: I0218 09:31:43.113177 4556 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3343de49-153e-43bb-9d22-703a5f3a0b0b-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Feb 18 09:31:43 crc kubenswrapper[4556]: I0218 09:31:43.113191 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qm7kd\" (UniqueName: \"kubernetes.io/projected/3343de49-153e-43bb-9d22-703a5f3a0b0b-kube-api-access-qm7kd\") on node \"crc\" DevicePath \"\"" Feb 18 09:31:43 crc kubenswrapper[4556]: I0218 09:31:43.113201 4556 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3343de49-153e-43bb-9d22-703a5f3a0b0b-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:31:43 crc kubenswrapper[4556]: I0218 09:31:43.627502 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wcldl" event={"ID":"3343de49-153e-43bb-9d22-703a5f3a0b0b","Type":"ContainerDied","Data":"6116284a4ada294af3d29dd4f9d1f36d16f247e4e7ebdb3c2a973ee1c93f2e72"} Feb 18 09:31:43 crc kubenswrapper[4556]: I0218 09:31:43.627549 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6116284a4ada294af3d29dd4f9d1f36d16f247e4e7ebdb3c2a973ee1c93f2e72" Feb 18 09:31:43 crc kubenswrapper[4556]: I0218 09:31:43.627555 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wcldl" Feb 18 09:31:43 crc kubenswrapper[4556]: I0218 09:31:43.738354 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hhn8t"] Feb 18 09:31:43 crc kubenswrapper[4556]: E0218 09:31:43.738668 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3343de49-153e-43bb-9d22-703a5f3a0b0b" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Feb 18 09:31:43 crc kubenswrapper[4556]: I0218 09:31:43.738686 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="3343de49-153e-43bb-9d22-703a5f3a0b0b" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Feb 18 09:31:43 crc kubenswrapper[4556]: I0218 09:31:43.738881 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="3343de49-153e-43bb-9d22-703a5f3a0b0b" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Feb 18 09:31:43 crc kubenswrapper[4556]: I0218 09:31:43.739416 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hhn8t" Feb 18 09:31:43 crc kubenswrapper[4556]: I0218 09:31:43.741376 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-54z7h" Feb 18 09:31:43 crc kubenswrapper[4556]: I0218 09:31:43.742140 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 18 09:31:43 crc kubenswrapper[4556]: I0218 09:31:43.743591 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Feb 18 09:31:43 crc kubenswrapper[4556]: I0218 09:31:43.743899 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 18 09:31:43 crc kubenswrapper[4556]: I0218 09:31:43.744221 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 18 09:31:43 crc kubenswrapper[4556]: I0218 09:31:43.749223 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hhn8t"] Feb 18 09:31:43 crc kubenswrapper[4556]: I0218 09:31:43.831714 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cff0286-97e0-422e-b55c-4f28711bde2d-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-hhn8t\" (UID: \"2cff0286-97e0-422e-b55c-4f28711bde2d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hhn8t" Feb 18 09:31:43 crc kubenswrapper[4556]: I0218 09:31:43.831979 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2cff0286-97e0-422e-b55c-4f28711bde2d-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-hhn8t\" (UID: \"2cff0286-97e0-422e-b55c-4f28711bde2d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hhn8t" Feb 18 09:31:43 crc kubenswrapper[4556]: I0218 09:31:43.832000 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/2cff0286-97e0-422e-b55c-4f28711bde2d-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-hhn8t\" (UID: \"2cff0286-97e0-422e-b55c-4f28711bde2d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hhn8t" Feb 18 09:31:43 crc kubenswrapper[4556]: I0218 09:31:43.832166 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwsqd\" (UniqueName: \"kubernetes.io/projected/2cff0286-97e0-422e-b55c-4f28711bde2d-kube-api-access-lwsqd\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-hhn8t\" (UID: \"2cff0286-97e0-422e-b55c-4f28711bde2d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hhn8t" Feb 18 09:31:43 crc kubenswrapper[4556]: I0218 09:31:43.832249 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2cff0286-97e0-422e-b55c-4f28711bde2d-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-hhn8t\" (UID: \"2cff0286-97e0-422e-b55c-4f28711bde2d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hhn8t" Feb 18 09:31:43 crc kubenswrapper[4556]: I0218 09:31:43.934054 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwsqd\" (UniqueName: \"kubernetes.io/projected/2cff0286-97e0-422e-b55c-4f28711bde2d-kube-api-access-lwsqd\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-hhn8t\" (UID: \"2cff0286-97e0-422e-b55c-4f28711bde2d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hhn8t" Feb 18 09:31:43 crc kubenswrapper[4556]: I0218 09:31:43.934109 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2cff0286-97e0-422e-b55c-4f28711bde2d-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-hhn8t\" (UID: \"2cff0286-97e0-422e-b55c-4f28711bde2d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hhn8t" Feb 18 09:31:43 crc kubenswrapper[4556]: I0218 09:31:43.934421 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cff0286-97e0-422e-b55c-4f28711bde2d-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-hhn8t\" (UID: \"2cff0286-97e0-422e-b55c-4f28711bde2d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hhn8t" Feb 18 09:31:43 crc kubenswrapper[4556]: I0218 09:31:43.934445 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2cff0286-97e0-422e-b55c-4f28711bde2d-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-hhn8t\" (UID: \"2cff0286-97e0-422e-b55c-4f28711bde2d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hhn8t" Feb 18 09:31:43 crc kubenswrapper[4556]: I0218 09:31:43.934463 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/2cff0286-97e0-422e-b55c-4f28711bde2d-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-hhn8t\" (UID: \"2cff0286-97e0-422e-b55c-4f28711bde2d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hhn8t" Feb 18 09:31:43 crc kubenswrapper[4556]: I0218 09:31:43.937666 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2cff0286-97e0-422e-b55c-4f28711bde2d-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-hhn8t\" (UID: \"2cff0286-97e0-422e-b55c-4f28711bde2d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hhn8t" Feb 18 09:31:43 crc kubenswrapper[4556]: I0218 09:31:43.937674 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2cff0286-97e0-422e-b55c-4f28711bde2d-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-hhn8t\" (UID: \"2cff0286-97e0-422e-b55c-4f28711bde2d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hhn8t" Feb 18 09:31:43 crc kubenswrapper[4556]: I0218 09:31:43.938797 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/2cff0286-97e0-422e-b55c-4f28711bde2d-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-hhn8t\" (UID: \"2cff0286-97e0-422e-b55c-4f28711bde2d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hhn8t" Feb 18 09:31:43 crc kubenswrapper[4556]: I0218 09:31:43.938813 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cff0286-97e0-422e-b55c-4f28711bde2d-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-hhn8t\" (UID: \"2cff0286-97e0-422e-b55c-4f28711bde2d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hhn8t" Feb 18 09:31:43 crc kubenswrapper[4556]: I0218 09:31:43.950429 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwsqd\" (UniqueName: \"kubernetes.io/projected/2cff0286-97e0-422e-b55c-4f28711bde2d-kube-api-access-lwsqd\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-hhn8t\" (UID: \"2cff0286-97e0-422e-b55c-4f28711bde2d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hhn8t" Feb 18 09:31:44 crc kubenswrapper[4556]: I0218 09:31:44.060396 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hhn8t" Feb 18 09:31:44 crc kubenswrapper[4556]: I0218 09:31:44.483179 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hhn8t"] Feb 18 09:31:44 crc kubenswrapper[4556]: I0218 09:31:44.634143 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hhn8t" event={"ID":"2cff0286-97e0-422e-b55c-4f28711bde2d","Type":"ContainerStarted","Data":"9294d7863964654bf752b5cb59e6b62c63525bdc4cd3ab26d0afb06e4437bc18"} Feb 18 09:31:45 crc kubenswrapper[4556]: I0218 09:31:45.641217 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hhn8t" event={"ID":"2cff0286-97e0-422e-b55c-4f28711bde2d","Type":"ContainerStarted","Data":"b5ea566219e7ce686e7970072acca2602e937c928ffd94559395da626656e7fd"} Feb 18 09:31:45 crc kubenswrapper[4556]: I0218 09:31:45.662219 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hhn8t" podStartSLOduration=2.06458857 podStartE2EDuration="2.662203009s" podCreationTimestamp="2026-02-18 09:31:43 +0000 UTC" firstStartedPulling="2026-02-18 09:31:44.486492571 +0000 UTC m=+1661.503453550" lastFinishedPulling="2026-02-18 09:31:45.084107009 +0000 UTC m=+1662.101067989" observedRunningTime="2026-02-18 09:31:45.655751249 +0000 UTC m=+1662.672712229" watchObservedRunningTime="2026-02-18 09:31:45.662203009 +0000 UTC m=+1662.679163990" Feb 18 09:31:49 crc kubenswrapper[4556]: I0218 09:31:49.282267 4556 scope.go:117] "RemoveContainer" containerID="507a9134b6862bc8be1ac03b7993c890ef64f16c21975e00aa96c53fe413415d" Feb 18 09:31:49 crc kubenswrapper[4556]: E0218 09:31:49.282876 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:32:04 crc kubenswrapper[4556]: I0218 09:32:04.282672 4556 scope.go:117] "RemoveContainer" containerID="507a9134b6862bc8be1ac03b7993c890ef64f16c21975e00aa96c53fe413415d" Feb 18 09:32:04 crc kubenswrapper[4556]: E0218 09:32:04.283464 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:32:18 crc kubenswrapper[4556]: I0218 09:32:18.282601 4556 scope.go:117] "RemoveContainer" containerID="507a9134b6862bc8be1ac03b7993c890ef64f16c21975e00aa96c53fe413415d" Feb 18 09:32:18 crc kubenswrapper[4556]: E0218 09:32:18.283521 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:32:30 crc kubenswrapper[4556]: I0218 09:32:30.282329 4556 scope.go:117] "RemoveContainer" containerID="507a9134b6862bc8be1ac03b7993c890ef64f16c21975e00aa96c53fe413415d" Feb 18 09:32:30 crc kubenswrapper[4556]: E0218 09:32:30.283030 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:32:45 crc kubenswrapper[4556]: I0218 09:32:45.283063 4556 scope.go:117] "RemoveContainer" containerID="507a9134b6862bc8be1ac03b7993c890ef64f16c21975e00aa96c53fe413415d" Feb 18 09:32:45 crc kubenswrapper[4556]: E0218 09:32:45.283867 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:32:58 crc kubenswrapper[4556]: I0218 09:32:58.283143 4556 scope.go:117] "RemoveContainer" containerID="507a9134b6862bc8be1ac03b7993c890ef64f16c21975e00aa96c53fe413415d" Feb 18 09:32:58 crc kubenswrapper[4556]: E0218 09:32:58.284039 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:33:11 crc kubenswrapper[4556]: I0218 09:33:11.283049 4556 scope.go:117] "RemoveContainer" containerID="507a9134b6862bc8be1ac03b7993c890ef64f16c21975e00aa96c53fe413415d" Feb 18 09:33:11 crc kubenswrapper[4556]: E0218 09:33:11.283875 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:33:24 crc kubenswrapper[4556]: I0218 09:33:24.282567 4556 scope.go:117] "RemoveContainer" containerID="507a9134b6862bc8be1ac03b7993c890ef64f16c21975e00aa96c53fe413415d" Feb 18 09:33:24 crc kubenswrapper[4556]: E0218 09:33:24.283121 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:33:36 crc kubenswrapper[4556]: I0218 09:33:36.282189 4556 scope.go:117] "RemoveContainer" containerID="507a9134b6862bc8be1ac03b7993c890ef64f16c21975e00aa96c53fe413415d" Feb 18 09:33:36 crc kubenswrapper[4556]: E0218 09:33:36.282704 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:33:50 crc kubenswrapper[4556]: I0218 09:33:50.282731 4556 scope.go:117] "RemoveContainer" containerID="507a9134b6862bc8be1ac03b7993c890ef64f16c21975e00aa96c53fe413415d" Feb 18 09:33:50 crc kubenswrapper[4556]: E0218 09:33:50.283384 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:33:54 crc kubenswrapper[4556]: I0218 09:33:54.856318 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zw26v"] Feb 18 09:33:54 crc kubenswrapper[4556]: I0218 09:33:54.879524 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zw26v" Feb 18 09:33:54 crc kubenswrapper[4556]: I0218 09:33:54.890557 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zw26v"] Feb 18 09:33:55 crc kubenswrapper[4556]: I0218 09:33:55.005235 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36a80da9-5d55-45c9-8f91-bcf47149d4a0-utilities\") pod \"certified-operators-zw26v\" (UID: \"36a80da9-5d55-45c9-8f91-bcf47149d4a0\") " pod="openshift-marketplace/certified-operators-zw26v" Feb 18 09:33:55 crc kubenswrapper[4556]: I0218 09:33:55.005439 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36a80da9-5d55-45c9-8f91-bcf47149d4a0-catalog-content\") pod \"certified-operators-zw26v\" (UID: \"36a80da9-5d55-45c9-8f91-bcf47149d4a0\") " pod="openshift-marketplace/certified-operators-zw26v" Feb 18 09:33:55 crc kubenswrapper[4556]: I0218 09:33:55.005557 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjlg9\" (UniqueName: \"kubernetes.io/projected/36a80da9-5d55-45c9-8f91-bcf47149d4a0-kube-api-access-cjlg9\") pod \"certified-operators-zw26v\" (UID: \"36a80da9-5d55-45c9-8f91-bcf47149d4a0\") " pod="openshift-marketplace/certified-operators-zw26v" Feb 18 09:33:55 crc kubenswrapper[4556]: I0218 09:33:55.107806 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36a80da9-5d55-45c9-8f91-bcf47149d4a0-catalog-content\") pod \"certified-operators-zw26v\" (UID: \"36a80da9-5d55-45c9-8f91-bcf47149d4a0\") " pod="openshift-marketplace/certified-operators-zw26v" Feb 18 09:33:55 crc kubenswrapper[4556]: I0218 09:33:55.107961 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjlg9\" (UniqueName: \"kubernetes.io/projected/36a80da9-5d55-45c9-8f91-bcf47149d4a0-kube-api-access-cjlg9\") pod \"certified-operators-zw26v\" (UID: \"36a80da9-5d55-45c9-8f91-bcf47149d4a0\") " pod="openshift-marketplace/certified-operators-zw26v" Feb 18 09:33:55 crc kubenswrapper[4556]: I0218 09:33:55.108072 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36a80da9-5d55-45c9-8f91-bcf47149d4a0-utilities\") pod \"certified-operators-zw26v\" (UID: \"36a80da9-5d55-45c9-8f91-bcf47149d4a0\") " pod="openshift-marketplace/certified-operators-zw26v" Feb 18 09:33:55 crc kubenswrapper[4556]: I0218 09:33:55.108398 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36a80da9-5d55-45c9-8f91-bcf47149d4a0-catalog-content\") pod \"certified-operators-zw26v\" (UID: \"36a80da9-5d55-45c9-8f91-bcf47149d4a0\") " pod="openshift-marketplace/certified-operators-zw26v" Feb 18 09:33:55 crc kubenswrapper[4556]: I0218 09:33:55.108516 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36a80da9-5d55-45c9-8f91-bcf47149d4a0-utilities\") pod \"certified-operators-zw26v\" (UID: \"36a80da9-5d55-45c9-8f91-bcf47149d4a0\") " pod="openshift-marketplace/certified-operators-zw26v" Feb 18 09:33:55 crc kubenswrapper[4556]: I0218 09:33:55.126574 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjlg9\" (UniqueName: \"kubernetes.io/projected/36a80da9-5d55-45c9-8f91-bcf47149d4a0-kube-api-access-cjlg9\") pod \"certified-operators-zw26v\" (UID: \"36a80da9-5d55-45c9-8f91-bcf47149d4a0\") " pod="openshift-marketplace/certified-operators-zw26v" Feb 18 09:33:55 crc kubenswrapper[4556]: I0218 09:33:55.209202 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zw26v" Feb 18 09:33:55 crc kubenswrapper[4556]: I0218 09:33:55.670240 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zw26v"] Feb 18 09:33:56 crc kubenswrapper[4556]: I0218 09:33:56.519875 4556 generic.go:334] "Generic (PLEG): container finished" podID="36a80da9-5d55-45c9-8f91-bcf47149d4a0" containerID="300423e35fc4e9712c44292607573876b5a74eb3ad5a842f43e4fd315bbb14aa" exitCode=0 Feb 18 09:33:56 crc kubenswrapper[4556]: I0218 09:33:56.519918 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zw26v" event={"ID":"36a80da9-5d55-45c9-8f91-bcf47149d4a0","Type":"ContainerDied","Data":"300423e35fc4e9712c44292607573876b5a74eb3ad5a842f43e4fd315bbb14aa"} Feb 18 09:33:56 crc kubenswrapper[4556]: I0218 09:33:56.520218 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zw26v" event={"ID":"36a80da9-5d55-45c9-8f91-bcf47149d4a0","Type":"ContainerStarted","Data":"e3737665ffd1722039ddf81f88680a3bb032b4061b05537b9ad1abfc97062eee"} Feb 18 09:33:56 crc kubenswrapper[4556]: I0218 09:33:56.522167 4556 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 18 09:33:57 crc kubenswrapper[4556]: I0218 09:33:57.534713 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zw26v" event={"ID":"36a80da9-5d55-45c9-8f91-bcf47149d4a0","Type":"ContainerStarted","Data":"6e1e2e3be0588f55f963b7c0929f1ec1c3077c8ab94b994bd71bae196c9e8384"} Feb 18 09:33:58 crc kubenswrapper[4556]: I0218 09:33:58.561332 4556 generic.go:334] "Generic (PLEG): container finished" podID="36a80da9-5d55-45c9-8f91-bcf47149d4a0" containerID="6e1e2e3be0588f55f963b7c0929f1ec1c3077c8ab94b994bd71bae196c9e8384" exitCode=0 Feb 18 09:33:58 crc kubenswrapper[4556]: I0218 09:33:58.561399 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zw26v" event={"ID":"36a80da9-5d55-45c9-8f91-bcf47149d4a0","Type":"ContainerDied","Data":"6e1e2e3be0588f55f963b7c0929f1ec1c3077c8ab94b994bd71bae196c9e8384"} Feb 18 09:33:58 crc kubenswrapper[4556]: I0218 09:33:58.561430 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zw26v" event={"ID":"36a80da9-5d55-45c9-8f91-bcf47149d4a0","Type":"ContainerStarted","Data":"435a0faca4634ff45a721685b8a0a92b908692540b1d3b176fc9070698963fdf"} Feb 18 09:33:58 crc kubenswrapper[4556]: I0218 09:33:58.582440 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zw26v" podStartSLOduration=3.103426652 podStartE2EDuration="4.582422227s" podCreationTimestamp="2026-02-18 09:33:54 +0000 UTC" firstStartedPulling="2026-02-18 09:33:56.521907552 +0000 UTC m=+1793.538868533" lastFinishedPulling="2026-02-18 09:33:58.000903129 +0000 UTC m=+1795.017864108" observedRunningTime="2026-02-18 09:33:58.578711487 +0000 UTC m=+1795.595672467" watchObservedRunningTime="2026-02-18 09:33:58.582422227 +0000 UTC m=+1795.599383206" Feb 18 09:34:04 crc kubenswrapper[4556]: I0218 09:34:04.283056 4556 scope.go:117] "RemoveContainer" containerID="507a9134b6862bc8be1ac03b7993c890ef64f16c21975e00aa96c53fe413415d" Feb 18 09:34:04 crc kubenswrapper[4556]: I0218 09:34:04.605389 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" event={"ID":"8dac7f27-d3d1-4778-9e54-f273035a1d37","Type":"ContainerStarted","Data":"fd40a6d72be7a9ed07291f69c89aeb615c5bf9535dc9706ff886f8f14c11f976"} Feb 18 09:34:05 crc kubenswrapper[4556]: I0218 09:34:05.209934 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zw26v" Feb 18 09:34:05 crc kubenswrapper[4556]: I0218 09:34:05.210393 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zw26v" Feb 18 09:34:05 crc kubenswrapper[4556]: I0218 09:34:05.250268 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zw26v" Feb 18 09:34:05 crc kubenswrapper[4556]: I0218 09:34:05.647319 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zw26v" Feb 18 09:34:05 crc kubenswrapper[4556]: I0218 09:34:05.692386 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zw26v"] Feb 18 09:34:07 crc kubenswrapper[4556]: I0218 09:34:07.632003 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zw26v" podUID="36a80da9-5d55-45c9-8f91-bcf47149d4a0" containerName="registry-server" containerID="cri-o://435a0faca4634ff45a721685b8a0a92b908692540b1d3b176fc9070698963fdf" gracePeriod=2 Feb 18 09:34:08 crc kubenswrapper[4556]: I0218 09:34:08.002356 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zw26v" Feb 18 09:34:08 crc kubenswrapper[4556]: I0218 09:34:08.097927 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36a80da9-5d55-45c9-8f91-bcf47149d4a0-catalog-content\") pod \"36a80da9-5d55-45c9-8f91-bcf47149d4a0\" (UID: \"36a80da9-5d55-45c9-8f91-bcf47149d4a0\") " Feb 18 09:34:08 crc kubenswrapper[4556]: I0218 09:34:08.098209 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36a80da9-5d55-45c9-8f91-bcf47149d4a0-utilities\") pod \"36a80da9-5d55-45c9-8f91-bcf47149d4a0\" (UID: \"36a80da9-5d55-45c9-8f91-bcf47149d4a0\") " Feb 18 09:34:08 crc kubenswrapper[4556]: I0218 09:34:08.098385 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cjlg9\" (UniqueName: \"kubernetes.io/projected/36a80da9-5d55-45c9-8f91-bcf47149d4a0-kube-api-access-cjlg9\") pod \"36a80da9-5d55-45c9-8f91-bcf47149d4a0\" (UID: \"36a80da9-5d55-45c9-8f91-bcf47149d4a0\") " Feb 18 09:34:08 crc kubenswrapper[4556]: I0218 09:34:08.098825 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36a80da9-5d55-45c9-8f91-bcf47149d4a0-utilities" (OuterVolumeSpecName: "utilities") pod "36a80da9-5d55-45c9-8f91-bcf47149d4a0" (UID: "36a80da9-5d55-45c9-8f91-bcf47149d4a0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:34:08 crc kubenswrapper[4556]: I0218 09:34:08.099259 4556 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36a80da9-5d55-45c9-8f91-bcf47149d4a0-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 09:34:08 crc kubenswrapper[4556]: I0218 09:34:08.107982 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36a80da9-5d55-45c9-8f91-bcf47149d4a0-kube-api-access-cjlg9" (OuterVolumeSpecName: "kube-api-access-cjlg9") pod "36a80da9-5d55-45c9-8f91-bcf47149d4a0" (UID: "36a80da9-5d55-45c9-8f91-bcf47149d4a0"). InnerVolumeSpecName "kube-api-access-cjlg9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:34:08 crc kubenswrapper[4556]: I0218 09:34:08.139327 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36a80da9-5d55-45c9-8f91-bcf47149d4a0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "36a80da9-5d55-45c9-8f91-bcf47149d4a0" (UID: "36a80da9-5d55-45c9-8f91-bcf47149d4a0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:34:08 crc kubenswrapper[4556]: I0218 09:34:08.201439 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cjlg9\" (UniqueName: \"kubernetes.io/projected/36a80da9-5d55-45c9-8f91-bcf47149d4a0-kube-api-access-cjlg9\") on node \"crc\" DevicePath \"\"" Feb 18 09:34:08 crc kubenswrapper[4556]: I0218 09:34:08.201473 4556 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36a80da9-5d55-45c9-8f91-bcf47149d4a0-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 09:34:08 crc kubenswrapper[4556]: I0218 09:34:08.640173 4556 generic.go:334] "Generic (PLEG): container finished" podID="36a80da9-5d55-45c9-8f91-bcf47149d4a0" containerID="435a0faca4634ff45a721685b8a0a92b908692540b1d3b176fc9070698963fdf" exitCode=0 Feb 18 09:34:08 crc kubenswrapper[4556]: I0218 09:34:08.640212 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zw26v" event={"ID":"36a80da9-5d55-45c9-8f91-bcf47149d4a0","Type":"ContainerDied","Data":"435a0faca4634ff45a721685b8a0a92b908692540b1d3b176fc9070698963fdf"} Feb 18 09:34:08 crc kubenswrapper[4556]: I0218 09:34:08.640408 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zw26v" event={"ID":"36a80da9-5d55-45c9-8f91-bcf47149d4a0","Type":"ContainerDied","Data":"e3737665ffd1722039ddf81f88680a3bb032b4061b05537b9ad1abfc97062eee"} Feb 18 09:34:08 crc kubenswrapper[4556]: I0218 09:34:08.640429 4556 scope.go:117] "RemoveContainer" containerID="435a0faca4634ff45a721685b8a0a92b908692540b1d3b176fc9070698963fdf" Feb 18 09:34:08 crc kubenswrapper[4556]: I0218 09:34:08.640245 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zw26v" Feb 18 09:34:08 crc kubenswrapper[4556]: I0218 09:34:08.658112 4556 scope.go:117] "RemoveContainer" containerID="6e1e2e3be0588f55f963b7c0929f1ec1c3077c8ab94b994bd71bae196c9e8384" Feb 18 09:34:08 crc kubenswrapper[4556]: I0218 09:34:08.676445 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zw26v"] Feb 18 09:34:08 crc kubenswrapper[4556]: I0218 09:34:08.683865 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zw26v"] Feb 18 09:34:08 crc kubenswrapper[4556]: I0218 09:34:08.690500 4556 scope.go:117] "RemoveContainer" containerID="300423e35fc4e9712c44292607573876b5a74eb3ad5a842f43e4fd315bbb14aa" Feb 18 09:34:08 crc kubenswrapper[4556]: I0218 09:34:08.720866 4556 scope.go:117] "RemoveContainer" containerID="435a0faca4634ff45a721685b8a0a92b908692540b1d3b176fc9070698963fdf" Feb 18 09:34:08 crc kubenswrapper[4556]: E0218 09:34:08.721307 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"435a0faca4634ff45a721685b8a0a92b908692540b1d3b176fc9070698963fdf\": container with ID starting with 435a0faca4634ff45a721685b8a0a92b908692540b1d3b176fc9070698963fdf not found: ID does not exist" containerID="435a0faca4634ff45a721685b8a0a92b908692540b1d3b176fc9070698963fdf" Feb 18 09:34:08 crc kubenswrapper[4556]: I0218 09:34:08.721348 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"435a0faca4634ff45a721685b8a0a92b908692540b1d3b176fc9070698963fdf"} err="failed to get container status \"435a0faca4634ff45a721685b8a0a92b908692540b1d3b176fc9070698963fdf\": rpc error: code = NotFound desc = could not find container \"435a0faca4634ff45a721685b8a0a92b908692540b1d3b176fc9070698963fdf\": container with ID starting with 435a0faca4634ff45a721685b8a0a92b908692540b1d3b176fc9070698963fdf not found: ID does not exist" Feb 18 09:34:08 crc kubenswrapper[4556]: I0218 09:34:08.721387 4556 scope.go:117] "RemoveContainer" containerID="6e1e2e3be0588f55f963b7c0929f1ec1c3077c8ab94b994bd71bae196c9e8384" Feb 18 09:34:08 crc kubenswrapper[4556]: E0218 09:34:08.722480 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e1e2e3be0588f55f963b7c0929f1ec1c3077c8ab94b994bd71bae196c9e8384\": container with ID starting with 6e1e2e3be0588f55f963b7c0929f1ec1c3077c8ab94b994bd71bae196c9e8384 not found: ID does not exist" containerID="6e1e2e3be0588f55f963b7c0929f1ec1c3077c8ab94b994bd71bae196c9e8384" Feb 18 09:34:08 crc kubenswrapper[4556]: I0218 09:34:08.722504 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e1e2e3be0588f55f963b7c0929f1ec1c3077c8ab94b994bd71bae196c9e8384"} err="failed to get container status \"6e1e2e3be0588f55f963b7c0929f1ec1c3077c8ab94b994bd71bae196c9e8384\": rpc error: code = NotFound desc = could not find container \"6e1e2e3be0588f55f963b7c0929f1ec1c3077c8ab94b994bd71bae196c9e8384\": container with ID starting with 6e1e2e3be0588f55f963b7c0929f1ec1c3077c8ab94b994bd71bae196c9e8384 not found: ID does not exist" Feb 18 09:34:08 crc kubenswrapper[4556]: I0218 09:34:08.722520 4556 scope.go:117] "RemoveContainer" containerID="300423e35fc4e9712c44292607573876b5a74eb3ad5a842f43e4fd315bbb14aa" Feb 18 09:34:08 crc kubenswrapper[4556]: E0218 09:34:08.722791 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"300423e35fc4e9712c44292607573876b5a74eb3ad5a842f43e4fd315bbb14aa\": container with ID starting with 300423e35fc4e9712c44292607573876b5a74eb3ad5a842f43e4fd315bbb14aa not found: ID does not exist" containerID="300423e35fc4e9712c44292607573876b5a74eb3ad5a842f43e4fd315bbb14aa" Feb 18 09:34:08 crc kubenswrapper[4556]: I0218 09:34:08.722832 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"300423e35fc4e9712c44292607573876b5a74eb3ad5a842f43e4fd315bbb14aa"} err="failed to get container status \"300423e35fc4e9712c44292607573876b5a74eb3ad5a842f43e4fd315bbb14aa\": rpc error: code = NotFound desc = could not find container \"300423e35fc4e9712c44292607573876b5a74eb3ad5a842f43e4fd315bbb14aa\": container with ID starting with 300423e35fc4e9712c44292607573876b5a74eb3ad5a842f43e4fd315bbb14aa not found: ID does not exist" Feb 18 09:34:08 crc kubenswrapper[4556]: E0218 09:34:08.803208 4556 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod36a80da9_5d55_45c9_8f91_bcf47149d4a0.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod36a80da9_5d55_45c9_8f91_bcf47149d4a0.slice/crio-e3737665ffd1722039ddf81f88680a3bb032b4061b05537b9ad1abfc97062eee\": RecentStats: unable to find data in memory cache]" Feb 18 09:34:09 crc kubenswrapper[4556]: I0218 09:34:09.293559 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36a80da9-5d55-45c9-8f91-bcf47149d4a0" path="/var/lib/kubelet/pods/36a80da9-5d55-45c9-8f91-bcf47149d4a0/volumes" Feb 18 09:34:31 crc kubenswrapper[4556]: I0218 09:34:31.809901 4556 generic.go:334] "Generic (PLEG): container finished" podID="2cff0286-97e0-422e-b55c-4f28711bde2d" containerID="b5ea566219e7ce686e7970072acca2602e937c928ffd94559395da626656e7fd" exitCode=0 Feb 18 09:34:31 crc kubenswrapper[4556]: I0218 09:34:31.809993 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hhn8t" event={"ID":"2cff0286-97e0-422e-b55c-4f28711bde2d","Type":"ContainerDied","Data":"b5ea566219e7ce686e7970072acca2602e937c928ffd94559395da626656e7fd"} Feb 18 09:34:33 crc kubenswrapper[4556]: I0218 09:34:33.161813 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hhn8t" Feb 18 09:34:33 crc kubenswrapper[4556]: I0218 09:34:33.252581 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2cff0286-97e0-422e-b55c-4f28711bde2d-ssh-key-openstack-edpm-ipam\") pod \"2cff0286-97e0-422e-b55c-4f28711bde2d\" (UID: \"2cff0286-97e0-422e-b55c-4f28711bde2d\") " Feb 18 09:34:33 crc kubenswrapper[4556]: I0218 09:34:33.252749 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lwsqd\" (UniqueName: \"kubernetes.io/projected/2cff0286-97e0-422e-b55c-4f28711bde2d-kube-api-access-lwsqd\") pod \"2cff0286-97e0-422e-b55c-4f28711bde2d\" (UID: \"2cff0286-97e0-422e-b55c-4f28711bde2d\") " Feb 18 09:34:33 crc kubenswrapper[4556]: I0218 09:34:33.252933 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cff0286-97e0-422e-b55c-4f28711bde2d-libvirt-combined-ca-bundle\") pod \"2cff0286-97e0-422e-b55c-4f28711bde2d\" (UID: \"2cff0286-97e0-422e-b55c-4f28711bde2d\") " Feb 18 09:34:33 crc kubenswrapper[4556]: I0218 09:34:33.253007 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2cff0286-97e0-422e-b55c-4f28711bde2d-inventory\") pod \"2cff0286-97e0-422e-b55c-4f28711bde2d\" (UID: \"2cff0286-97e0-422e-b55c-4f28711bde2d\") " Feb 18 09:34:33 crc kubenswrapper[4556]: I0218 09:34:33.253042 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/2cff0286-97e0-422e-b55c-4f28711bde2d-libvirt-secret-0\") pod \"2cff0286-97e0-422e-b55c-4f28711bde2d\" (UID: \"2cff0286-97e0-422e-b55c-4f28711bde2d\") " Feb 18 09:34:33 crc kubenswrapper[4556]: I0218 09:34:33.258107 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2cff0286-97e0-422e-b55c-4f28711bde2d-kube-api-access-lwsqd" (OuterVolumeSpecName: "kube-api-access-lwsqd") pod "2cff0286-97e0-422e-b55c-4f28711bde2d" (UID: "2cff0286-97e0-422e-b55c-4f28711bde2d"). InnerVolumeSpecName "kube-api-access-lwsqd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:34:33 crc kubenswrapper[4556]: I0218 09:34:33.258218 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2cff0286-97e0-422e-b55c-4f28711bde2d-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "2cff0286-97e0-422e-b55c-4f28711bde2d" (UID: "2cff0286-97e0-422e-b55c-4f28711bde2d"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:34:33 crc kubenswrapper[4556]: I0218 09:34:33.277394 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2cff0286-97e0-422e-b55c-4f28711bde2d-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "2cff0286-97e0-422e-b55c-4f28711bde2d" (UID: "2cff0286-97e0-422e-b55c-4f28711bde2d"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:34:33 crc kubenswrapper[4556]: I0218 09:34:33.277444 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2cff0286-97e0-422e-b55c-4f28711bde2d-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "2cff0286-97e0-422e-b55c-4f28711bde2d" (UID: "2cff0286-97e0-422e-b55c-4f28711bde2d"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:34:33 crc kubenswrapper[4556]: I0218 09:34:33.277689 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2cff0286-97e0-422e-b55c-4f28711bde2d-inventory" (OuterVolumeSpecName: "inventory") pod "2cff0286-97e0-422e-b55c-4f28711bde2d" (UID: "2cff0286-97e0-422e-b55c-4f28711bde2d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:34:33 crc kubenswrapper[4556]: I0218 09:34:33.355297 4556 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cff0286-97e0-422e-b55c-4f28711bde2d-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:34:33 crc kubenswrapper[4556]: I0218 09:34:33.355331 4556 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2cff0286-97e0-422e-b55c-4f28711bde2d-inventory\") on node \"crc\" DevicePath \"\"" Feb 18 09:34:33 crc kubenswrapper[4556]: I0218 09:34:33.355344 4556 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/2cff0286-97e0-422e-b55c-4f28711bde2d-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Feb 18 09:34:33 crc kubenswrapper[4556]: I0218 09:34:33.355353 4556 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2cff0286-97e0-422e-b55c-4f28711bde2d-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 18 09:34:33 crc kubenswrapper[4556]: I0218 09:34:33.355363 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lwsqd\" (UniqueName: \"kubernetes.io/projected/2cff0286-97e0-422e-b55c-4f28711bde2d-kube-api-access-lwsqd\") on node \"crc\" DevicePath \"\"" Feb 18 09:34:33 crc kubenswrapper[4556]: I0218 09:34:33.825793 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hhn8t" event={"ID":"2cff0286-97e0-422e-b55c-4f28711bde2d","Type":"ContainerDied","Data":"9294d7863964654bf752b5cb59e6b62c63525bdc4cd3ab26d0afb06e4437bc18"} Feb 18 09:34:33 crc kubenswrapper[4556]: I0218 09:34:33.825832 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9294d7863964654bf752b5cb59e6b62c63525bdc4cd3ab26d0afb06e4437bc18" Feb 18 09:34:33 crc kubenswrapper[4556]: I0218 09:34:33.825856 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hhn8t" Feb 18 09:34:33 crc kubenswrapper[4556]: I0218 09:34:33.891318 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-snrkh"] Feb 18 09:34:33 crc kubenswrapper[4556]: E0218 09:34:33.891709 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cff0286-97e0-422e-b55c-4f28711bde2d" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Feb 18 09:34:33 crc kubenswrapper[4556]: I0218 09:34:33.891726 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cff0286-97e0-422e-b55c-4f28711bde2d" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Feb 18 09:34:33 crc kubenswrapper[4556]: E0218 09:34:33.891739 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36a80da9-5d55-45c9-8f91-bcf47149d4a0" containerName="registry-server" Feb 18 09:34:33 crc kubenswrapper[4556]: I0218 09:34:33.891745 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="36a80da9-5d55-45c9-8f91-bcf47149d4a0" containerName="registry-server" Feb 18 09:34:33 crc kubenswrapper[4556]: E0218 09:34:33.891755 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36a80da9-5d55-45c9-8f91-bcf47149d4a0" containerName="extract-content" Feb 18 09:34:33 crc kubenswrapper[4556]: I0218 09:34:33.891760 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="36a80da9-5d55-45c9-8f91-bcf47149d4a0" containerName="extract-content" Feb 18 09:34:33 crc kubenswrapper[4556]: E0218 09:34:33.891782 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36a80da9-5d55-45c9-8f91-bcf47149d4a0" containerName="extract-utilities" Feb 18 09:34:33 crc kubenswrapper[4556]: I0218 09:34:33.891787 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="36a80da9-5d55-45c9-8f91-bcf47149d4a0" containerName="extract-utilities" Feb 18 09:34:33 crc kubenswrapper[4556]: I0218 09:34:33.891976 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="36a80da9-5d55-45c9-8f91-bcf47149d4a0" containerName="registry-server" Feb 18 09:34:33 crc kubenswrapper[4556]: I0218 09:34:33.891996 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="2cff0286-97e0-422e-b55c-4f28711bde2d" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Feb 18 09:34:33 crc kubenswrapper[4556]: I0218 09:34:33.892620 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-snrkh" Feb 18 09:34:33 crc kubenswrapper[4556]: I0218 09:34:33.894991 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 18 09:34:33 crc kubenswrapper[4556]: I0218 09:34:33.895249 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Feb 18 09:34:33 crc kubenswrapper[4556]: I0218 09:34:33.895373 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 18 09:34:33 crc kubenswrapper[4556]: I0218 09:34:33.895493 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Feb 18 09:34:33 crc kubenswrapper[4556]: I0218 09:34:33.897623 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-54z7h" Feb 18 09:34:33 crc kubenswrapper[4556]: I0218 09:34:33.901778 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-snrkh"] Feb 18 09:34:33 crc kubenswrapper[4556]: I0218 09:34:33.907077 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 18 09:34:33 crc kubenswrapper[4556]: I0218 09:34:33.907853 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Feb 18 09:34:34 crc kubenswrapper[4556]: I0218 09:34:34.068703 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1ab2e8c-4abd-4421-ac68-20821b8ba938-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-snrkh\" (UID: \"a1ab2e8c-4abd-4421-ac68-20821b8ba938\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-snrkh" Feb 18 09:34:34 crc kubenswrapper[4556]: I0218 09:34:34.068844 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a1ab2e8c-4abd-4421-ac68-20821b8ba938-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-snrkh\" (UID: \"a1ab2e8c-4abd-4421-ac68-20821b8ba938\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-snrkh" Feb 18 09:34:34 crc kubenswrapper[4556]: I0218 09:34:34.068873 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/a1ab2e8c-4abd-4421-ac68-20821b8ba938-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-snrkh\" (UID: \"a1ab2e8c-4abd-4421-ac68-20821b8ba938\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-snrkh" Feb 18 09:34:34 crc kubenswrapper[4556]: I0218 09:34:34.069273 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/a1ab2e8c-4abd-4421-ac68-20821b8ba938-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-snrkh\" (UID: \"a1ab2e8c-4abd-4421-ac68-20821b8ba938\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-snrkh" Feb 18 09:34:34 crc kubenswrapper[4556]: I0218 09:34:34.069338 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a1ab2e8c-4abd-4421-ac68-20821b8ba938-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-snrkh\" (UID: \"a1ab2e8c-4abd-4421-ac68-20821b8ba938\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-snrkh" Feb 18 09:34:34 crc kubenswrapper[4556]: I0218 09:34:34.069400 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gglzk\" (UniqueName: \"kubernetes.io/projected/a1ab2e8c-4abd-4421-ac68-20821b8ba938-kube-api-access-gglzk\") pod \"nova-edpm-deployment-openstack-edpm-ipam-snrkh\" (UID: \"a1ab2e8c-4abd-4421-ac68-20821b8ba938\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-snrkh" Feb 18 09:34:34 crc kubenswrapper[4556]: I0218 09:34:34.069486 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/a1ab2e8c-4abd-4421-ac68-20821b8ba938-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-snrkh\" (UID: \"a1ab2e8c-4abd-4421-ac68-20821b8ba938\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-snrkh" Feb 18 09:34:34 crc kubenswrapper[4556]: I0218 09:34:34.069520 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/a1ab2e8c-4abd-4421-ac68-20821b8ba938-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-snrkh\" (UID: \"a1ab2e8c-4abd-4421-ac68-20821b8ba938\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-snrkh" Feb 18 09:34:34 crc kubenswrapper[4556]: I0218 09:34:34.069783 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/a1ab2e8c-4abd-4421-ac68-20821b8ba938-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-snrkh\" (UID: \"a1ab2e8c-4abd-4421-ac68-20821b8ba938\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-snrkh" Feb 18 09:34:34 crc kubenswrapper[4556]: I0218 09:34:34.069850 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/a1ab2e8c-4abd-4421-ac68-20821b8ba938-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-snrkh\" (UID: \"a1ab2e8c-4abd-4421-ac68-20821b8ba938\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-snrkh" Feb 18 09:34:34 crc kubenswrapper[4556]: I0218 09:34:34.069992 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/a1ab2e8c-4abd-4421-ac68-20821b8ba938-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-snrkh\" (UID: \"a1ab2e8c-4abd-4421-ac68-20821b8ba938\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-snrkh" Feb 18 09:34:34 crc kubenswrapper[4556]: I0218 09:34:34.172399 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1ab2e8c-4abd-4421-ac68-20821b8ba938-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-snrkh\" (UID: \"a1ab2e8c-4abd-4421-ac68-20821b8ba938\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-snrkh" Feb 18 09:34:34 crc kubenswrapper[4556]: I0218 09:34:34.172504 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a1ab2e8c-4abd-4421-ac68-20821b8ba938-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-snrkh\" (UID: \"a1ab2e8c-4abd-4421-ac68-20821b8ba938\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-snrkh" Feb 18 09:34:34 crc kubenswrapper[4556]: I0218 09:34:34.172548 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/a1ab2e8c-4abd-4421-ac68-20821b8ba938-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-snrkh\" (UID: \"a1ab2e8c-4abd-4421-ac68-20821b8ba938\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-snrkh" Feb 18 09:34:34 crc kubenswrapper[4556]: I0218 09:34:34.172757 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/a1ab2e8c-4abd-4421-ac68-20821b8ba938-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-snrkh\" (UID: \"a1ab2e8c-4abd-4421-ac68-20821b8ba938\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-snrkh" Feb 18 09:34:34 crc kubenswrapper[4556]: I0218 09:34:34.172813 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a1ab2e8c-4abd-4421-ac68-20821b8ba938-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-snrkh\" (UID: \"a1ab2e8c-4abd-4421-ac68-20821b8ba938\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-snrkh" Feb 18 09:34:34 crc kubenswrapper[4556]: I0218 09:34:34.172999 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gglzk\" (UniqueName: \"kubernetes.io/projected/a1ab2e8c-4abd-4421-ac68-20821b8ba938-kube-api-access-gglzk\") pod \"nova-edpm-deployment-openstack-edpm-ipam-snrkh\" (UID: \"a1ab2e8c-4abd-4421-ac68-20821b8ba938\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-snrkh" Feb 18 09:34:34 crc kubenswrapper[4556]: I0218 09:34:34.173214 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/a1ab2e8c-4abd-4421-ac68-20821b8ba938-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-snrkh\" (UID: \"a1ab2e8c-4abd-4421-ac68-20821b8ba938\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-snrkh" Feb 18 09:34:34 crc kubenswrapper[4556]: I0218 09:34:34.173703 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/a1ab2e8c-4abd-4421-ac68-20821b8ba938-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-snrkh\" (UID: \"a1ab2e8c-4abd-4421-ac68-20821b8ba938\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-snrkh" Feb 18 09:34:34 crc kubenswrapper[4556]: I0218 09:34:34.173757 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/a1ab2e8c-4abd-4421-ac68-20821b8ba938-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-snrkh\" (UID: \"a1ab2e8c-4abd-4421-ac68-20821b8ba938\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-snrkh" Feb 18 09:34:34 crc kubenswrapper[4556]: I0218 09:34:34.173812 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/a1ab2e8c-4abd-4421-ac68-20821b8ba938-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-snrkh\" (UID: \"a1ab2e8c-4abd-4421-ac68-20821b8ba938\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-snrkh" Feb 18 09:34:34 crc kubenswrapper[4556]: I0218 09:34:34.173906 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/a1ab2e8c-4abd-4421-ac68-20821b8ba938-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-snrkh\" (UID: \"a1ab2e8c-4abd-4421-ac68-20821b8ba938\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-snrkh" Feb 18 09:34:34 crc kubenswrapper[4556]: I0218 09:34:34.174909 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/a1ab2e8c-4abd-4421-ac68-20821b8ba938-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-snrkh\" (UID: \"a1ab2e8c-4abd-4421-ac68-20821b8ba938\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-snrkh" Feb 18 09:34:34 crc kubenswrapper[4556]: I0218 09:34:34.176633 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/a1ab2e8c-4abd-4421-ac68-20821b8ba938-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-snrkh\" (UID: \"a1ab2e8c-4abd-4421-ac68-20821b8ba938\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-snrkh" Feb 18 09:34:34 crc kubenswrapper[4556]: I0218 09:34:34.176699 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a1ab2e8c-4abd-4421-ac68-20821b8ba938-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-snrkh\" (UID: \"a1ab2e8c-4abd-4421-ac68-20821b8ba938\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-snrkh" Feb 18 09:34:34 crc kubenswrapper[4556]: I0218 09:34:34.177041 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1ab2e8c-4abd-4421-ac68-20821b8ba938-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-snrkh\" (UID: \"a1ab2e8c-4abd-4421-ac68-20821b8ba938\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-snrkh" Feb 18 09:34:34 crc kubenswrapper[4556]: I0218 09:34:34.177315 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/a1ab2e8c-4abd-4421-ac68-20821b8ba938-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-snrkh\" (UID: \"a1ab2e8c-4abd-4421-ac68-20821b8ba938\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-snrkh" Feb 18 09:34:34 crc kubenswrapper[4556]: I0218 09:34:34.177456 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a1ab2e8c-4abd-4421-ac68-20821b8ba938-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-snrkh\" (UID: \"a1ab2e8c-4abd-4421-ac68-20821b8ba938\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-snrkh" Feb 18 09:34:34 crc kubenswrapper[4556]: I0218 09:34:34.177595 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/a1ab2e8c-4abd-4421-ac68-20821b8ba938-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-snrkh\" (UID: \"a1ab2e8c-4abd-4421-ac68-20821b8ba938\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-snrkh" Feb 18 09:34:34 crc kubenswrapper[4556]: I0218 09:34:34.178445 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/a1ab2e8c-4abd-4421-ac68-20821b8ba938-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-snrkh\" (UID: \"a1ab2e8c-4abd-4421-ac68-20821b8ba938\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-snrkh" Feb 18 09:34:34 crc kubenswrapper[4556]: I0218 09:34:34.178483 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/a1ab2e8c-4abd-4421-ac68-20821b8ba938-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-snrkh\" (UID: \"a1ab2e8c-4abd-4421-ac68-20821b8ba938\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-snrkh" Feb 18 09:34:34 crc kubenswrapper[4556]: I0218 09:34:34.178527 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/a1ab2e8c-4abd-4421-ac68-20821b8ba938-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-snrkh\" (UID: \"a1ab2e8c-4abd-4421-ac68-20821b8ba938\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-snrkh" Feb 18 09:34:34 crc kubenswrapper[4556]: I0218 09:34:34.186329 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gglzk\" (UniqueName: \"kubernetes.io/projected/a1ab2e8c-4abd-4421-ac68-20821b8ba938-kube-api-access-gglzk\") pod \"nova-edpm-deployment-openstack-edpm-ipam-snrkh\" (UID: \"a1ab2e8c-4abd-4421-ac68-20821b8ba938\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-snrkh" Feb 18 09:34:34 crc kubenswrapper[4556]: I0218 09:34:34.209463 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-snrkh" Feb 18 09:34:34 crc kubenswrapper[4556]: I0218 09:34:34.652664 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-snrkh"] Feb 18 09:34:34 crc kubenswrapper[4556]: I0218 09:34:34.836637 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-snrkh" event={"ID":"a1ab2e8c-4abd-4421-ac68-20821b8ba938","Type":"ContainerStarted","Data":"efaa705a27e148bb71a8f8e0c0adcd515fa83d03bc8a9747f3906221d3d24044"} Feb 18 09:34:35 crc kubenswrapper[4556]: I0218 09:34:35.846406 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-snrkh" event={"ID":"a1ab2e8c-4abd-4421-ac68-20821b8ba938","Type":"ContainerStarted","Data":"a62b0c0d226dbf45ac37340548e9c0d17948b199cb999b0479fe3263e503add1"} Feb 18 09:34:35 crc kubenswrapper[4556]: I0218 09:34:35.865011 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-snrkh" podStartSLOduration=2.351057557 podStartE2EDuration="2.864997612s" podCreationTimestamp="2026-02-18 09:34:33 +0000 UTC" firstStartedPulling="2026-02-18 09:34:34.654464276 +0000 UTC m=+1831.671425257" lastFinishedPulling="2026-02-18 09:34:35.168404332 +0000 UTC m=+1832.185365312" observedRunningTime="2026-02-18 09:34:35.862491896 +0000 UTC m=+1832.879452876" watchObservedRunningTime="2026-02-18 09:34:35.864997612 +0000 UTC m=+1832.881958593" Feb 18 09:36:20 crc kubenswrapper[4556]: I0218 09:36:20.612887 4556 generic.go:334] "Generic (PLEG): container finished" podID="a1ab2e8c-4abd-4421-ac68-20821b8ba938" containerID="a62b0c0d226dbf45ac37340548e9c0d17948b199cb999b0479fe3263e503add1" exitCode=0 Feb 18 09:36:20 crc kubenswrapper[4556]: I0218 09:36:20.612960 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-snrkh" event={"ID":"a1ab2e8c-4abd-4421-ac68-20821b8ba938","Type":"ContainerDied","Data":"a62b0c0d226dbf45ac37340548e9c0d17948b199cb999b0479fe3263e503add1"} Feb 18 09:36:21 crc kubenswrapper[4556]: I0218 09:36:21.957772 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-snrkh" Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.024443 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gglzk\" (UniqueName: \"kubernetes.io/projected/a1ab2e8c-4abd-4421-ac68-20821b8ba938-kube-api-access-gglzk\") pod \"a1ab2e8c-4abd-4421-ac68-20821b8ba938\" (UID: \"a1ab2e8c-4abd-4421-ac68-20821b8ba938\") " Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.024499 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/a1ab2e8c-4abd-4421-ac68-20821b8ba938-nova-migration-ssh-key-0\") pod \"a1ab2e8c-4abd-4421-ac68-20821b8ba938\" (UID: \"a1ab2e8c-4abd-4421-ac68-20821b8ba938\") " Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.024559 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1ab2e8c-4abd-4421-ac68-20821b8ba938-nova-combined-ca-bundle\") pod \"a1ab2e8c-4abd-4421-ac68-20821b8ba938\" (UID: \"a1ab2e8c-4abd-4421-ac68-20821b8ba938\") " Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.024640 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/a1ab2e8c-4abd-4421-ac68-20821b8ba938-nova-cell1-compute-config-1\") pod \"a1ab2e8c-4abd-4421-ac68-20821b8ba938\" (UID: \"a1ab2e8c-4abd-4421-ac68-20821b8ba938\") " Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.024694 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a1ab2e8c-4abd-4421-ac68-20821b8ba938-inventory\") pod \"a1ab2e8c-4abd-4421-ac68-20821b8ba938\" (UID: \"a1ab2e8c-4abd-4421-ac68-20821b8ba938\") " Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.024738 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/a1ab2e8c-4abd-4421-ac68-20821b8ba938-nova-migration-ssh-key-1\") pod \"a1ab2e8c-4abd-4421-ac68-20821b8ba938\" (UID: \"a1ab2e8c-4abd-4421-ac68-20821b8ba938\") " Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.024769 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/a1ab2e8c-4abd-4421-ac68-20821b8ba938-nova-cell1-compute-config-2\") pod \"a1ab2e8c-4abd-4421-ac68-20821b8ba938\" (UID: \"a1ab2e8c-4abd-4421-ac68-20821b8ba938\") " Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.024792 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/a1ab2e8c-4abd-4421-ac68-20821b8ba938-nova-extra-config-0\") pod \"a1ab2e8c-4abd-4421-ac68-20821b8ba938\" (UID: \"a1ab2e8c-4abd-4421-ac68-20821b8ba938\") " Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.024836 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a1ab2e8c-4abd-4421-ac68-20821b8ba938-ssh-key-openstack-edpm-ipam\") pod \"a1ab2e8c-4abd-4421-ac68-20821b8ba938\" (UID: \"a1ab2e8c-4abd-4421-ac68-20821b8ba938\") " Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.024878 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/a1ab2e8c-4abd-4421-ac68-20821b8ba938-nova-cell1-compute-config-3\") pod \"a1ab2e8c-4abd-4421-ac68-20821b8ba938\" (UID: \"a1ab2e8c-4abd-4421-ac68-20821b8ba938\") " Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.024907 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/a1ab2e8c-4abd-4421-ac68-20821b8ba938-nova-cell1-compute-config-0\") pod \"a1ab2e8c-4abd-4421-ac68-20821b8ba938\" (UID: \"a1ab2e8c-4abd-4421-ac68-20821b8ba938\") " Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.031781 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1ab2e8c-4abd-4421-ac68-20821b8ba938-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "a1ab2e8c-4abd-4421-ac68-20821b8ba938" (UID: "a1ab2e8c-4abd-4421-ac68-20821b8ba938"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.046319 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1ab2e8c-4abd-4421-ac68-20821b8ba938-kube-api-access-gglzk" (OuterVolumeSpecName: "kube-api-access-gglzk") pod "a1ab2e8c-4abd-4421-ac68-20821b8ba938" (UID: "a1ab2e8c-4abd-4421-ac68-20821b8ba938"). InnerVolumeSpecName "kube-api-access-gglzk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.051725 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1ab2e8c-4abd-4421-ac68-20821b8ba938-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "a1ab2e8c-4abd-4421-ac68-20821b8ba938" (UID: "a1ab2e8c-4abd-4421-ac68-20821b8ba938"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.059565 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1ab2e8c-4abd-4421-ac68-20821b8ba938-nova-cell1-compute-config-2" (OuterVolumeSpecName: "nova-cell1-compute-config-2") pod "a1ab2e8c-4abd-4421-ac68-20821b8ba938" (UID: "a1ab2e8c-4abd-4421-ac68-20821b8ba938"). InnerVolumeSpecName "nova-cell1-compute-config-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.059985 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1ab2e8c-4abd-4421-ac68-20821b8ba938-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "a1ab2e8c-4abd-4421-ac68-20821b8ba938" (UID: "a1ab2e8c-4abd-4421-ac68-20821b8ba938"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.060696 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1ab2e8c-4abd-4421-ac68-20821b8ba938-nova-cell1-compute-config-3" (OuterVolumeSpecName: "nova-cell1-compute-config-3") pod "a1ab2e8c-4abd-4421-ac68-20821b8ba938" (UID: "a1ab2e8c-4abd-4421-ac68-20821b8ba938"). InnerVolumeSpecName "nova-cell1-compute-config-3". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.062923 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1ab2e8c-4abd-4421-ac68-20821b8ba938-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "a1ab2e8c-4abd-4421-ac68-20821b8ba938" (UID: "a1ab2e8c-4abd-4421-ac68-20821b8ba938"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.062991 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1ab2e8c-4abd-4421-ac68-20821b8ba938-inventory" (OuterVolumeSpecName: "inventory") pod "a1ab2e8c-4abd-4421-ac68-20821b8ba938" (UID: "a1ab2e8c-4abd-4421-ac68-20821b8ba938"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.064969 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1ab2e8c-4abd-4421-ac68-20821b8ba938-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "a1ab2e8c-4abd-4421-ac68-20821b8ba938" (UID: "a1ab2e8c-4abd-4421-ac68-20821b8ba938"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.066688 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1ab2e8c-4abd-4421-ac68-20821b8ba938-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "a1ab2e8c-4abd-4421-ac68-20821b8ba938" (UID: "a1ab2e8c-4abd-4421-ac68-20821b8ba938"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.074850 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1ab2e8c-4abd-4421-ac68-20821b8ba938-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "a1ab2e8c-4abd-4421-ac68-20821b8ba938" (UID: "a1ab2e8c-4abd-4421-ac68-20821b8ba938"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.127391 4556 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a1ab2e8c-4abd-4421-ac68-20821b8ba938-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.127422 4556 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/a1ab2e8c-4abd-4421-ac68-20821b8ba938-nova-cell1-compute-config-3\") on node \"crc\" DevicePath \"\"" Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.127432 4556 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/a1ab2e8c-4abd-4421-ac68-20821b8ba938-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.127442 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gglzk\" (UniqueName: \"kubernetes.io/projected/a1ab2e8c-4abd-4421-ac68-20821b8ba938-kube-api-access-gglzk\") on node \"crc\" DevicePath \"\"" Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.127452 4556 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/a1ab2e8c-4abd-4421-ac68-20821b8ba938-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.127461 4556 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1ab2e8c-4abd-4421-ac68-20821b8ba938-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.127469 4556 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/a1ab2e8c-4abd-4421-ac68-20821b8ba938-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.127481 4556 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a1ab2e8c-4abd-4421-ac68-20821b8ba938-inventory\") on node \"crc\" DevicePath \"\"" Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.127492 4556 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/a1ab2e8c-4abd-4421-ac68-20821b8ba938-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.127501 4556 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/a1ab2e8c-4abd-4421-ac68-20821b8ba938-nova-cell1-compute-config-2\") on node \"crc\" DevicePath \"\"" Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.127512 4556 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/a1ab2e8c-4abd-4421-ac68-20821b8ba938-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.632232 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-snrkh" event={"ID":"a1ab2e8c-4abd-4421-ac68-20821b8ba938","Type":"ContainerDied","Data":"efaa705a27e148bb71a8f8e0c0adcd515fa83d03bc8a9747f3906221d3d24044"} Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.632547 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="efaa705a27e148bb71a8f8e0c0adcd515fa83d03bc8a9747f3906221d3d24044" Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.632291 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-snrkh" Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.741595 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-l92vs"] Feb 18 09:36:22 crc kubenswrapper[4556]: E0218 09:36:22.742065 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1ab2e8c-4abd-4421-ac68-20821b8ba938" containerName="nova-edpm-deployment-openstack-edpm-ipam" Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.742086 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1ab2e8c-4abd-4421-ac68-20821b8ba938" containerName="nova-edpm-deployment-openstack-edpm-ipam" Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.742355 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1ab2e8c-4abd-4421-ac68-20821b8ba938" containerName="nova-edpm-deployment-openstack-edpm-ipam" Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.743014 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-l92vs" Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.745274 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.745624 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.745816 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.746224 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.746587 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-54z7h" Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.757372 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-l92vs"] Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.842989 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/8ea53a2a-db68-43fb-98c4-26b9ca9f816d-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-l92vs\" (UID: \"8ea53a2a-db68-43fb-98c4-26b9ca9f816d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-l92vs" Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.843237 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8ea53a2a-db68-43fb-98c4-26b9ca9f816d-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-l92vs\" (UID: \"8ea53a2a-db68-43fb-98c4-26b9ca9f816d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-l92vs" Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.843434 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8ea53a2a-db68-43fb-98c4-26b9ca9f816d-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-l92vs\" (UID: \"8ea53a2a-db68-43fb-98c4-26b9ca9f816d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-l92vs" Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.843608 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/8ea53a2a-db68-43fb-98c4-26b9ca9f816d-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-l92vs\" (UID: \"8ea53a2a-db68-43fb-98c4-26b9ca9f816d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-l92vs" Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.843686 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ea53a2a-db68-43fb-98c4-26b9ca9f816d-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-l92vs\" (UID: \"8ea53a2a-db68-43fb-98c4-26b9ca9f816d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-l92vs" Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.843897 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68jdc\" (UniqueName: \"kubernetes.io/projected/8ea53a2a-db68-43fb-98c4-26b9ca9f816d-kube-api-access-68jdc\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-l92vs\" (UID: \"8ea53a2a-db68-43fb-98c4-26b9ca9f816d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-l92vs" Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.844170 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/8ea53a2a-db68-43fb-98c4-26b9ca9f816d-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-l92vs\" (UID: \"8ea53a2a-db68-43fb-98c4-26b9ca9f816d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-l92vs" Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.946254 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/8ea53a2a-db68-43fb-98c4-26b9ca9f816d-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-l92vs\" (UID: \"8ea53a2a-db68-43fb-98c4-26b9ca9f816d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-l92vs" Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.946320 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8ea53a2a-db68-43fb-98c4-26b9ca9f816d-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-l92vs\" (UID: \"8ea53a2a-db68-43fb-98c4-26b9ca9f816d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-l92vs" Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.946373 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8ea53a2a-db68-43fb-98c4-26b9ca9f816d-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-l92vs\" (UID: \"8ea53a2a-db68-43fb-98c4-26b9ca9f816d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-l92vs" Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.946438 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/8ea53a2a-db68-43fb-98c4-26b9ca9f816d-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-l92vs\" (UID: \"8ea53a2a-db68-43fb-98c4-26b9ca9f816d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-l92vs" Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.946460 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ea53a2a-db68-43fb-98c4-26b9ca9f816d-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-l92vs\" (UID: \"8ea53a2a-db68-43fb-98c4-26b9ca9f816d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-l92vs" Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.946512 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68jdc\" (UniqueName: \"kubernetes.io/projected/8ea53a2a-db68-43fb-98c4-26b9ca9f816d-kube-api-access-68jdc\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-l92vs\" (UID: \"8ea53a2a-db68-43fb-98c4-26b9ca9f816d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-l92vs" Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.946599 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/8ea53a2a-db68-43fb-98c4-26b9ca9f816d-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-l92vs\" (UID: \"8ea53a2a-db68-43fb-98c4-26b9ca9f816d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-l92vs" Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.952435 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/8ea53a2a-db68-43fb-98c4-26b9ca9f816d-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-l92vs\" (UID: \"8ea53a2a-db68-43fb-98c4-26b9ca9f816d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-l92vs" Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.952787 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8ea53a2a-db68-43fb-98c4-26b9ca9f816d-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-l92vs\" (UID: \"8ea53a2a-db68-43fb-98c4-26b9ca9f816d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-l92vs" Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.953876 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/8ea53a2a-db68-43fb-98c4-26b9ca9f816d-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-l92vs\" (UID: \"8ea53a2a-db68-43fb-98c4-26b9ca9f816d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-l92vs" Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.954020 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ea53a2a-db68-43fb-98c4-26b9ca9f816d-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-l92vs\" (UID: \"8ea53a2a-db68-43fb-98c4-26b9ca9f816d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-l92vs" Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.954180 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8ea53a2a-db68-43fb-98c4-26b9ca9f816d-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-l92vs\" (UID: \"8ea53a2a-db68-43fb-98c4-26b9ca9f816d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-l92vs" Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.954570 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/8ea53a2a-db68-43fb-98c4-26b9ca9f816d-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-l92vs\" (UID: \"8ea53a2a-db68-43fb-98c4-26b9ca9f816d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-l92vs" Feb 18 09:36:22 crc kubenswrapper[4556]: I0218 09:36:22.966026 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68jdc\" (UniqueName: \"kubernetes.io/projected/8ea53a2a-db68-43fb-98c4-26b9ca9f816d-kube-api-access-68jdc\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-l92vs\" (UID: \"8ea53a2a-db68-43fb-98c4-26b9ca9f816d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-l92vs" Feb 18 09:36:23 crc kubenswrapper[4556]: I0218 09:36:23.057636 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-l92vs" Feb 18 09:36:23 crc kubenswrapper[4556]: I0218 09:36:23.512078 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-l92vs"] Feb 18 09:36:23 crc kubenswrapper[4556]: I0218 09:36:23.642138 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-l92vs" event={"ID":"8ea53a2a-db68-43fb-98c4-26b9ca9f816d","Type":"ContainerStarted","Data":"2fffdf7a2da7c447a0d0e5d1fddd1de1c3be27b283628d06976d17b9032721f1"} Feb 18 09:36:24 crc kubenswrapper[4556]: I0218 09:36:24.651447 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-l92vs" event={"ID":"8ea53a2a-db68-43fb-98c4-26b9ca9f816d","Type":"ContainerStarted","Data":"de71f07e16059c52e1d7b6ea2881a7bd5f6c765e11f606957a1a96a7bc99e10f"} Feb 18 09:36:24 crc kubenswrapper[4556]: I0218 09:36:24.671628 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-l92vs" podStartSLOduration=2.06215837 podStartE2EDuration="2.671610879s" podCreationTimestamp="2026-02-18 09:36:22 +0000 UTC" firstStartedPulling="2026-02-18 09:36:23.51332355 +0000 UTC m=+1940.530284530" lastFinishedPulling="2026-02-18 09:36:24.122776059 +0000 UTC m=+1941.139737039" observedRunningTime="2026-02-18 09:36:24.665494822 +0000 UTC m=+1941.682455801" watchObservedRunningTime="2026-02-18 09:36:24.671610879 +0000 UTC m=+1941.688571859" Feb 18 09:36:31 crc kubenswrapper[4556]: I0218 09:36:31.727287 4556 patch_prober.go:28] interesting pod/machine-config-daemon-f76hs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 09:36:31 crc kubenswrapper[4556]: I0218 09:36:31.728139 4556 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 09:37:01 crc kubenswrapper[4556]: I0218 09:37:01.728311 4556 patch_prober.go:28] interesting pod/machine-config-daemon-f76hs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 09:37:01 crc kubenswrapper[4556]: I0218 09:37:01.729287 4556 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 09:37:31 crc kubenswrapper[4556]: I0218 09:37:31.727306 4556 patch_prober.go:28] interesting pod/machine-config-daemon-f76hs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 09:37:31 crc kubenswrapper[4556]: I0218 09:37:31.727898 4556 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 09:37:31 crc kubenswrapper[4556]: I0218 09:37:31.728180 4556 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" Feb 18 09:37:31 crc kubenswrapper[4556]: I0218 09:37:31.728765 4556 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fd40a6d72be7a9ed07291f69c89aeb615c5bf9535dc9706ff886f8f14c11f976"} pod="openshift-machine-config-operator/machine-config-daemon-f76hs" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 18 09:37:31 crc kubenswrapper[4556]: I0218 09:37:31.728827 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerName="machine-config-daemon" containerID="cri-o://fd40a6d72be7a9ed07291f69c89aeb615c5bf9535dc9706ff886f8f14c11f976" gracePeriod=600 Feb 18 09:37:32 crc kubenswrapper[4556]: I0218 09:37:32.267428 4556 generic.go:334] "Generic (PLEG): container finished" podID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerID="fd40a6d72be7a9ed07291f69c89aeb615c5bf9535dc9706ff886f8f14c11f976" exitCode=0 Feb 18 09:37:32 crc kubenswrapper[4556]: I0218 09:37:32.267533 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" event={"ID":"8dac7f27-d3d1-4778-9e54-f273035a1d37","Type":"ContainerDied","Data":"fd40a6d72be7a9ed07291f69c89aeb615c5bf9535dc9706ff886f8f14c11f976"} Feb 18 09:37:32 crc kubenswrapper[4556]: I0218 09:37:32.267871 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" event={"ID":"8dac7f27-d3d1-4778-9e54-f273035a1d37","Type":"ContainerStarted","Data":"04a4196781cc895fa1780ee09f471180228eaaff00e4cc56bf534ca86fb1d640"} Feb 18 09:37:32 crc kubenswrapper[4556]: I0218 09:37:32.267899 4556 scope.go:117] "RemoveContainer" containerID="507a9134b6862bc8be1ac03b7993c890ef64f16c21975e00aa96c53fe413415d" Feb 18 09:38:11 crc kubenswrapper[4556]: I0218 09:38:11.630766 4556 generic.go:334] "Generic (PLEG): container finished" podID="8ea53a2a-db68-43fb-98c4-26b9ca9f816d" containerID="de71f07e16059c52e1d7b6ea2881a7bd5f6c765e11f606957a1a96a7bc99e10f" exitCode=0 Feb 18 09:38:11 crc kubenswrapper[4556]: I0218 09:38:11.630849 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-l92vs" event={"ID":"8ea53a2a-db68-43fb-98c4-26b9ca9f816d","Type":"ContainerDied","Data":"de71f07e16059c52e1d7b6ea2881a7bd5f6c765e11f606957a1a96a7bc99e10f"} Feb 18 09:38:12 crc kubenswrapper[4556]: I0218 09:38:12.962636 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-l92vs" Feb 18 09:38:13 crc kubenswrapper[4556]: I0218 09:38:13.024272 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/8ea53a2a-db68-43fb-98c4-26b9ca9f816d-ceilometer-compute-config-data-1\") pod \"8ea53a2a-db68-43fb-98c4-26b9ca9f816d\" (UID: \"8ea53a2a-db68-43fb-98c4-26b9ca9f816d\") " Feb 18 09:38:13 crc kubenswrapper[4556]: I0218 09:38:13.024337 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/8ea53a2a-db68-43fb-98c4-26b9ca9f816d-ceilometer-compute-config-data-2\") pod \"8ea53a2a-db68-43fb-98c4-26b9ca9f816d\" (UID: \"8ea53a2a-db68-43fb-98c4-26b9ca9f816d\") " Feb 18 09:38:13 crc kubenswrapper[4556]: I0218 09:38:13.024373 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ea53a2a-db68-43fb-98c4-26b9ca9f816d-telemetry-combined-ca-bundle\") pod \"8ea53a2a-db68-43fb-98c4-26b9ca9f816d\" (UID: \"8ea53a2a-db68-43fb-98c4-26b9ca9f816d\") " Feb 18 09:38:13 crc kubenswrapper[4556]: I0218 09:38:13.024398 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-68jdc\" (UniqueName: \"kubernetes.io/projected/8ea53a2a-db68-43fb-98c4-26b9ca9f816d-kube-api-access-68jdc\") pod \"8ea53a2a-db68-43fb-98c4-26b9ca9f816d\" (UID: \"8ea53a2a-db68-43fb-98c4-26b9ca9f816d\") " Feb 18 09:38:13 crc kubenswrapper[4556]: I0218 09:38:13.024419 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8ea53a2a-db68-43fb-98c4-26b9ca9f816d-ssh-key-openstack-edpm-ipam\") pod \"8ea53a2a-db68-43fb-98c4-26b9ca9f816d\" (UID: \"8ea53a2a-db68-43fb-98c4-26b9ca9f816d\") " Feb 18 09:38:13 crc kubenswrapper[4556]: I0218 09:38:13.024452 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/8ea53a2a-db68-43fb-98c4-26b9ca9f816d-ceilometer-compute-config-data-0\") pod \"8ea53a2a-db68-43fb-98c4-26b9ca9f816d\" (UID: \"8ea53a2a-db68-43fb-98c4-26b9ca9f816d\") " Feb 18 09:38:13 crc kubenswrapper[4556]: I0218 09:38:13.024625 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8ea53a2a-db68-43fb-98c4-26b9ca9f816d-inventory\") pod \"8ea53a2a-db68-43fb-98c4-26b9ca9f816d\" (UID: \"8ea53a2a-db68-43fb-98c4-26b9ca9f816d\") " Feb 18 09:38:13 crc kubenswrapper[4556]: I0218 09:38:13.043687 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ea53a2a-db68-43fb-98c4-26b9ca9f816d-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "8ea53a2a-db68-43fb-98c4-26b9ca9f816d" (UID: "8ea53a2a-db68-43fb-98c4-26b9ca9f816d"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:38:13 crc kubenswrapper[4556]: I0218 09:38:13.043696 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ea53a2a-db68-43fb-98c4-26b9ca9f816d-kube-api-access-68jdc" (OuterVolumeSpecName: "kube-api-access-68jdc") pod "8ea53a2a-db68-43fb-98c4-26b9ca9f816d" (UID: "8ea53a2a-db68-43fb-98c4-26b9ca9f816d"). InnerVolumeSpecName "kube-api-access-68jdc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:38:13 crc kubenswrapper[4556]: I0218 09:38:13.047883 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ea53a2a-db68-43fb-98c4-26b9ca9f816d-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "8ea53a2a-db68-43fb-98c4-26b9ca9f816d" (UID: "8ea53a2a-db68-43fb-98c4-26b9ca9f816d"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:38:13 crc kubenswrapper[4556]: I0218 09:38:13.048828 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ea53a2a-db68-43fb-98c4-26b9ca9f816d-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "8ea53a2a-db68-43fb-98c4-26b9ca9f816d" (UID: "8ea53a2a-db68-43fb-98c4-26b9ca9f816d"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:38:13 crc kubenswrapper[4556]: I0218 09:38:13.048846 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ea53a2a-db68-43fb-98c4-26b9ca9f816d-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "8ea53a2a-db68-43fb-98c4-26b9ca9f816d" (UID: "8ea53a2a-db68-43fb-98c4-26b9ca9f816d"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:38:13 crc kubenswrapper[4556]: I0218 09:38:13.049224 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ea53a2a-db68-43fb-98c4-26b9ca9f816d-inventory" (OuterVolumeSpecName: "inventory") pod "8ea53a2a-db68-43fb-98c4-26b9ca9f816d" (UID: "8ea53a2a-db68-43fb-98c4-26b9ca9f816d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:38:13 crc kubenswrapper[4556]: I0218 09:38:13.049624 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ea53a2a-db68-43fb-98c4-26b9ca9f816d-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "8ea53a2a-db68-43fb-98c4-26b9ca9f816d" (UID: "8ea53a2a-db68-43fb-98c4-26b9ca9f816d"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:38:13 crc kubenswrapper[4556]: I0218 09:38:13.127755 4556 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8ea53a2a-db68-43fb-98c4-26b9ca9f816d-inventory\") on node \"crc\" DevicePath \"\"" Feb 18 09:38:13 crc kubenswrapper[4556]: I0218 09:38:13.127795 4556 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/8ea53a2a-db68-43fb-98c4-26b9ca9f816d-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Feb 18 09:38:13 crc kubenswrapper[4556]: I0218 09:38:13.127809 4556 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/8ea53a2a-db68-43fb-98c4-26b9ca9f816d-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Feb 18 09:38:13 crc kubenswrapper[4556]: I0218 09:38:13.127822 4556 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ea53a2a-db68-43fb-98c4-26b9ca9f816d-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 09:38:13 crc kubenswrapper[4556]: I0218 09:38:13.127834 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-68jdc\" (UniqueName: \"kubernetes.io/projected/8ea53a2a-db68-43fb-98c4-26b9ca9f816d-kube-api-access-68jdc\") on node \"crc\" DevicePath \"\"" Feb 18 09:38:13 crc kubenswrapper[4556]: I0218 09:38:13.127844 4556 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8ea53a2a-db68-43fb-98c4-26b9ca9f816d-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 18 09:38:13 crc kubenswrapper[4556]: I0218 09:38:13.127855 4556 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/8ea53a2a-db68-43fb-98c4-26b9ca9f816d-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Feb 18 09:38:13 crc kubenswrapper[4556]: I0218 09:38:13.658300 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-l92vs" event={"ID":"8ea53a2a-db68-43fb-98c4-26b9ca9f816d","Type":"ContainerDied","Data":"2fffdf7a2da7c447a0d0e5d1fddd1de1c3be27b283628d06976d17b9032721f1"} Feb 18 09:38:13 crc kubenswrapper[4556]: I0218 09:38:13.658361 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-l92vs" Feb 18 09:38:13 crc kubenswrapper[4556]: I0218 09:38:13.658376 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2fffdf7a2da7c447a0d0e5d1fddd1de1c3be27b283628d06976d17b9032721f1" Feb 18 09:38:59 crc kubenswrapper[4556]: I0218 09:38:59.842196 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Feb 18 09:38:59 crc kubenswrapper[4556]: E0218 09:38:59.843390 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ea53a2a-db68-43fb-98c4-26b9ca9f816d" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Feb 18 09:38:59 crc kubenswrapper[4556]: I0218 09:38:59.843411 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ea53a2a-db68-43fb-98c4-26b9ca9f816d" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Feb 18 09:38:59 crc kubenswrapper[4556]: I0218 09:38:59.843695 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ea53a2a-db68-43fb-98c4-26b9ca9f816d" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Feb 18 09:38:59 crc kubenswrapper[4556]: I0218 09:38:59.844443 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Feb 18 09:38:59 crc kubenswrapper[4556]: I0218 09:38:59.846223 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Feb 18 09:38:59 crc kubenswrapper[4556]: I0218 09:38:59.846438 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-9vt27" Feb 18 09:38:59 crc kubenswrapper[4556]: I0218 09:38:59.847436 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Feb 18 09:38:59 crc kubenswrapper[4556]: I0218 09:38:59.847458 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Feb 18 09:38:59 crc kubenswrapper[4556]: I0218 09:38:59.854020 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Feb 18 09:38:59 crc kubenswrapper[4556]: I0218 09:38:59.873650 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6j8k\" (UniqueName: \"kubernetes.io/projected/7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59-kube-api-access-x6j8k\") pod \"tempest-tests-tempest\" (UID: \"7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59\") " pod="openstack/tempest-tests-tempest" Feb 18 09:38:59 crc kubenswrapper[4556]: I0218 09:38:59.873694 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59\") " pod="openstack/tempest-tests-tempest" Feb 18 09:38:59 crc kubenswrapper[4556]: I0218 09:38:59.873726 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59\") " pod="openstack/tempest-tests-tempest" Feb 18 09:38:59 crc kubenswrapper[4556]: I0218 09:38:59.873784 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59\") " pod="openstack/tempest-tests-tempest" Feb 18 09:38:59 crc kubenswrapper[4556]: I0218 09:38:59.873944 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59-config-data\") pod \"tempest-tests-tempest\" (UID: \"7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59\") " pod="openstack/tempest-tests-tempest" Feb 18 09:38:59 crc kubenswrapper[4556]: I0218 09:38:59.874117 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59\") " pod="openstack/tempest-tests-tempest" Feb 18 09:38:59 crc kubenswrapper[4556]: I0218 09:38:59.874437 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59\") " pod="openstack/tempest-tests-tempest" Feb 18 09:38:59 crc kubenswrapper[4556]: I0218 09:38:59.874471 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59\") " pod="openstack/tempest-tests-tempest" Feb 18 09:38:59 crc kubenswrapper[4556]: I0218 09:38:59.874555 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"tempest-tests-tempest\" (UID: \"7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59\") " pod="openstack/tempest-tests-tempest" Feb 18 09:38:59 crc kubenswrapper[4556]: I0218 09:38:59.976685 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59\") " pod="openstack/tempest-tests-tempest" Feb 18 09:38:59 crc kubenswrapper[4556]: I0218 09:38:59.976804 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59-config-data\") pod \"tempest-tests-tempest\" (UID: \"7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59\") " pod="openstack/tempest-tests-tempest" Feb 18 09:38:59 crc kubenswrapper[4556]: I0218 09:38:59.976865 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59\") " pod="openstack/tempest-tests-tempest" Feb 18 09:38:59 crc kubenswrapper[4556]: I0218 09:38:59.977009 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59\") " pod="openstack/tempest-tests-tempest" Feb 18 09:38:59 crc kubenswrapper[4556]: I0218 09:38:59.977034 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59\") " pod="openstack/tempest-tests-tempest" Feb 18 09:38:59 crc kubenswrapper[4556]: I0218 09:38:59.977080 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"tempest-tests-tempest\" (UID: \"7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59\") " pod="openstack/tempest-tests-tempest" Feb 18 09:38:59 crc kubenswrapper[4556]: I0218 09:38:59.977123 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59\") " pod="openstack/tempest-tests-tempest" Feb 18 09:38:59 crc kubenswrapper[4556]: I0218 09:38:59.977128 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6j8k\" (UniqueName: \"kubernetes.io/projected/7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59-kube-api-access-x6j8k\") pod \"tempest-tests-tempest\" (UID: \"7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59\") " pod="openstack/tempest-tests-tempest" Feb 18 09:38:59 crc kubenswrapper[4556]: I0218 09:38:59.977220 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59\") " pod="openstack/tempest-tests-tempest" Feb 18 09:38:59 crc kubenswrapper[4556]: I0218 09:38:59.977264 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59\") " pod="openstack/tempest-tests-tempest" Feb 18 09:38:59 crc kubenswrapper[4556]: I0218 09:38:59.977509 4556 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"tempest-tests-tempest\" (UID: \"7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/tempest-tests-tempest" Feb 18 09:38:59 crc kubenswrapper[4556]: I0218 09:38:59.978216 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59\") " pod="openstack/tempest-tests-tempest" Feb 18 09:38:59 crc kubenswrapper[4556]: I0218 09:38:59.978784 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59\") " pod="openstack/tempest-tests-tempest" Feb 18 09:38:59 crc kubenswrapper[4556]: I0218 09:38:59.979421 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59-config-data\") pod \"tempest-tests-tempest\" (UID: \"7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59\") " pod="openstack/tempest-tests-tempest" Feb 18 09:38:59 crc kubenswrapper[4556]: I0218 09:38:59.984221 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59\") " pod="openstack/tempest-tests-tempest" Feb 18 09:38:59 crc kubenswrapper[4556]: I0218 09:38:59.984309 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59\") " pod="openstack/tempest-tests-tempest" Feb 18 09:38:59 crc kubenswrapper[4556]: I0218 09:38:59.984915 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59\") " pod="openstack/tempest-tests-tempest" Feb 18 09:38:59 crc kubenswrapper[4556]: I0218 09:38:59.994751 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6j8k\" (UniqueName: \"kubernetes.io/projected/7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59-kube-api-access-x6j8k\") pod \"tempest-tests-tempest\" (UID: \"7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59\") " pod="openstack/tempest-tests-tempest" Feb 18 09:39:00 crc kubenswrapper[4556]: I0218 09:39:00.005526 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"tempest-tests-tempest\" (UID: \"7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59\") " pod="openstack/tempest-tests-tempest" Feb 18 09:39:00 crc kubenswrapper[4556]: I0218 09:39:00.166702 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Feb 18 09:39:00 crc kubenswrapper[4556]: I0218 09:39:00.590144 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Feb 18 09:39:00 crc kubenswrapper[4556]: I0218 09:39:00.597734 4556 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 18 09:39:01 crc kubenswrapper[4556]: I0218 09:39:01.028852 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59","Type":"ContainerStarted","Data":"aeca925c01875991035f5d939d3e2f0d20910b37f24006ee7a735490518a300b"} Feb 18 09:39:31 crc kubenswrapper[4556]: I0218 09:39:31.733138 4556 patch_prober.go:28] interesting pod/machine-config-daemon-f76hs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 09:39:31 crc kubenswrapper[4556]: I0218 09:39:31.733990 4556 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 09:39:32 crc kubenswrapper[4556]: E0218 09:39:32.320037 4556 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Feb 18 09:39:32 crc kubenswrapper[4556]: E0218 09:39:32.320538 4556 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-x6j8k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 18 09:39:32 crc kubenswrapper[4556]: E0218 09:39:32.322206 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59" Feb 18 09:39:32 crc kubenswrapper[4556]: E0218 09:39:32.368850 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59" Feb 18 09:39:46 crc kubenswrapper[4556]: I0218 09:39:46.762870 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Feb 18 09:39:48 crc kubenswrapper[4556]: I0218 09:39:48.524432 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59","Type":"ContainerStarted","Data":"3cd1bb4b8d0bc13d13b1197e1a12491e89329de392c8c6c6c3c344d4dbf9b3e6"} Feb 18 09:39:48 crc kubenswrapper[4556]: I0218 09:39:48.545696 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=4.383309238 podStartE2EDuration="50.54548552s" podCreationTimestamp="2026-02-18 09:38:58 +0000 UTC" firstStartedPulling="2026-02-18 09:39:00.597513393 +0000 UTC m=+2097.614474373" lastFinishedPulling="2026-02-18 09:39:46.759689675 +0000 UTC m=+2143.776650655" observedRunningTime="2026-02-18 09:39:48.541720097 +0000 UTC m=+2145.558681076" watchObservedRunningTime="2026-02-18 09:39:48.54548552 +0000 UTC m=+2145.562446499" Feb 18 09:40:01 crc kubenswrapper[4556]: I0218 09:40:01.727473 4556 patch_prober.go:28] interesting pod/machine-config-daemon-f76hs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 09:40:01 crc kubenswrapper[4556]: I0218 09:40:01.728196 4556 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 09:40:31 crc kubenswrapper[4556]: I0218 09:40:31.727210 4556 patch_prober.go:28] interesting pod/machine-config-daemon-f76hs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 09:40:31 crc kubenswrapper[4556]: I0218 09:40:31.728025 4556 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 09:40:31 crc kubenswrapper[4556]: I0218 09:40:31.728082 4556 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" Feb 18 09:40:31 crc kubenswrapper[4556]: I0218 09:40:31.728977 4556 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"04a4196781cc895fa1780ee09f471180228eaaff00e4cc56bf534ca86fb1d640"} pod="openshift-machine-config-operator/machine-config-daemon-f76hs" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 18 09:40:31 crc kubenswrapper[4556]: I0218 09:40:31.729043 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerName="machine-config-daemon" containerID="cri-o://04a4196781cc895fa1780ee09f471180228eaaff00e4cc56bf534ca86fb1d640" gracePeriod=600 Feb 18 09:40:31 crc kubenswrapper[4556]: E0218 09:40:31.850178 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:40:31 crc kubenswrapper[4556]: I0218 09:40:31.941877 4556 generic.go:334] "Generic (PLEG): container finished" podID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerID="04a4196781cc895fa1780ee09f471180228eaaff00e4cc56bf534ca86fb1d640" exitCode=0 Feb 18 09:40:31 crc kubenswrapper[4556]: I0218 09:40:31.941951 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" event={"ID":"8dac7f27-d3d1-4778-9e54-f273035a1d37","Type":"ContainerDied","Data":"04a4196781cc895fa1780ee09f471180228eaaff00e4cc56bf534ca86fb1d640"} Feb 18 09:40:31 crc kubenswrapper[4556]: I0218 09:40:31.942241 4556 scope.go:117] "RemoveContainer" containerID="fd40a6d72be7a9ed07291f69c89aeb615c5bf9535dc9706ff886f8f14c11f976" Feb 18 09:40:31 crc kubenswrapper[4556]: I0218 09:40:31.942888 4556 scope.go:117] "RemoveContainer" containerID="04a4196781cc895fa1780ee09f471180228eaaff00e4cc56bf534ca86fb1d640" Feb 18 09:40:31 crc kubenswrapper[4556]: E0218 09:40:31.943235 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:40:42 crc kubenswrapper[4556]: I0218 09:40:42.283055 4556 scope.go:117] "RemoveContainer" containerID="04a4196781cc895fa1780ee09f471180228eaaff00e4cc56bf534ca86fb1d640" Feb 18 09:40:42 crc kubenswrapper[4556]: E0218 09:40:42.284171 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:40:48 crc kubenswrapper[4556]: I0218 09:40:48.068108 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-sgff2"] Feb 18 09:40:48 crc kubenswrapper[4556]: I0218 09:40:48.070406 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sgff2" Feb 18 09:40:48 crc kubenswrapper[4556]: I0218 09:40:48.084199 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-sgff2"] Feb 18 09:40:48 crc kubenswrapper[4556]: I0218 09:40:48.121834 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3029129-132b-48c3-829d-81fad7c3c131-catalog-content\") pod \"redhat-marketplace-sgff2\" (UID: \"f3029129-132b-48c3-829d-81fad7c3c131\") " pod="openshift-marketplace/redhat-marketplace-sgff2" Feb 18 09:40:48 crc kubenswrapper[4556]: I0218 09:40:48.121977 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3029129-132b-48c3-829d-81fad7c3c131-utilities\") pod \"redhat-marketplace-sgff2\" (UID: \"f3029129-132b-48c3-829d-81fad7c3c131\") " pod="openshift-marketplace/redhat-marketplace-sgff2" Feb 18 09:40:48 crc kubenswrapper[4556]: I0218 09:40:48.122098 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gm9xd\" (UniqueName: \"kubernetes.io/projected/f3029129-132b-48c3-829d-81fad7c3c131-kube-api-access-gm9xd\") pod \"redhat-marketplace-sgff2\" (UID: \"f3029129-132b-48c3-829d-81fad7c3c131\") " pod="openshift-marketplace/redhat-marketplace-sgff2" Feb 18 09:40:48 crc kubenswrapper[4556]: I0218 09:40:48.224313 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3029129-132b-48c3-829d-81fad7c3c131-catalog-content\") pod \"redhat-marketplace-sgff2\" (UID: \"f3029129-132b-48c3-829d-81fad7c3c131\") " pod="openshift-marketplace/redhat-marketplace-sgff2" Feb 18 09:40:48 crc kubenswrapper[4556]: I0218 09:40:48.224410 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3029129-132b-48c3-829d-81fad7c3c131-utilities\") pod \"redhat-marketplace-sgff2\" (UID: \"f3029129-132b-48c3-829d-81fad7c3c131\") " pod="openshift-marketplace/redhat-marketplace-sgff2" Feb 18 09:40:48 crc kubenswrapper[4556]: I0218 09:40:48.224499 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gm9xd\" (UniqueName: \"kubernetes.io/projected/f3029129-132b-48c3-829d-81fad7c3c131-kube-api-access-gm9xd\") pod \"redhat-marketplace-sgff2\" (UID: \"f3029129-132b-48c3-829d-81fad7c3c131\") " pod="openshift-marketplace/redhat-marketplace-sgff2" Feb 18 09:40:48 crc kubenswrapper[4556]: I0218 09:40:48.224889 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3029129-132b-48c3-829d-81fad7c3c131-catalog-content\") pod \"redhat-marketplace-sgff2\" (UID: \"f3029129-132b-48c3-829d-81fad7c3c131\") " pod="openshift-marketplace/redhat-marketplace-sgff2" Feb 18 09:40:48 crc kubenswrapper[4556]: I0218 09:40:48.224899 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3029129-132b-48c3-829d-81fad7c3c131-utilities\") pod \"redhat-marketplace-sgff2\" (UID: \"f3029129-132b-48c3-829d-81fad7c3c131\") " pod="openshift-marketplace/redhat-marketplace-sgff2" Feb 18 09:40:48 crc kubenswrapper[4556]: I0218 09:40:48.242785 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gm9xd\" (UniqueName: \"kubernetes.io/projected/f3029129-132b-48c3-829d-81fad7c3c131-kube-api-access-gm9xd\") pod \"redhat-marketplace-sgff2\" (UID: \"f3029129-132b-48c3-829d-81fad7c3c131\") " pod="openshift-marketplace/redhat-marketplace-sgff2" Feb 18 09:40:48 crc kubenswrapper[4556]: I0218 09:40:48.389378 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sgff2" Feb 18 09:40:48 crc kubenswrapper[4556]: I0218 09:40:48.835545 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-sgff2"] Feb 18 09:40:49 crc kubenswrapper[4556]: I0218 09:40:49.093793 4556 generic.go:334] "Generic (PLEG): container finished" podID="f3029129-132b-48c3-829d-81fad7c3c131" containerID="4e1471985ac36c8fec1616da2bf12177992237c424e465d138e388749fd01150" exitCode=0 Feb 18 09:40:49 crc kubenswrapper[4556]: I0218 09:40:49.093908 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sgff2" event={"ID":"f3029129-132b-48c3-829d-81fad7c3c131","Type":"ContainerDied","Data":"4e1471985ac36c8fec1616da2bf12177992237c424e465d138e388749fd01150"} Feb 18 09:40:49 crc kubenswrapper[4556]: I0218 09:40:49.094130 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sgff2" event={"ID":"f3029129-132b-48c3-829d-81fad7c3c131","Type":"ContainerStarted","Data":"d852c768bd2c03816ac5f13292d750a55a284a2b3cd8dd19383f9be26ef1ec9a"} Feb 18 09:40:50 crc kubenswrapper[4556]: I0218 09:40:50.103668 4556 generic.go:334] "Generic (PLEG): container finished" podID="f3029129-132b-48c3-829d-81fad7c3c131" containerID="3c9a437f81a0ab2c6916a4e85fa7a9221bc690597fbb5c83f45bd3bc9b2bafd1" exitCode=0 Feb 18 09:40:50 crc kubenswrapper[4556]: I0218 09:40:50.103825 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sgff2" event={"ID":"f3029129-132b-48c3-829d-81fad7c3c131","Type":"ContainerDied","Data":"3c9a437f81a0ab2c6916a4e85fa7a9221bc690597fbb5c83f45bd3bc9b2bafd1"} Feb 18 09:40:51 crc kubenswrapper[4556]: I0218 09:40:51.117371 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sgff2" event={"ID":"f3029129-132b-48c3-829d-81fad7c3c131","Type":"ContainerStarted","Data":"7a617e82d4743a8f22f16aa199e8ad05ccd638bd4f021568cfeeb3021d1bf681"} Feb 18 09:40:51 crc kubenswrapper[4556]: I0218 09:40:51.134409 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-sgff2" podStartSLOduration=1.6007023230000001 podStartE2EDuration="3.134377241s" podCreationTimestamp="2026-02-18 09:40:48 +0000 UTC" firstStartedPulling="2026-02-18 09:40:49.09572311 +0000 UTC m=+2206.112684090" lastFinishedPulling="2026-02-18 09:40:50.629398028 +0000 UTC m=+2207.646359008" observedRunningTime="2026-02-18 09:40:51.133859315 +0000 UTC m=+2208.150820295" watchObservedRunningTime="2026-02-18 09:40:51.134377241 +0000 UTC m=+2208.151338221" Feb 18 09:40:56 crc kubenswrapper[4556]: I0218 09:40:56.733588 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-pt4t9"] Feb 18 09:40:56 crc kubenswrapper[4556]: I0218 09:40:56.738668 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pt4t9" Feb 18 09:40:56 crc kubenswrapper[4556]: I0218 09:40:56.773668 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pt4t9"] Feb 18 09:40:56 crc kubenswrapper[4556]: I0218 09:40:56.856898 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6trj\" (UniqueName: \"kubernetes.io/projected/76fe1400-c43c-4bfd-bab3-377464edcbe3-kube-api-access-z6trj\") pod \"community-operators-pt4t9\" (UID: \"76fe1400-c43c-4bfd-bab3-377464edcbe3\") " pod="openshift-marketplace/community-operators-pt4t9" Feb 18 09:40:56 crc kubenswrapper[4556]: I0218 09:40:56.857224 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76fe1400-c43c-4bfd-bab3-377464edcbe3-catalog-content\") pod \"community-operators-pt4t9\" (UID: \"76fe1400-c43c-4bfd-bab3-377464edcbe3\") " pod="openshift-marketplace/community-operators-pt4t9" Feb 18 09:40:56 crc kubenswrapper[4556]: I0218 09:40:56.857293 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76fe1400-c43c-4bfd-bab3-377464edcbe3-utilities\") pod \"community-operators-pt4t9\" (UID: \"76fe1400-c43c-4bfd-bab3-377464edcbe3\") " pod="openshift-marketplace/community-operators-pt4t9" Feb 18 09:40:56 crc kubenswrapper[4556]: I0218 09:40:56.959182 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76fe1400-c43c-4bfd-bab3-377464edcbe3-catalog-content\") pod \"community-operators-pt4t9\" (UID: \"76fe1400-c43c-4bfd-bab3-377464edcbe3\") " pod="openshift-marketplace/community-operators-pt4t9" Feb 18 09:40:56 crc kubenswrapper[4556]: I0218 09:40:56.959245 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76fe1400-c43c-4bfd-bab3-377464edcbe3-utilities\") pod \"community-operators-pt4t9\" (UID: \"76fe1400-c43c-4bfd-bab3-377464edcbe3\") " pod="openshift-marketplace/community-operators-pt4t9" Feb 18 09:40:56 crc kubenswrapper[4556]: I0218 09:40:56.959365 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6trj\" (UniqueName: \"kubernetes.io/projected/76fe1400-c43c-4bfd-bab3-377464edcbe3-kube-api-access-z6trj\") pod \"community-operators-pt4t9\" (UID: \"76fe1400-c43c-4bfd-bab3-377464edcbe3\") " pod="openshift-marketplace/community-operators-pt4t9" Feb 18 09:40:56 crc kubenswrapper[4556]: I0218 09:40:56.959704 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76fe1400-c43c-4bfd-bab3-377464edcbe3-catalog-content\") pod \"community-operators-pt4t9\" (UID: \"76fe1400-c43c-4bfd-bab3-377464edcbe3\") " pod="openshift-marketplace/community-operators-pt4t9" Feb 18 09:40:56 crc kubenswrapper[4556]: I0218 09:40:56.959776 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76fe1400-c43c-4bfd-bab3-377464edcbe3-utilities\") pod \"community-operators-pt4t9\" (UID: \"76fe1400-c43c-4bfd-bab3-377464edcbe3\") " pod="openshift-marketplace/community-operators-pt4t9" Feb 18 09:40:56 crc kubenswrapper[4556]: I0218 09:40:56.979234 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6trj\" (UniqueName: \"kubernetes.io/projected/76fe1400-c43c-4bfd-bab3-377464edcbe3-kube-api-access-z6trj\") pod \"community-operators-pt4t9\" (UID: \"76fe1400-c43c-4bfd-bab3-377464edcbe3\") " pod="openshift-marketplace/community-operators-pt4t9" Feb 18 09:40:57 crc kubenswrapper[4556]: I0218 09:40:57.062609 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pt4t9" Feb 18 09:40:57 crc kubenswrapper[4556]: I0218 09:40:57.288232 4556 scope.go:117] "RemoveContainer" containerID="04a4196781cc895fa1780ee09f471180228eaaff00e4cc56bf534ca86fb1d640" Feb 18 09:40:57 crc kubenswrapper[4556]: E0218 09:40:57.289892 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:40:57 crc kubenswrapper[4556]: I0218 09:40:57.557297 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pt4t9"] Feb 18 09:40:57 crc kubenswrapper[4556]: W0218 09:40:57.557390 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod76fe1400_c43c_4bfd_bab3_377464edcbe3.slice/crio-5d30d6c2229ba631046c348989df567270053599a5180fd7fc952bee41a25cb5 WatchSource:0}: Error finding container 5d30d6c2229ba631046c348989df567270053599a5180fd7fc952bee41a25cb5: Status 404 returned error can't find the container with id 5d30d6c2229ba631046c348989df567270053599a5180fd7fc952bee41a25cb5 Feb 18 09:40:58 crc kubenswrapper[4556]: I0218 09:40:58.204224 4556 generic.go:334] "Generic (PLEG): container finished" podID="76fe1400-c43c-4bfd-bab3-377464edcbe3" containerID="da8477c2acd7577f20c3b7c0d0ce34f3d91ef6d16987027f386111246285a92e" exitCode=0 Feb 18 09:40:58 crc kubenswrapper[4556]: I0218 09:40:58.205244 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pt4t9" event={"ID":"76fe1400-c43c-4bfd-bab3-377464edcbe3","Type":"ContainerDied","Data":"da8477c2acd7577f20c3b7c0d0ce34f3d91ef6d16987027f386111246285a92e"} Feb 18 09:40:58 crc kubenswrapper[4556]: I0218 09:40:58.205301 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pt4t9" event={"ID":"76fe1400-c43c-4bfd-bab3-377464edcbe3","Type":"ContainerStarted","Data":"5d30d6c2229ba631046c348989df567270053599a5180fd7fc952bee41a25cb5"} Feb 18 09:40:58 crc kubenswrapper[4556]: I0218 09:40:58.390114 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-sgff2" Feb 18 09:40:58 crc kubenswrapper[4556]: I0218 09:40:58.390592 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-sgff2" Feb 18 09:40:58 crc kubenswrapper[4556]: I0218 09:40:58.434855 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-sgff2" Feb 18 09:40:59 crc kubenswrapper[4556]: I0218 09:40:59.215214 4556 generic.go:334] "Generic (PLEG): container finished" podID="76fe1400-c43c-4bfd-bab3-377464edcbe3" containerID="7c8c2daa1ccfcc43b893c248d90ee42f17c829e9c6b127ff268a5a0b0354e13d" exitCode=0 Feb 18 09:40:59 crc kubenswrapper[4556]: I0218 09:40:59.215318 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pt4t9" event={"ID":"76fe1400-c43c-4bfd-bab3-377464edcbe3","Type":"ContainerDied","Data":"7c8c2daa1ccfcc43b893c248d90ee42f17c829e9c6b127ff268a5a0b0354e13d"} Feb 18 09:40:59 crc kubenswrapper[4556]: I0218 09:40:59.253143 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-sgff2" Feb 18 09:41:00 crc kubenswrapper[4556]: I0218 09:41:00.228259 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pt4t9" event={"ID":"76fe1400-c43c-4bfd-bab3-377464edcbe3","Type":"ContainerStarted","Data":"794780a9b22c521895a5b1d122c6d9a2e130bc3aa0ea6228d84a7ac87f1401c5"} Feb 18 09:41:00 crc kubenswrapper[4556]: I0218 09:41:00.248228 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-pt4t9" podStartSLOduration=2.7057196919999997 podStartE2EDuration="4.248008861s" podCreationTimestamp="2026-02-18 09:40:56 +0000 UTC" firstStartedPulling="2026-02-18 09:40:58.207387357 +0000 UTC m=+2215.224348338" lastFinishedPulling="2026-02-18 09:40:59.749676526 +0000 UTC m=+2216.766637507" observedRunningTime="2026-02-18 09:41:00.243996071 +0000 UTC m=+2217.260957051" watchObservedRunningTime="2026-02-18 09:41:00.248008861 +0000 UTC m=+2217.264969841" Feb 18 09:41:00 crc kubenswrapper[4556]: I0218 09:41:00.708033 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-sgff2"] Feb 18 09:41:01 crc kubenswrapper[4556]: I0218 09:41:01.238853 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-sgff2" podUID="f3029129-132b-48c3-829d-81fad7c3c131" containerName="registry-server" containerID="cri-o://7a617e82d4743a8f22f16aa199e8ad05ccd638bd4f021568cfeeb3021d1bf681" gracePeriod=2 Feb 18 09:41:01 crc kubenswrapper[4556]: I0218 09:41:01.691697 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sgff2" Feb 18 09:41:01 crc kubenswrapper[4556]: I0218 09:41:01.791280 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3029129-132b-48c3-829d-81fad7c3c131-utilities\") pod \"f3029129-132b-48c3-829d-81fad7c3c131\" (UID: \"f3029129-132b-48c3-829d-81fad7c3c131\") " Feb 18 09:41:01 crc kubenswrapper[4556]: I0218 09:41:01.792163 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3029129-132b-48c3-829d-81fad7c3c131-utilities" (OuterVolumeSpecName: "utilities") pod "f3029129-132b-48c3-829d-81fad7c3c131" (UID: "f3029129-132b-48c3-829d-81fad7c3c131"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:41:01 crc kubenswrapper[4556]: I0218 09:41:01.792353 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3029129-132b-48c3-829d-81fad7c3c131-catalog-content\") pod \"f3029129-132b-48c3-829d-81fad7c3c131\" (UID: \"f3029129-132b-48c3-829d-81fad7c3c131\") " Feb 18 09:41:01 crc kubenswrapper[4556]: I0218 09:41:01.792397 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gm9xd\" (UniqueName: \"kubernetes.io/projected/f3029129-132b-48c3-829d-81fad7c3c131-kube-api-access-gm9xd\") pod \"f3029129-132b-48c3-829d-81fad7c3c131\" (UID: \"f3029129-132b-48c3-829d-81fad7c3c131\") " Feb 18 09:41:01 crc kubenswrapper[4556]: I0218 09:41:01.794197 4556 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3029129-132b-48c3-829d-81fad7c3c131-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 09:41:01 crc kubenswrapper[4556]: I0218 09:41:01.801377 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3029129-132b-48c3-829d-81fad7c3c131-kube-api-access-gm9xd" (OuterVolumeSpecName: "kube-api-access-gm9xd") pod "f3029129-132b-48c3-829d-81fad7c3c131" (UID: "f3029129-132b-48c3-829d-81fad7c3c131"). InnerVolumeSpecName "kube-api-access-gm9xd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:41:01 crc kubenswrapper[4556]: I0218 09:41:01.808611 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3029129-132b-48c3-829d-81fad7c3c131-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f3029129-132b-48c3-829d-81fad7c3c131" (UID: "f3029129-132b-48c3-829d-81fad7c3c131"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:41:01 crc kubenswrapper[4556]: I0218 09:41:01.896281 4556 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3029129-132b-48c3-829d-81fad7c3c131-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 09:41:01 crc kubenswrapper[4556]: I0218 09:41:01.896309 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gm9xd\" (UniqueName: \"kubernetes.io/projected/f3029129-132b-48c3-829d-81fad7c3c131-kube-api-access-gm9xd\") on node \"crc\" DevicePath \"\"" Feb 18 09:41:02 crc kubenswrapper[4556]: I0218 09:41:02.249292 4556 generic.go:334] "Generic (PLEG): container finished" podID="f3029129-132b-48c3-829d-81fad7c3c131" containerID="7a617e82d4743a8f22f16aa199e8ad05ccd638bd4f021568cfeeb3021d1bf681" exitCode=0 Feb 18 09:41:02 crc kubenswrapper[4556]: I0218 09:41:02.249358 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sgff2" event={"ID":"f3029129-132b-48c3-829d-81fad7c3c131","Type":"ContainerDied","Data":"7a617e82d4743a8f22f16aa199e8ad05ccd638bd4f021568cfeeb3021d1bf681"} Feb 18 09:41:02 crc kubenswrapper[4556]: I0218 09:41:02.249398 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sgff2" Feb 18 09:41:02 crc kubenswrapper[4556]: I0218 09:41:02.249439 4556 scope.go:117] "RemoveContainer" containerID="7a617e82d4743a8f22f16aa199e8ad05ccd638bd4f021568cfeeb3021d1bf681" Feb 18 09:41:02 crc kubenswrapper[4556]: I0218 09:41:02.249411 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sgff2" event={"ID":"f3029129-132b-48c3-829d-81fad7c3c131","Type":"ContainerDied","Data":"d852c768bd2c03816ac5f13292d750a55a284a2b3cd8dd19383f9be26ef1ec9a"} Feb 18 09:41:02 crc kubenswrapper[4556]: I0218 09:41:02.281521 4556 scope.go:117] "RemoveContainer" containerID="3c9a437f81a0ab2c6916a4e85fa7a9221bc690597fbb5c83f45bd3bc9b2bafd1" Feb 18 09:41:02 crc kubenswrapper[4556]: I0218 09:41:02.292667 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-sgff2"] Feb 18 09:41:02 crc kubenswrapper[4556]: I0218 09:41:02.297551 4556 scope.go:117] "RemoveContainer" containerID="4e1471985ac36c8fec1616da2bf12177992237c424e465d138e388749fd01150" Feb 18 09:41:02 crc kubenswrapper[4556]: I0218 09:41:02.302181 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-sgff2"] Feb 18 09:41:02 crc kubenswrapper[4556]: I0218 09:41:02.331871 4556 scope.go:117] "RemoveContainer" containerID="7a617e82d4743a8f22f16aa199e8ad05ccd638bd4f021568cfeeb3021d1bf681" Feb 18 09:41:02 crc kubenswrapper[4556]: E0218 09:41:02.332386 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a617e82d4743a8f22f16aa199e8ad05ccd638bd4f021568cfeeb3021d1bf681\": container with ID starting with 7a617e82d4743a8f22f16aa199e8ad05ccd638bd4f021568cfeeb3021d1bf681 not found: ID does not exist" containerID="7a617e82d4743a8f22f16aa199e8ad05ccd638bd4f021568cfeeb3021d1bf681" Feb 18 09:41:02 crc kubenswrapper[4556]: I0218 09:41:02.332419 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a617e82d4743a8f22f16aa199e8ad05ccd638bd4f021568cfeeb3021d1bf681"} err="failed to get container status \"7a617e82d4743a8f22f16aa199e8ad05ccd638bd4f021568cfeeb3021d1bf681\": rpc error: code = NotFound desc = could not find container \"7a617e82d4743a8f22f16aa199e8ad05ccd638bd4f021568cfeeb3021d1bf681\": container with ID starting with 7a617e82d4743a8f22f16aa199e8ad05ccd638bd4f021568cfeeb3021d1bf681 not found: ID does not exist" Feb 18 09:41:02 crc kubenswrapper[4556]: I0218 09:41:02.332450 4556 scope.go:117] "RemoveContainer" containerID="3c9a437f81a0ab2c6916a4e85fa7a9221bc690597fbb5c83f45bd3bc9b2bafd1" Feb 18 09:41:02 crc kubenswrapper[4556]: E0218 09:41:02.332676 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c9a437f81a0ab2c6916a4e85fa7a9221bc690597fbb5c83f45bd3bc9b2bafd1\": container with ID starting with 3c9a437f81a0ab2c6916a4e85fa7a9221bc690597fbb5c83f45bd3bc9b2bafd1 not found: ID does not exist" containerID="3c9a437f81a0ab2c6916a4e85fa7a9221bc690597fbb5c83f45bd3bc9b2bafd1" Feb 18 09:41:02 crc kubenswrapper[4556]: I0218 09:41:02.332721 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c9a437f81a0ab2c6916a4e85fa7a9221bc690597fbb5c83f45bd3bc9b2bafd1"} err="failed to get container status \"3c9a437f81a0ab2c6916a4e85fa7a9221bc690597fbb5c83f45bd3bc9b2bafd1\": rpc error: code = NotFound desc = could not find container \"3c9a437f81a0ab2c6916a4e85fa7a9221bc690597fbb5c83f45bd3bc9b2bafd1\": container with ID starting with 3c9a437f81a0ab2c6916a4e85fa7a9221bc690597fbb5c83f45bd3bc9b2bafd1 not found: ID does not exist" Feb 18 09:41:02 crc kubenswrapper[4556]: I0218 09:41:02.332739 4556 scope.go:117] "RemoveContainer" containerID="4e1471985ac36c8fec1616da2bf12177992237c424e465d138e388749fd01150" Feb 18 09:41:02 crc kubenswrapper[4556]: E0218 09:41:02.332984 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e1471985ac36c8fec1616da2bf12177992237c424e465d138e388749fd01150\": container with ID starting with 4e1471985ac36c8fec1616da2bf12177992237c424e465d138e388749fd01150 not found: ID does not exist" containerID="4e1471985ac36c8fec1616da2bf12177992237c424e465d138e388749fd01150" Feb 18 09:41:02 crc kubenswrapper[4556]: I0218 09:41:02.333021 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e1471985ac36c8fec1616da2bf12177992237c424e465d138e388749fd01150"} err="failed to get container status \"4e1471985ac36c8fec1616da2bf12177992237c424e465d138e388749fd01150\": rpc error: code = NotFound desc = could not find container \"4e1471985ac36c8fec1616da2bf12177992237c424e465d138e388749fd01150\": container with ID starting with 4e1471985ac36c8fec1616da2bf12177992237c424e465d138e388749fd01150 not found: ID does not exist" Feb 18 09:41:03 crc kubenswrapper[4556]: I0218 09:41:03.111263 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-c52fs"] Feb 18 09:41:03 crc kubenswrapper[4556]: E0218 09:41:03.111753 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3029129-132b-48c3-829d-81fad7c3c131" containerName="registry-server" Feb 18 09:41:03 crc kubenswrapper[4556]: I0218 09:41:03.111776 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3029129-132b-48c3-829d-81fad7c3c131" containerName="registry-server" Feb 18 09:41:03 crc kubenswrapper[4556]: E0218 09:41:03.111789 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3029129-132b-48c3-829d-81fad7c3c131" containerName="extract-utilities" Feb 18 09:41:03 crc kubenswrapper[4556]: I0218 09:41:03.111797 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3029129-132b-48c3-829d-81fad7c3c131" containerName="extract-utilities" Feb 18 09:41:03 crc kubenswrapper[4556]: E0218 09:41:03.111817 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3029129-132b-48c3-829d-81fad7c3c131" containerName="extract-content" Feb 18 09:41:03 crc kubenswrapper[4556]: I0218 09:41:03.111824 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3029129-132b-48c3-829d-81fad7c3c131" containerName="extract-content" Feb 18 09:41:03 crc kubenswrapper[4556]: I0218 09:41:03.112062 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3029129-132b-48c3-829d-81fad7c3c131" containerName="registry-server" Feb 18 09:41:03 crc kubenswrapper[4556]: I0218 09:41:03.113377 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c52fs" Feb 18 09:41:03 crc kubenswrapper[4556]: I0218 09:41:03.124896 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-c52fs"] Feb 18 09:41:03 crc kubenswrapper[4556]: I0218 09:41:03.224122 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmv5f\" (UniqueName: \"kubernetes.io/projected/a6ed3aad-6ec1-443d-8699-5703362a6131-kube-api-access-zmv5f\") pod \"redhat-operators-c52fs\" (UID: \"a6ed3aad-6ec1-443d-8699-5703362a6131\") " pod="openshift-marketplace/redhat-operators-c52fs" Feb 18 09:41:03 crc kubenswrapper[4556]: I0218 09:41:03.224351 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6ed3aad-6ec1-443d-8699-5703362a6131-utilities\") pod \"redhat-operators-c52fs\" (UID: \"a6ed3aad-6ec1-443d-8699-5703362a6131\") " pod="openshift-marketplace/redhat-operators-c52fs" Feb 18 09:41:03 crc kubenswrapper[4556]: I0218 09:41:03.224444 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6ed3aad-6ec1-443d-8699-5703362a6131-catalog-content\") pod \"redhat-operators-c52fs\" (UID: \"a6ed3aad-6ec1-443d-8699-5703362a6131\") " pod="openshift-marketplace/redhat-operators-c52fs" Feb 18 09:41:03 crc kubenswrapper[4556]: I0218 09:41:03.318521 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3029129-132b-48c3-829d-81fad7c3c131" path="/var/lib/kubelet/pods/f3029129-132b-48c3-829d-81fad7c3c131/volumes" Feb 18 09:41:03 crc kubenswrapper[4556]: I0218 09:41:03.326183 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmv5f\" (UniqueName: \"kubernetes.io/projected/a6ed3aad-6ec1-443d-8699-5703362a6131-kube-api-access-zmv5f\") pod \"redhat-operators-c52fs\" (UID: \"a6ed3aad-6ec1-443d-8699-5703362a6131\") " pod="openshift-marketplace/redhat-operators-c52fs" Feb 18 09:41:03 crc kubenswrapper[4556]: I0218 09:41:03.326306 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6ed3aad-6ec1-443d-8699-5703362a6131-utilities\") pod \"redhat-operators-c52fs\" (UID: \"a6ed3aad-6ec1-443d-8699-5703362a6131\") " pod="openshift-marketplace/redhat-operators-c52fs" Feb 18 09:41:03 crc kubenswrapper[4556]: I0218 09:41:03.326358 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6ed3aad-6ec1-443d-8699-5703362a6131-catalog-content\") pod \"redhat-operators-c52fs\" (UID: \"a6ed3aad-6ec1-443d-8699-5703362a6131\") " pod="openshift-marketplace/redhat-operators-c52fs" Feb 18 09:41:03 crc kubenswrapper[4556]: I0218 09:41:03.327393 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6ed3aad-6ec1-443d-8699-5703362a6131-utilities\") pod \"redhat-operators-c52fs\" (UID: \"a6ed3aad-6ec1-443d-8699-5703362a6131\") " pod="openshift-marketplace/redhat-operators-c52fs" Feb 18 09:41:03 crc kubenswrapper[4556]: I0218 09:41:03.327749 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6ed3aad-6ec1-443d-8699-5703362a6131-catalog-content\") pod \"redhat-operators-c52fs\" (UID: \"a6ed3aad-6ec1-443d-8699-5703362a6131\") " pod="openshift-marketplace/redhat-operators-c52fs" Feb 18 09:41:03 crc kubenswrapper[4556]: I0218 09:41:03.344703 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmv5f\" (UniqueName: \"kubernetes.io/projected/a6ed3aad-6ec1-443d-8699-5703362a6131-kube-api-access-zmv5f\") pod \"redhat-operators-c52fs\" (UID: \"a6ed3aad-6ec1-443d-8699-5703362a6131\") " pod="openshift-marketplace/redhat-operators-c52fs" Feb 18 09:41:03 crc kubenswrapper[4556]: I0218 09:41:03.429590 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c52fs" Feb 18 09:41:03 crc kubenswrapper[4556]: I0218 09:41:03.880685 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-c52fs"] Feb 18 09:41:04 crc kubenswrapper[4556]: I0218 09:41:04.277449 4556 generic.go:334] "Generic (PLEG): container finished" podID="a6ed3aad-6ec1-443d-8699-5703362a6131" containerID="8cfff50a13374adc5af106e75b4d6926615a5b7449f8ec151894aa3a08764c36" exitCode=0 Feb 18 09:41:04 crc kubenswrapper[4556]: I0218 09:41:04.277639 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c52fs" event={"ID":"a6ed3aad-6ec1-443d-8699-5703362a6131","Type":"ContainerDied","Data":"8cfff50a13374adc5af106e75b4d6926615a5b7449f8ec151894aa3a08764c36"} Feb 18 09:41:04 crc kubenswrapper[4556]: I0218 09:41:04.277824 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c52fs" event={"ID":"a6ed3aad-6ec1-443d-8699-5703362a6131","Type":"ContainerStarted","Data":"6f235f4cf644621390679a3fb7fbbb2b0c0b79c9ae10db754c1e12554ec897e6"} Feb 18 09:41:05 crc kubenswrapper[4556]: I0218 09:41:05.290859 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c52fs" event={"ID":"a6ed3aad-6ec1-443d-8699-5703362a6131","Type":"ContainerStarted","Data":"b20ba8502043359354a9d0d258e4b9f461c932e2b43490d456ce048051b9aebb"} Feb 18 09:41:07 crc kubenswrapper[4556]: I0218 09:41:07.064316 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-pt4t9" Feb 18 09:41:07 crc kubenswrapper[4556]: I0218 09:41:07.064772 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-pt4t9" Feb 18 09:41:07 crc kubenswrapper[4556]: I0218 09:41:07.108382 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-pt4t9" Feb 18 09:41:07 crc kubenswrapper[4556]: I0218 09:41:07.309269 4556 generic.go:334] "Generic (PLEG): container finished" podID="a6ed3aad-6ec1-443d-8699-5703362a6131" containerID="b20ba8502043359354a9d0d258e4b9f461c932e2b43490d456ce048051b9aebb" exitCode=0 Feb 18 09:41:07 crc kubenswrapper[4556]: I0218 09:41:07.309350 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c52fs" event={"ID":"a6ed3aad-6ec1-443d-8699-5703362a6131","Type":"ContainerDied","Data":"b20ba8502043359354a9d0d258e4b9f461c932e2b43490d456ce048051b9aebb"} Feb 18 09:41:07 crc kubenswrapper[4556]: I0218 09:41:07.353355 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-pt4t9" Feb 18 09:41:08 crc kubenswrapper[4556]: I0218 09:41:08.282910 4556 scope.go:117] "RemoveContainer" containerID="04a4196781cc895fa1780ee09f471180228eaaff00e4cc56bf534ca86fb1d640" Feb 18 09:41:08 crc kubenswrapper[4556]: E0218 09:41:08.283925 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:41:08 crc kubenswrapper[4556]: I0218 09:41:08.323578 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c52fs" event={"ID":"a6ed3aad-6ec1-443d-8699-5703362a6131","Type":"ContainerStarted","Data":"8b4cf8225c1f49804f7c6bae498246872fe8b28765c450e06103ae307241af43"} Feb 18 09:41:08 crc kubenswrapper[4556]: I0218 09:41:08.348731 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-c52fs" podStartSLOduration=1.8428492699999999 podStartE2EDuration="5.348715457s" podCreationTimestamp="2026-02-18 09:41:03 +0000 UTC" firstStartedPulling="2026-02-18 09:41:04.280002548 +0000 UTC m=+2221.296963529" lastFinishedPulling="2026-02-18 09:41:07.785868736 +0000 UTC m=+2224.802829716" observedRunningTime="2026-02-18 09:41:08.340169655 +0000 UTC m=+2225.357130626" watchObservedRunningTime="2026-02-18 09:41:08.348715457 +0000 UTC m=+2225.365676437" Feb 18 09:41:09 crc kubenswrapper[4556]: I0218 09:41:09.504029 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pt4t9"] Feb 18 09:41:09 crc kubenswrapper[4556]: I0218 09:41:09.504500 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-pt4t9" podUID="76fe1400-c43c-4bfd-bab3-377464edcbe3" containerName="registry-server" containerID="cri-o://794780a9b22c521895a5b1d122c6d9a2e130bc3aa0ea6228d84a7ac87f1401c5" gracePeriod=2 Feb 18 09:41:09 crc kubenswrapper[4556]: I0218 09:41:09.945937 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pt4t9" Feb 18 09:41:09 crc kubenswrapper[4556]: I0218 09:41:09.982860 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76fe1400-c43c-4bfd-bab3-377464edcbe3-catalog-content\") pod \"76fe1400-c43c-4bfd-bab3-377464edcbe3\" (UID: \"76fe1400-c43c-4bfd-bab3-377464edcbe3\") " Feb 18 09:41:09 crc kubenswrapper[4556]: I0218 09:41:09.983130 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76fe1400-c43c-4bfd-bab3-377464edcbe3-utilities\") pod \"76fe1400-c43c-4bfd-bab3-377464edcbe3\" (UID: \"76fe1400-c43c-4bfd-bab3-377464edcbe3\") " Feb 18 09:41:09 crc kubenswrapper[4556]: I0218 09:41:09.983425 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z6trj\" (UniqueName: \"kubernetes.io/projected/76fe1400-c43c-4bfd-bab3-377464edcbe3-kube-api-access-z6trj\") pod \"76fe1400-c43c-4bfd-bab3-377464edcbe3\" (UID: \"76fe1400-c43c-4bfd-bab3-377464edcbe3\") " Feb 18 09:41:09 crc kubenswrapper[4556]: I0218 09:41:09.983764 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76fe1400-c43c-4bfd-bab3-377464edcbe3-utilities" (OuterVolumeSpecName: "utilities") pod "76fe1400-c43c-4bfd-bab3-377464edcbe3" (UID: "76fe1400-c43c-4bfd-bab3-377464edcbe3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:41:09 crc kubenswrapper[4556]: I0218 09:41:09.985199 4556 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76fe1400-c43c-4bfd-bab3-377464edcbe3-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 09:41:09 crc kubenswrapper[4556]: I0218 09:41:09.993325 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76fe1400-c43c-4bfd-bab3-377464edcbe3-kube-api-access-z6trj" (OuterVolumeSpecName: "kube-api-access-z6trj") pod "76fe1400-c43c-4bfd-bab3-377464edcbe3" (UID: "76fe1400-c43c-4bfd-bab3-377464edcbe3"). InnerVolumeSpecName "kube-api-access-z6trj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:41:10 crc kubenswrapper[4556]: I0218 09:41:10.021248 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76fe1400-c43c-4bfd-bab3-377464edcbe3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "76fe1400-c43c-4bfd-bab3-377464edcbe3" (UID: "76fe1400-c43c-4bfd-bab3-377464edcbe3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:41:10 crc kubenswrapper[4556]: I0218 09:41:10.087185 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z6trj\" (UniqueName: \"kubernetes.io/projected/76fe1400-c43c-4bfd-bab3-377464edcbe3-kube-api-access-z6trj\") on node \"crc\" DevicePath \"\"" Feb 18 09:41:10 crc kubenswrapper[4556]: I0218 09:41:10.087224 4556 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76fe1400-c43c-4bfd-bab3-377464edcbe3-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 09:41:10 crc kubenswrapper[4556]: I0218 09:41:10.368231 4556 generic.go:334] "Generic (PLEG): container finished" podID="76fe1400-c43c-4bfd-bab3-377464edcbe3" containerID="794780a9b22c521895a5b1d122c6d9a2e130bc3aa0ea6228d84a7ac87f1401c5" exitCode=0 Feb 18 09:41:10 crc kubenswrapper[4556]: I0218 09:41:10.368320 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pt4t9" Feb 18 09:41:10 crc kubenswrapper[4556]: I0218 09:41:10.368341 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pt4t9" event={"ID":"76fe1400-c43c-4bfd-bab3-377464edcbe3","Type":"ContainerDied","Data":"794780a9b22c521895a5b1d122c6d9a2e130bc3aa0ea6228d84a7ac87f1401c5"} Feb 18 09:41:10 crc kubenswrapper[4556]: I0218 09:41:10.368552 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pt4t9" event={"ID":"76fe1400-c43c-4bfd-bab3-377464edcbe3","Type":"ContainerDied","Data":"5d30d6c2229ba631046c348989df567270053599a5180fd7fc952bee41a25cb5"} Feb 18 09:41:10 crc kubenswrapper[4556]: I0218 09:41:10.368574 4556 scope.go:117] "RemoveContainer" containerID="794780a9b22c521895a5b1d122c6d9a2e130bc3aa0ea6228d84a7ac87f1401c5" Feb 18 09:41:10 crc kubenswrapper[4556]: I0218 09:41:10.388339 4556 scope.go:117] "RemoveContainer" containerID="7c8c2daa1ccfcc43b893c248d90ee42f17c829e9c6b127ff268a5a0b0354e13d" Feb 18 09:41:10 crc kubenswrapper[4556]: I0218 09:41:10.402534 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pt4t9"] Feb 18 09:41:10 crc kubenswrapper[4556]: I0218 09:41:10.411217 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-pt4t9"] Feb 18 09:41:10 crc kubenswrapper[4556]: I0218 09:41:10.415385 4556 scope.go:117] "RemoveContainer" containerID="da8477c2acd7577f20c3b7c0d0ce34f3d91ef6d16987027f386111246285a92e" Feb 18 09:41:10 crc kubenswrapper[4556]: I0218 09:41:10.445849 4556 scope.go:117] "RemoveContainer" containerID="794780a9b22c521895a5b1d122c6d9a2e130bc3aa0ea6228d84a7ac87f1401c5" Feb 18 09:41:10 crc kubenswrapper[4556]: E0218 09:41:10.446379 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"794780a9b22c521895a5b1d122c6d9a2e130bc3aa0ea6228d84a7ac87f1401c5\": container with ID starting with 794780a9b22c521895a5b1d122c6d9a2e130bc3aa0ea6228d84a7ac87f1401c5 not found: ID does not exist" containerID="794780a9b22c521895a5b1d122c6d9a2e130bc3aa0ea6228d84a7ac87f1401c5" Feb 18 09:41:10 crc kubenswrapper[4556]: I0218 09:41:10.446423 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"794780a9b22c521895a5b1d122c6d9a2e130bc3aa0ea6228d84a7ac87f1401c5"} err="failed to get container status \"794780a9b22c521895a5b1d122c6d9a2e130bc3aa0ea6228d84a7ac87f1401c5\": rpc error: code = NotFound desc = could not find container \"794780a9b22c521895a5b1d122c6d9a2e130bc3aa0ea6228d84a7ac87f1401c5\": container with ID starting with 794780a9b22c521895a5b1d122c6d9a2e130bc3aa0ea6228d84a7ac87f1401c5 not found: ID does not exist" Feb 18 09:41:10 crc kubenswrapper[4556]: I0218 09:41:10.446464 4556 scope.go:117] "RemoveContainer" containerID="7c8c2daa1ccfcc43b893c248d90ee42f17c829e9c6b127ff268a5a0b0354e13d" Feb 18 09:41:10 crc kubenswrapper[4556]: E0218 09:41:10.446888 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c8c2daa1ccfcc43b893c248d90ee42f17c829e9c6b127ff268a5a0b0354e13d\": container with ID starting with 7c8c2daa1ccfcc43b893c248d90ee42f17c829e9c6b127ff268a5a0b0354e13d not found: ID does not exist" containerID="7c8c2daa1ccfcc43b893c248d90ee42f17c829e9c6b127ff268a5a0b0354e13d" Feb 18 09:41:10 crc kubenswrapper[4556]: I0218 09:41:10.447009 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c8c2daa1ccfcc43b893c248d90ee42f17c829e9c6b127ff268a5a0b0354e13d"} err="failed to get container status \"7c8c2daa1ccfcc43b893c248d90ee42f17c829e9c6b127ff268a5a0b0354e13d\": rpc error: code = NotFound desc = could not find container \"7c8c2daa1ccfcc43b893c248d90ee42f17c829e9c6b127ff268a5a0b0354e13d\": container with ID starting with 7c8c2daa1ccfcc43b893c248d90ee42f17c829e9c6b127ff268a5a0b0354e13d not found: ID does not exist" Feb 18 09:41:10 crc kubenswrapper[4556]: I0218 09:41:10.447108 4556 scope.go:117] "RemoveContainer" containerID="da8477c2acd7577f20c3b7c0d0ce34f3d91ef6d16987027f386111246285a92e" Feb 18 09:41:10 crc kubenswrapper[4556]: E0218 09:41:10.447467 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da8477c2acd7577f20c3b7c0d0ce34f3d91ef6d16987027f386111246285a92e\": container with ID starting with da8477c2acd7577f20c3b7c0d0ce34f3d91ef6d16987027f386111246285a92e not found: ID does not exist" containerID="da8477c2acd7577f20c3b7c0d0ce34f3d91ef6d16987027f386111246285a92e" Feb 18 09:41:10 crc kubenswrapper[4556]: I0218 09:41:10.447498 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da8477c2acd7577f20c3b7c0d0ce34f3d91ef6d16987027f386111246285a92e"} err="failed to get container status \"da8477c2acd7577f20c3b7c0d0ce34f3d91ef6d16987027f386111246285a92e\": rpc error: code = NotFound desc = could not find container \"da8477c2acd7577f20c3b7c0d0ce34f3d91ef6d16987027f386111246285a92e\": container with ID starting with da8477c2acd7577f20c3b7c0d0ce34f3d91ef6d16987027f386111246285a92e not found: ID does not exist" Feb 18 09:41:11 crc kubenswrapper[4556]: I0218 09:41:11.293507 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76fe1400-c43c-4bfd-bab3-377464edcbe3" path="/var/lib/kubelet/pods/76fe1400-c43c-4bfd-bab3-377464edcbe3/volumes" Feb 18 09:41:13 crc kubenswrapper[4556]: I0218 09:41:13.430577 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-c52fs" Feb 18 09:41:13 crc kubenswrapper[4556]: I0218 09:41:13.430878 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-c52fs" Feb 18 09:41:13 crc kubenswrapper[4556]: I0218 09:41:13.468750 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-c52fs" Feb 18 09:41:14 crc kubenswrapper[4556]: I0218 09:41:14.456706 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-c52fs" Feb 18 09:41:14 crc kubenswrapper[4556]: I0218 09:41:14.703355 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-c52fs"] Feb 18 09:41:16 crc kubenswrapper[4556]: I0218 09:41:16.429313 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-c52fs" podUID="a6ed3aad-6ec1-443d-8699-5703362a6131" containerName="registry-server" containerID="cri-o://8b4cf8225c1f49804f7c6bae498246872fe8b28765c450e06103ae307241af43" gracePeriod=2 Feb 18 09:41:16 crc kubenswrapper[4556]: I0218 09:41:16.874193 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c52fs" Feb 18 09:41:16 crc kubenswrapper[4556]: I0218 09:41:16.948429 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6ed3aad-6ec1-443d-8699-5703362a6131-utilities\") pod \"a6ed3aad-6ec1-443d-8699-5703362a6131\" (UID: \"a6ed3aad-6ec1-443d-8699-5703362a6131\") " Feb 18 09:41:16 crc kubenswrapper[4556]: I0218 09:41:16.948643 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6ed3aad-6ec1-443d-8699-5703362a6131-catalog-content\") pod \"a6ed3aad-6ec1-443d-8699-5703362a6131\" (UID: \"a6ed3aad-6ec1-443d-8699-5703362a6131\") " Feb 18 09:41:16 crc kubenswrapper[4556]: I0218 09:41:16.948775 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zmv5f\" (UniqueName: \"kubernetes.io/projected/a6ed3aad-6ec1-443d-8699-5703362a6131-kube-api-access-zmv5f\") pod \"a6ed3aad-6ec1-443d-8699-5703362a6131\" (UID: \"a6ed3aad-6ec1-443d-8699-5703362a6131\") " Feb 18 09:41:16 crc kubenswrapper[4556]: I0218 09:41:16.949397 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6ed3aad-6ec1-443d-8699-5703362a6131-utilities" (OuterVolumeSpecName: "utilities") pod "a6ed3aad-6ec1-443d-8699-5703362a6131" (UID: "a6ed3aad-6ec1-443d-8699-5703362a6131"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:41:16 crc kubenswrapper[4556]: I0218 09:41:16.949621 4556 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6ed3aad-6ec1-443d-8699-5703362a6131-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 09:41:16 crc kubenswrapper[4556]: I0218 09:41:16.955076 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6ed3aad-6ec1-443d-8699-5703362a6131-kube-api-access-zmv5f" (OuterVolumeSpecName: "kube-api-access-zmv5f") pod "a6ed3aad-6ec1-443d-8699-5703362a6131" (UID: "a6ed3aad-6ec1-443d-8699-5703362a6131"). InnerVolumeSpecName "kube-api-access-zmv5f". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:41:17 crc kubenswrapper[4556]: I0218 09:41:17.053382 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zmv5f\" (UniqueName: \"kubernetes.io/projected/a6ed3aad-6ec1-443d-8699-5703362a6131-kube-api-access-zmv5f\") on node \"crc\" DevicePath \"\"" Feb 18 09:41:17 crc kubenswrapper[4556]: I0218 09:41:17.056954 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6ed3aad-6ec1-443d-8699-5703362a6131-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a6ed3aad-6ec1-443d-8699-5703362a6131" (UID: "a6ed3aad-6ec1-443d-8699-5703362a6131"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:41:17 crc kubenswrapper[4556]: I0218 09:41:17.155065 4556 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6ed3aad-6ec1-443d-8699-5703362a6131-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 09:41:17 crc kubenswrapper[4556]: I0218 09:41:17.442241 4556 generic.go:334] "Generic (PLEG): container finished" podID="a6ed3aad-6ec1-443d-8699-5703362a6131" containerID="8b4cf8225c1f49804f7c6bae498246872fe8b28765c450e06103ae307241af43" exitCode=0 Feb 18 09:41:17 crc kubenswrapper[4556]: I0218 09:41:17.442303 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c52fs" event={"ID":"a6ed3aad-6ec1-443d-8699-5703362a6131","Type":"ContainerDied","Data":"8b4cf8225c1f49804f7c6bae498246872fe8b28765c450e06103ae307241af43"} Feb 18 09:41:17 crc kubenswrapper[4556]: I0218 09:41:17.442694 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c52fs" event={"ID":"a6ed3aad-6ec1-443d-8699-5703362a6131","Type":"ContainerDied","Data":"6f235f4cf644621390679a3fb7fbbb2b0c0b79c9ae10db754c1e12554ec897e6"} Feb 18 09:41:17 crc kubenswrapper[4556]: I0218 09:41:17.442733 4556 scope.go:117] "RemoveContainer" containerID="8b4cf8225c1f49804f7c6bae498246872fe8b28765c450e06103ae307241af43" Feb 18 09:41:17 crc kubenswrapper[4556]: I0218 09:41:17.442328 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c52fs" Feb 18 09:41:17 crc kubenswrapper[4556]: I0218 09:41:17.470013 4556 scope.go:117] "RemoveContainer" containerID="b20ba8502043359354a9d0d258e4b9f461c932e2b43490d456ce048051b9aebb" Feb 18 09:41:17 crc kubenswrapper[4556]: I0218 09:41:17.494373 4556 scope.go:117] "RemoveContainer" containerID="8cfff50a13374adc5af106e75b4d6926615a5b7449f8ec151894aa3a08764c36" Feb 18 09:41:17 crc kubenswrapper[4556]: I0218 09:41:17.496037 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-c52fs"] Feb 18 09:41:17 crc kubenswrapper[4556]: I0218 09:41:17.504415 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-c52fs"] Feb 18 09:41:17 crc kubenswrapper[4556]: I0218 09:41:17.529669 4556 scope.go:117] "RemoveContainer" containerID="8b4cf8225c1f49804f7c6bae498246872fe8b28765c450e06103ae307241af43" Feb 18 09:41:17 crc kubenswrapper[4556]: E0218 09:41:17.530048 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b4cf8225c1f49804f7c6bae498246872fe8b28765c450e06103ae307241af43\": container with ID starting with 8b4cf8225c1f49804f7c6bae498246872fe8b28765c450e06103ae307241af43 not found: ID does not exist" containerID="8b4cf8225c1f49804f7c6bae498246872fe8b28765c450e06103ae307241af43" Feb 18 09:41:17 crc kubenswrapper[4556]: I0218 09:41:17.530091 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b4cf8225c1f49804f7c6bae498246872fe8b28765c450e06103ae307241af43"} err="failed to get container status \"8b4cf8225c1f49804f7c6bae498246872fe8b28765c450e06103ae307241af43\": rpc error: code = NotFound desc = could not find container \"8b4cf8225c1f49804f7c6bae498246872fe8b28765c450e06103ae307241af43\": container with ID starting with 8b4cf8225c1f49804f7c6bae498246872fe8b28765c450e06103ae307241af43 not found: ID does not exist" Feb 18 09:41:17 crc kubenswrapper[4556]: I0218 09:41:17.530121 4556 scope.go:117] "RemoveContainer" containerID="b20ba8502043359354a9d0d258e4b9f461c932e2b43490d456ce048051b9aebb" Feb 18 09:41:17 crc kubenswrapper[4556]: E0218 09:41:17.530481 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b20ba8502043359354a9d0d258e4b9f461c932e2b43490d456ce048051b9aebb\": container with ID starting with b20ba8502043359354a9d0d258e4b9f461c932e2b43490d456ce048051b9aebb not found: ID does not exist" containerID="b20ba8502043359354a9d0d258e4b9f461c932e2b43490d456ce048051b9aebb" Feb 18 09:41:17 crc kubenswrapper[4556]: I0218 09:41:17.530505 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b20ba8502043359354a9d0d258e4b9f461c932e2b43490d456ce048051b9aebb"} err="failed to get container status \"b20ba8502043359354a9d0d258e4b9f461c932e2b43490d456ce048051b9aebb\": rpc error: code = NotFound desc = could not find container \"b20ba8502043359354a9d0d258e4b9f461c932e2b43490d456ce048051b9aebb\": container with ID starting with b20ba8502043359354a9d0d258e4b9f461c932e2b43490d456ce048051b9aebb not found: ID does not exist" Feb 18 09:41:17 crc kubenswrapper[4556]: I0218 09:41:17.530519 4556 scope.go:117] "RemoveContainer" containerID="8cfff50a13374adc5af106e75b4d6926615a5b7449f8ec151894aa3a08764c36" Feb 18 09:41:17 crc kubenswrapper[4556]: E0218 09:41:17.530742 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8cfff50a13374adc5af106e75b4d6926615a5b7449f8ec151894aa3a08764c36\": container with ID starting with 8cfff50a13374adc5af106e75b4d6926615a5b7449f8ec151894aa3a08764c36 not found: ID does not exist" containerID="8cfff50a13374adc5af106e75b4d6926615a5b7449f8ec151894aa3a08764c36" Feb 18 09:41:17 crc kubenswrapper[4556]: I0218 09:41:17.530762 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8cfff50a13374adc5af106e75b4d6926615a5b7449f8ec151894aa3a08764c36"} err="failed to get container status \"8cfff50a13374adc5af106e75b4d6926615a5b7449f8ec151894aa3a08764c36\": rpc error: code = NotFound desc = could not find container \"8cfff50a13374adc5af106e75b4d6926615a5b7449f8ec151894aa3a08764c36\": container with ID starting with 8cfff50a13374adc5af106e75b4d6926615a5b7449f8ec151894aa3a08764c36 not found: ID does not exist" Feb 18 09:41:19 crc kubenswrapper[4556]: I0218 09:41:19.290517 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6ed3aad-6ec1-443d-8699-5703362a6131" path="/var/lib/kubelet/pods/a6ed3aad-6ec1-443d-8699-5703362a6131/volumes" Feb 18 09:41:20 crc kubenswrapper[4556]: I0218 09:41:20.283677 4556 scope.go:117] "RemoveContainer" containerID="04a4196781cc895fa1780ee09f471180228eaaff00e4cc56bf534ca86fb1d640" Feb 18 09:41:20 crc kubenswrapper[4556]: E0218 09:41:20.284317 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:41:34 crc kubenswrapper[4556]: I0218 09:41:34.282420 4556 scope.go:117] "RemoveContainer" containerID="04a4196781cc895fa1780ee09f471180228eaaff00e4cc56bf534ca86fb1d640" Feb 18 09:41:34 crc kubenswrapper[4556]: E0218 09:41:34.283395 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:41:48 crc kubenswrapper[4556]: I0218 09:41:48.282510 4556 scope.go:117] "RemoveContainer" containerID="04a4196781cc895fa1780ee09f471180228eaaff00e4cc56bf534ca86fb1d640" Feb 18 09:41:48 crc kubenswrapper[4556]: E0218 09:41:48.283685 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:41:59 crc kubenswrapper[4556]: I0218 09:41:59.283669 4556 scope.go:117] "RemoveContainer" containerID="04a4196781cc895fa1780ee09f471180228eaaff00e4cc56bf534ca86fb1d640" Feb 18 09:41:59 crc kubenswrapper[4556]: E0218 09:41:59.284778 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:42:13 crc kubenswrapper[4556]: I0218 09:42:13.289443 4556 scope.go:117] "RemoveContainer" containerID="04a4196781cc895fa1780ee09f471180228eaaff00e4cc56bf534ca86fb1d640" Feb 18 09:42:13 crc kubenswrapper[4556]: E0218 09:42:13.290489 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:42:28 crc kubenswrapper[4556]: I0218 09:42:28.283185 4556 scope.go:117] "RemoveContainer" containerID="04a4196781cc895fa1780ee09f471180228eaaff00e4cc56bf534ca86fb1d640" Feb 18 09:42:28 crc kubenswrapper[4556]: E0218 09:42:28.284075 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:42:43 crc kubenswrapper[4556]: I0218 09:42:43.288484 4556 scope.go:117] "RemoveContainer" containerID="04a4196781cc895fa1780ee09f471180228eaaff00e4cc56bf534ca86fb1d640" Feb 18 09:42:43 crc kubenswrapper[4556]: E0218 09:42:43.289434 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:42:54 crc kubenswrapper[4556]: I0218 09:42:54.282719 4556 scope.go:117] "RemoveContainer" containerID="04a4196781cc895fa1780ee09f471180228eaaff00e4cc56bf534ca86fb1d640" Feb 18 09:42:54 crc kubenswrapper[4556]: E0218 09:42:54.283915 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:43:06 crc kubenswrapper[4556]: I0218 09:43:06.283355 4556 scope.go:117] "RemoveContainer" containerID="04a4196781cc895fa1780ee09f471180228eaaff00e4cc56bf534ca86fb1d640" Feb 18 09:43:06 crc kubenswrapper[4556]: E0218 09:43:06.284646 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:43:20 crc kubenswrapper[4556]: I0218 09:43:20.282813 4556 scope.go:117] "RemoveContainer" containerID="04a4196781cc895fa1780ee09f471180228eaaff00e4cc56bf534ca86fb1d640" Feb 18 09:43:20 crc kubenswrapper[4556]: E0218 09:43:20.283843 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:43:32 crc kubenswrapper[4556]: I0218 09:43:32.283389 4556 scope.go:117] "RemoveContainer" containerID="04a4196781cc895fa1780ee09f471180228eaaff00e4cc56bf534ca86fb1d640" Feb 18 09:43:32 crc kubenswrapper[4556]: E0218 09:43:32.284479 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:43:44 crc kubenswrapper[4556]: I0218 09:43:44.283268 4556 scope.go:117] "RemoveContainer" containerID="04a4196781cc895fa1780ee09f471180228eaaff00e4cc56bf534ca86fb1d640" Feb 18 09:43:44 crc kubenswrapper[4556]: E0218 09:43:44.284480 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:43:56 crc kubenswrapper[4556]: I0218 09:43:56.283624 4556 scope.go:117] "RemoveContainer" containerID="04a4196781cc895fa1780ee09f471180228eaaff00e4cc56bf534ca86fb1d640" Feb 18 09:43:56 crc kubenswrapper[4556]: E0218 09:43:56.284865 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:43:56 crc kubenswrapper[4556]: I0218 09:43:56.784674 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kz4hb"] Feb 18 09:43:56 crc kubenswrapper[4556]: E0218 09:43:56.785055 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76fe1400-c43c-4bfd-bab3-377464edcbe3" containerName="registry-server" Feb 18 09:43:56 crc kubenswrapper[4556]: I0218 09:43:56.785076 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="76fe1400-c43c-4bfd-bab3-377464edcbe3" containerName="registry-server" Feb 18 09:43:56 crc kubenswrapper[4556]: E0218 09:43:56.785093 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76fe1400-c43c-4bfd-bab3-377464edcbe3" containerName="extract-content" Feb 18 09:43:56 crc kubenswrapper[4556]: I0218 09:43:56.785100 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="76fe1400-c43c-4bfd-bab3-377464edcbe3" containerName="extract-content" Feb 18 09:43:56 crc kubenswrapper[4556]: E0218 09:43:56.785128 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6ed3aad-6ec1-443d-8699-5703362a6131" containerName="extract-utilities" Feb 18 09:43:56 crc kubenswrapper[4556]: I0218 09:43:56.785134 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6ed3aad-6ec1-443d-8699-5703362a6131" containerName="extract-utilities" Feb 18 09:43:56 crc kubenswrapper[4556]: E0218 09:43:56.785146 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6ed3aad-6ec1-443d-8699-5703362a6131" containerName="registry-server" Feb 18 09:43:56 crc kubenswrapper[4556]: I0218 09:43:56.785168 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6ed3aad-6ec1-443d-8699-5703362a6131" containerName="registry-server" Feb 18 09:43:56 crc kubenswrapper[4556]: E0218 09:43:56.785188 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6ed3aad-6ec1-443d-8699-5703362a6131" containerName="extract-content" Feb 18 09:43:56 crc kubenswrapper[4556]: I0218 09:43:56.785195 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6ed3aad-6ec1-443d-8699-5703362a6131" containerName="extract-content" Feb 18 09:43:56 crc kubenswrapper[4556]: E0218 09:43:56.785205 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76fe1400-c43c-4bfd-bab3-377464edcbe3" containerName="extract-utilities" Feb 18 09:43:56 crc kubenswrapper[4556]: I0218 09:43:56.785211 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="76fe1400-c43c-4bfd-bab3-377464edcbe3" containerName="extract-utilities" Feb 18 09:43:56 crc kubenswrapper[4556]: I0218 09:43:56.785401 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="76fe1400-c43c-4bfd-bab3-377464edcbe3" containerName="registry-server" Feb 18 09:43:56 crc kubenswrapper[4556]: I0218 09:43:56.785419 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6ed3aad-6ec1-443d-8699-5703362a6131" containerName="registry-server" Feb 18 09:43:56 crc kubenswrapper[4556]: I0218 09:43:56.786690 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kz4hb" Feb 18 09:43:56 crc kubenswrapper[4556]: I0218 09:43:56.806258 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kz4hb"] Feb 18 09:43:56 crc kubenswrapper[4556]: I0218 09:43:56.807534 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ec28886-ec9a-4e81-888b-38bd2adbe1ec-catalog-content\") pod \"certified-operators-kz4hb\" (UID: \"7ec28886-ec9a-4e81-888b-38bd2adbe1ec\") " pod="openshift-marketplace/certified-operators-kz4hb" Feb 18 09:43:56 crc kubenswrapper[4556]: I0218 09:43:56.807627 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5r8ws\" (UniqueName: \"kubernetes.io/projected/7ec28886-ec9a-4e81-888b-38bd2adbe1ec-kube-api-access-5r8ws\") pod \"certified-operators-kz4hb\" (UID: \"7ec28886-ec9a-4e81-888b-38bd2adbe1ec\") " pod="openshift-marketplace/certified-operators-kz4hb" Feb 18 09:43:56 crc kubenswrapper[4556]: I0218 09:43:56.807653 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ec28886-ec9a-4e81-888b-38bd2adbe1ec-utilities\") pod \"certified-operators-kz4hb\" (UID: \"7ec28886-ec9a-4e81-888b-38bd2adbe1ec\") " pod="openshift-marketplace/certified-operators-kz4hb" Feb 18 09:43:56 crc kubenswrapper[4556]: I0218 09:43:56.910277 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ec28886-ec9a-4e81-888b-38bd2adbe1ec-catalog-content\") pod \"certified-operators-kz4hb\" (UID: \"7ec28886-ec9a-4e81-888b-38bd2adbe1ec\") " pod="openshift-marketplace/certified-operators-kz4hb" Feb 18 09:43:56 crc kubenswrapper[4556]: I0218 09:43:56.910714 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5r8ws\" (UniqueName: \"kubernetes.io/projected/7ec28886-ec9a-4e81-888b-38bd2adbe1ec-kube-api-access-5r8ws\") pod \"certified-operators-kz4hb\" (UID: \"7ec28886-ec9a-4e81-888b-38bd2adbe1ec\") " pod="openshift-marketplace/certified-operators-kz4hb" Feb 18 09:43:56 crc kubenswrapper[4556]: I0218 09:43:56.910772 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ec28886-ec9a-4e81-888b-38bd2adbe1ec-utilities\") pod \"certified-operators-kz4hb\" (UID: \"7ec28886-ec9a-4e81-888b-38bd2adbe1ec\") " pod="openshift-marketplace/certified-operators-kz4hb" Feb 18 09:43:56 crc kubenswrapper[4556]: I0218 09:43:56.910872 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ec28886-ec9a-4e81-888b-38bd2adbe1ec-catalog-content\") pod \"certified-operators-kz4hb\" (UID: \"7ec28886-ec9a-4e81-888b-38bd2adbe1ec\") " pod="openshift-marketplace/certified-operators-kz4hb" Feb 18 09:43:56 crc kubenswrapper[4556]: I0218 09:43:56.911291 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ec28886-ec9a-4e81-888b-38bd2adbe1ec-utilities\") pod \"certified-operators-kz4hb\" (UID: \"7ec28886-ec9a-4e81-888b-38bd2adbe1ec\") " pod="openshift-marketplace/certified-operators-kz4hb" Feb 18 09:43:56 crc kubenswrapper[4556]: I0218 09:43:56.947209 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5r8ws\" (UniqueName: \"kubernetes.io/projected/7ec28886-ec9a-4e81-888b-38bd2adbe1ec-kube-api-access-5r8ws\") pod \"certified-operators-kz4hb\" (UID: \"7ec28886-ec9a-4e81-888b-38bd2adbe1ec\") " pod="openshift-marketplace/certified-operators-kz4hb" Feb 18 09:43:57 crc kubenswrapper[4556]: I0218 09:43:57.105891 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kz4hb" Feb 18 09:43:57 crc kubenswrapper[4556]: I0218 09:43:57.537859 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kz4hb"] Feb 18 09:43:58 crc kubenswrapper[4556]: I0218 09:43:58.009444 4556 generic.go:334] "Generic (PLEG): container finished" podID="7ec28886-ec9a-4e81-888b-38bd2adbe1ec" containerID="527e88281e1e34117048512131a742284c58b095745e790a5e4bb5aefad9e140" exitCode=0 Feb 18 09:43:58 crc kubenswrapper[4556]: I0218 09:43:58.009495 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kz4hb" event={"ID":"7ec28886-ec9a-4e81-888b-38bd2adbe1ec","Type":"ContainerDied","Data":"527e88281e1e34117048512131a742284c58b095745e790a5e4bb5aefad9e140"} Feb 18 09:43:58 crc kubenswrapper[4556]: I0218 09:43:58.009528 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kz4hb" event={"ID":"7ec28886-ec9a-4e81-888b-38bd2adbe1ec","Type":"ContainerStarted","Data":"526f07a35ec9b1a5bcdf16195c28ef1abedc360f4843b1194b021243c8f870da"} Feb 18 09:43:59 crc kubenswrapper[4556]: I0218 09:43:59.020277 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kz4hb" event={"ID":"7ec28886-ec9a-4e81-888b-38bd2adbe1ec","Type":"ContainerStarted","Data":"ceb751ab9b056945bb49b858dfd5fe73d93281795ca5070d70a3207350d1332b"} Feb 18 09:44:00 crc kubenswrapper[4556]: I0218 09:44:00.037840 4556 generic.go:334] "Generic (PLEG): container finished" podID="7ec28886-ec9a-4e81-888b-38bd2adbe1ec" containerID="ceb751ab9b056945bb49b858dfd5fe73d93281795ca5070d70a3207350d1332b" exitCode=0 Feb 18 09:44:00 crc kubenswrapper[4556]: I0218 09:44:00.037949 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kz4hb" event={"ID":"7ec28886-ec9a-4e81-888b-38bd2adbe1ec","Type":"ContainerDied","Data":"ceb751ab9b056945bb49b858dfd5fe73d93281795ca5070d70a3207350d1332b"} Feb 18 09:44:01 crc kubenswrapper[4556]: I0218 09:44:01.051363 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kz4hb" event={"ID":"7ec28886-ec9a-4e81-888b-38bd2adbe1ec","Type":"ContainerStarted","Data":"2627d5854177a4a5980a53d5f666aa6c9ce0aba651546581604b1e530d6dccb8"} Feb 18 09:44:01 crc kubenswrapper[4556]: I0218 09:44:01.070148 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kz4hb" podStartSLOduration=2.559182182 podStartE2EDuration="5.070101636s" podCreationTimestamp="2026-02-18 09:43:56 +0000 UTC" firstStartedPulling="2026-02-18 09:43:58.011316779 +0000 UTC m=+2395.028277758" lastFinishedPulling="2026-02-18 09:44:00.522236232 +0000 UTC m=+2397.539197212" observedRunningTime="2026-02-18 09:44:01.068691457 +0000 UTC m=+2398.085652426" watchObservedRunningTime="2026-02-18 09:44:01.070101636 +0000 UTC m=+2398.087062616" Feb 18 09:44:07 crc kubenswrapper[4556]: I0218 09:44:07.106044 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kz4hb" Feb 18 09:44:07 crc kubenswrapper[4556]: I0218 09:44:07.106543 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kz4hb" Feb 18 09:44:07 crc kubenswrapper[4556]: I0218 09:44:07.146665 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kz4hb" Feb 18 09:44:08 crc kubenswrapper[4556]: I0218 09:44:08.159716 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kz4hb" Feb 18 09:44:08 crc kubenswrapper[4556]: I0218 09:44:08.207518 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kz4hb"] Feb 18 09:44:08 crc kubenswrapper[4556]: I0218 09:44:08.283064 4556 scope.go:117] "RemoveContainer" containerID="04a4196781cc895fa1780ee09f471180228eaaff00e4cc56bf534ca86fb1d640" Feb 18 09:44:08 crc kubenswrapper[4556]: E0218 09:44:08.283428 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:44:10 crc kubenswrapper[4556]: I0218 09:44:10.137919 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kz4hb" podUID="7ec28886-ec9a-4e81-888b-38bd2adbe1ec" containerName="registry-server" containerID="cri-o://2627d5854177a4a5980a53d5f666aa6c9ce0aba651546581604b1e530d6dccb8" gracePeriod=2 Feb 18 09:44:10 crc kubenswrapper[4556]: I0218 09:44:10.582383 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kz4hb" Feb 18 09:44:10 crc kubenswrapper[4556]: I0218 09:44:10.735706 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ec28886-ec9a-4e81-888b-38bd2adbe1ec-utilities\") pod \"7ec28886-ec9a-4e81-888b-38bd2adbe1ec\" (UID: \"7ec28886-ec9a-4e81-888b-38bd2adbe1ec\") " Feb 18 09:44:10 crc kubenswrapper[4556]: I0218 09:44:10.736317 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ec28886-ec9a-4e81-888b-38bd2adbe1ec-catalog-content\") pod \"7ec28886-ec9a-4e81-888b-38bd2adbe1ec\" (UID: \"7ec28886-ec9a-4e81-888b-38bd2adbe1ec\") " Feb 18 09:44:10 crc kubenswrapper[4556]: I0218 09:44:10.736394 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5r8ws\" (UniqueName: \"kubernetes.io/projected/7ec28886-ec9a-4e81-888b-38bd2adbe1ec-kube-api-access-5r8ws\") pod \"7ec28886-ec9a-4e81-888b-38bd2adbe1ec\" (UID: \"7ec28886-ec9a-4e81-888b-38bd2adbe1ec\") " Feb 18 09:44:10 crc kubenswrapper[4556]: I0218 09:44:10.736663 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ec28886-ec9a-4e81-888b-38bd2adbe1ec-utilities" (OuterVolumeSpecName: "utilities") pod "7ec28886-ec9a-4e81-888b-38bd2adbe1ec" (UID: "7ec28886-ec9a-4e81-888b-38bd2adbe1ec"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:44:10 crc kubenswrapper[4556]: I0218 09:44:10.737116 4556 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ec28886-ec9a-4e81-888b-38bd2adbe1ec-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 09:44:10 crc kubenswrapper[4556]: I0218 09:44:10.741435 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ec28886-ec9a-4e81-888b-38bd2adbe1ec-kube-api-access-5r8ws" (OuterVolumeSpecName: "kube-api-access-5r8ws") pod "7ec28886-ec9a-4e81-888b-38bd2adbe1ec" (UID: "7ec28886-ec9a-4e81-888b-38bd2adbe1ec"). InnerVolumeSpecName "kube-api-access-5r8ws". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:44:10 crc kubenswrapper[4556]: I0218 09:44:10.775503 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ec28886-ec9a-4e81-888b-38bd2adbe1ec-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7ec28886-ec9a-4e81-888b-38bd2adbe1ec" (UID: "7ec28886-ec9a-4e81-888b-38bd2adbe1ec"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:44:10 crc kubenswrapper[4556]: I0218 09:44:10.839867 4556 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ec28886-ec9a-4e81-888b-38bd2adbe1ec-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 09:44:10 crc kubenswrapper[4556]: I0218 09:44:10.839903 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5r8ws\" (UniqueName: \"kubernetes.io/projected/7ec28886-ec9a-4e81-888b-38bd2adbe1ec-kube-api-access-5r8ws\") on node \"crc\" DevicePath \"\"" Feb 18 09:44:11 crc kubenswrapper[4556]: I0218 09:44:11.152732 4556 generic.go:334] "Generic (PLEG): container finished" podID="7ec28886-ec9a-4e81-888b-38bd2adbe1ec" containerID="2627d5854177a4a5980a53d5f666aa6c9ce0aba651546581604b1e530d6dccb8" exitCode=0 Feb 18 09:44:11 crc kubenswrapper[4556]: I0218 09:44:11.152782 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kz4hb" event={"ID":"7ec28886-ec9a-4e81-888b-38bd2adbe1ec","Type":"ContainerDied","Data":"2627d5854177a4a5980a53d5f666aa6c9ce0aba651546581604b1e530d6dccb8"} Feb 18 09:44:11 crc kubenswrapper[4556]: I0218 09:44:11.152796 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kz4hb" Feb 18 09:44:11 crc kubenswrapper[4556]: I0218 09:44:11.152819 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kz4hb" event={"ID":"7ec28886-ec9a-4e81-888b-38bd2adbe1ec","Type":"ContainerDied","Data":"526f07a35ec9b1a5bcdf16195c28ef1abedc360f4843b1194b021243c8f870da"} Feb 18 09:44:11 crc kubenswrapper[4556]: I0218 09:44:11.152843 4556 scope.go:117] "RemoveContainer" containerID="2627d5854177a4a5980a53d5f666aa6c9ce0aba651546581604b1e530d6dccb8" Feb 18 09:44:11 crc kubenswrapper[4556]: I0218 09:44:11.185878 4556 scope.go:117] "RemoveContainer" containerID="ceb751ab9b056945bb49b858dfd5fe73d93281795ca5070d70a3207350d1332b" Feb 18 09:44:11 crc kubenswrapper[4556]: I0218 09:44:11.187928 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kz4hb"] Feb 18 09:44:11 crc kubenswrapper[4556]: I0218 09:44:11.197056 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kz4hb"] Feb 18 09:44:11 crc kubenswrapper[4556]: I0218 09:44:11.207140 4556 scope.go:117] "RemoveContainer" containerID="527e88281e1e34117048512131a742284c58b095745e790a5e4bb5aefad9e140" Feb 18 09:44:11 crc kubenswrapper[4556]: I0218 09:44:11.237374 4556 scope.go:117] "RemoveContainer" containerID="2627d5854177a4a5980a53d5f666aa6c9ce0aba651546581604b1e530d6dccb8" Feb 18 09:44:11 crc kubenswrapper[4556]: E0218 09:44:11.243550 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2627d5854177a4a5980a53d5f666aa6c9ce0aba651546581604b1e530d6dccb8\": container with ID starting with 2627d5854177a4a5980a53d5f666aa6c9ce0aba651546581604b1e530d6dccb8 not found: ID does not exist" containerID="2627d5854177a4a5980a53d5f666aa6c9ce0aba651546581604b1e530d6dccb8" Feb 18 09:44:11 crc kubenswrapper[4556]: I0218 09:44:11.243611 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2627d5854177a4a5980a53d5f666aa6c9ce0aba651546581604b1e530d6dccb8"} err="failed to get container status \"2627d5854177a4a5980a53d5f666aa6c9ce0aba651546581604b1e530d6dccb8\": rpc error: code = NotFound desc = could not find container \"2627d5854177a4a5980a53d5f666aa6c9ce0aba651546581604b1e530d6dccb8\": container with ID starting with 2627d5854177a4a5980a53d5f666aa6c9ce0aba651546581604b1e530d6dccb8 not found: ID does not exist" Feb 18 09:44:11 crc kubenswrapper[4556]: I0218 09:44:11.243646 4556 scope.go:117] "RemoveContainer" containerID="ceb751ab9b056945bb49b858dfd5fe73d93281795ca5070d70a3207350d1332b" Feb 18 09:44:11 crc kubenswrapper[4556]: E0218 09:44:11.244008 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ceb751ab9b056945bb49b858dfd5fe73d93281795ca5070d70a3207350d1332b\": container with ID starting with ceb751ab9b056945bb49b858dfd5fe73d93281795ca5070d70a3207350d1332b not found: ID does not exist" containerID="ceb751ab9b056945bb49b858dfd5fe73d93281795ca5070d70a3207350d1332b" Feb 18 09:44:11 crc kubenswrapper[4556]: I0218 09:44:11.244058 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ceb751ab9b056945bb49b858dfd5fe73d93281795ca5070d70a3207350d1332b"} err="failed to get container status \"ceb751ab9b056945bb49b858dfd5fe73d93281795ca5070d70a3207350d1332b\": rpc error: code = NotFound desc = could not find container \"ceb751ab9b056945bb49b858dfd5fe73d93281795ca5070d70a3207350d1332b\": container with ID starting with ceb751ab9b056945bb49b858dfd5fe73d93281795ca5070d70a3207350d1332b not found: ID does not exist" Feb 18 09:44:11 crc kubenswrapper[4556]: I0218 09:44:11.244091 4556 scope.go:117] "RemoveContainer" containerID="527e88281e1e34117048512131a742284c58b095745e790a5e4bb5aefad9e140" Feb 18 09:44:11 crc kubenswrapper[4556]: E0218 09:44:11.244427 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"527e88281e1e34117048512131a742284c58b095745e790a5e4bb5aefad9e140\": container with ID starting with 527e88281e1e34117048512131a742284c58b095745e790a5e4bb5aefad9e140 not found: ID does not exist" containerID="527e88281e1e34117048512131a742284c58b095745e790a5e4bb5aefad9e140" Feb 18 09:44:11 crc kubenswrapper[4556]: I0218 09:44:11.244458 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"527e88281e1e34117048512131a742284c58b095745e790a5e4bb5aefad9e140"} err="failed to get container status \"527e88281e1e34117048512131a742284c58b095745e790a5e4bb5aefad9e140\": rpc error: code = NotFound desc = could not find container \"527e88281e1e34117048512131a742284c58b095745e790a5e4bb5aefad9e140\": container with ID starting with 527e88281e1e34117048512131a742284c58b095745e790a5e4bb5aefad9e140 not found: ID does not exist" Feb 18 09:44:11 crc kubenswrapper[4556]: I0218 09:44:11.292623 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ec28886-ec9a-4e81-888b-38bd2adbe1ec" path="/var/lib/kubelet/pods/7ec28886-ec9a-4e81-888b-38bd2adbe1ec/volumes" Feb 18 09:44:11 crc kubenswrapper[4556]: E0218 09:44:11.359045 4556 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7ec28886_ec9a_4e81_888b_38bd2adbe1ec.slice/crio-526f07a35ec9b1a5bcdf16195c28ef1abedc360f4843b1194b021243c8f870da\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7ec28886_ec9a_4e81_888b_38bd2adbe1ec.slice\": RecentStats: unable to find data in memory cache]" Feb 18 09:44:20 crc kubenswrapper[4556]: I0218 09:44:20.282080 4556 scope.go:117] "RemoveContainer" containerID="04a4196781cc895fa1780ee09f471180228eaaff00e4cc56bf534ca86fb1d640" Feb 18 09:44:20 crc kubenswrapper[4556]: E0218 09:44:20.282815 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:44:34 crc kubenswrapper[4556]: I0218 09:44:34.282840 4556 scope.go:117] "RemoveContainer" containerID="04a4196781cc895fa1780ee09f471180228eaaff00e4cc56bf534ca86fb1d640" Feb 18 09:44:34 crc kubenswrapper[4556]: E0218 09:44:34.283842 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:44:45 crc kubenswrapper[4556]: I0218 09:44:45.282750 4556 scope.go:117] "RemoveContainer" containerID="04a4196781cc895fa1780ee09f471180228eaaff00e4cc56bf534ca86fb1d640" Feb 18 09:44:45 crc kubenswrapper[4556]: E0218 09:44:45.283969 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:44:59 crc kubenswrapper[4556]: I0218 09:44:59.282900 4556 scope.go:117] "RemoveContainer" containerID="04a4196781cc895fa1780ee09f471180228eaaff00e4cc56bf534ca86fb1d640" Feb 18 09:44:59 crc kubenswrapper[4556]: E0218 09:44:59.283908 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:45:00 crc kubenswrapper[4556]: I0218 09:45:00.140057 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29523465-26jw4"] Feb 18 09:45:00 crc kubenswrapper[4556]: E0218 09:45:00.140476 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ec28886-ec9a-4e81-888b-38bd2adbe1ec" containerName="registry-server" Feb 18 09:45:00 crc kubenswrapper[4556]: I0218 09:45:00.140507 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ec28886-ec9a-4e81-888b-38bd2adbe1ec" containerName="registry-server" Feb 18 09:45:00 crc kubenswrapper[4556]: E0218 09:45:00.140519 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ec28886-ec9a-4e81-888b-38bd2adbe1ec" containerName="extract-content" Feb 18 09:45:00 crc kubenswrapper[4556]: I0218 09:45:00.140525 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ec28886-ec9a-4e81-888b-38bd2adbe1ec" containerName="extract-content" Feb 18 09:45:00 crc kubenswrapper[4556]: E0218 09:45:00.140555 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ec28886-ec9a-4e81-888b-38bd2adbe1ec" containerName="extract-utilities" Feb 18 09:45:00 crc kubenswrapper[4556]: I0218 09:45:00.140564 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ec28886-ec9a-4e81-888b-38bd2adbe1ec" containerName="extract-utilities" Feb 18 09:45:00 crc kubenswrapper[4556]: I0218 09:45:00.140778 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ec28886-ec9a-4e81-888b-38bd2adbe1ec" containerName="registry-server" Feb 18 09:45:00 crc kubenswrapper[4556]: I0218 09:45:00.141461 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29523465-26jw4" Feb 18 09:45:00 crc kubenswrapper[4556]: I0218 09:45:00.143439 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 18 09:45:00 crc kubenswrapper[4556]: I0218 09:45:00.143447 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 18 09:45:00 crc kubenswrapper[4556]: I0218 09:45:00.150557 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29523465-26jw4"] Feb 18 09:45:00 crc kubenswrapper[4556]: I0218 09:45:00.175344 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llj54\" (UniqueName: \"kubernetes.io/projected/052e47c9-ab25-4b23-97c1-1512ef91ebda-kube-api-access-llj54\") pod \"collect-profiles-29523465-26jw4\" (UID: \"052e47c9-ab25-4b23-97c1-1512ef91ebda\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523465-26jw4" Feb 18 09:45:00 crc kubenswrapper[4556]: I0218 09:45:00.175411 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/052e47c9-ab25-4b23-97c1-1512ef91ebda-secret-volume\") pod \"collect-profiles-29523465-26jw4\" (UID: \"052e47c9-ab25-4b23-97c1-1512ef91ebda\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523465-26jw4" Feb 18 09:45:00 crc kubenswrapper[4556]: I0218 09:45:00.175482 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/052e47c9-ab25-4b23-97c1-1512ef91ebda-config-volume\") pod \"collect-profiles-29523465-26jw4\" (UID: \"052e47c9-ab25-4b23-97c1-1512ef91ebda\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523465-26jw4" Feb 18 09:45:00 crc kubenswrapper[4556]: I0218 09:45:00.277569 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/052e47c9-ab25-4b23-97c1-1512ef91ebda-config-volume\") pod \"collect-profiles-29523465-26jw4\" (UID: \"052e47c9-ab25-4b23-97c1-1512ef91ebda\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523465-26jw4" Feb 18 09:45:00 crc kubenswrapper[4556]: I0218 09:45:00.277795 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llj54\" (UniqueName: \"kubernetes.io/projected/052e47c9-ab25-4b23-97c1-1512ef91ebda-kube-api-access-llj54\") pod \"collect-profiles-29523465-26jw4\" (UID: \"052e47c9-ab25-4b23-97c1-1512ef91ebda\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523465-26jw4" Feb 18 09:45:00 crc kubenswrapper[4556]: I0218 09:45:00.278293 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/052e47c9-ab25-4b23-97c1-1512ef91ebda-secret-volume\") pod \"collect-profiles-29523465-26jw4\" (UID: \"052e47c9-ab25-4b23-97c1-1512ef91ebda\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523465-26jw4" Feb 18 09:45:00 crc kubenswrapper[4556]: I0218 09:45:00.278504 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/052e47c9-ab25-4b23-97c1-1512ef91ebda-config-volume\") pod \"collect-profiles-29523465-26jw4\" (UID: \"052e47c9-ab25-4b23-97c1-1512ef91ebda\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523465-26jw4" Feb 18 09:45:00 crc kubenswrapper[4556]: I0218 09:45:00.284870 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/052e47c9-ab25-4b23-97c1-1512ef91ebda-secret-volume\") pod \"collect-profiles-29523465-26jw4\" (UID: \"052e47c9-ab25-4b23-97c1-1512ef91ebda\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523465-26jw4" Feb 18 09:45:00 crc kubenswrapper[4556]: I0218 09:45:00.292765 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llj54\" (UniqueName: \"kubernetes.io/projected/052e47c9-ab25-4b23-97c1-1512ef91ebda-kube-api-access-llj54\") pod \"collect-profiles-29523465-26jw4\" (UID: \"052e47c9-ab25-4b23-97c1-1512ef91ebda\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523465-26jw4" Feb 18 09:45:00 crc kubenswrapper[4556]: I0218 09:45:00.472501 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29523465-26jw4" Feb 18 09:45:00 crc kubenswrapper[4556]: I0218 09:45:00.904087 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29523465-26jw4"] Feb 18 09:45:01 crc kubenswrapper[4556]: I0218 09:45:01.646287 4556 generic.go:334] "Generic (PLEG): container finished" podID="052e47c9-ab25-4b23-97c1-1512ef91ebda" containerID="174a6595ff5c90c2cbff52baa9c0fefd20436c8b96f48189cf97b83e81ef7748" exitCode=0 Feb 18 09:45:01 crc kubenswrapper[4556]: I0218 09:45:01.646411 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29523465-26jw4" event={"ID":"052e47c9-ab25-4b23-97c1-1512ef91ebda","Type":"ContainerDied","Data":"174a6595ff5c90c2cbff52baa9c0fefd20436c8b96f48189cf97b83e81ef7748"} Feb 18 09:45:01 crc kubenswrapper[4556]: I0218 09:45:01.646725 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29523465-26jw4" event={"ID":"052e47c9-ab25-4b23-97c1-1512ef91ebda","Type":"ContainerStarted","Data":"719fd9d13097db0acb0ba11c8639e46f15738b17ce10fa2c775e312390d66640"} Feb 18 09:45:02 crc kubenswrapper[4556]: I0218 09:45:02.989203 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29523465-26jw4" Feb 18 09:45:03 crc kubenswrapper[4556]: I0218 09:45:03.040135 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/052e47c9-ab25-4b23-97c1-1512ef91ebda-secret-volume\") pod \"052e47c9-ab25-4b23-97c1-1512ef91ebda\" (UID: \"052e47c9-ab25-4b23-97c1-1512ef91ebda\") " Feb 18 09:45:03 crc kubenswrapper[4556]: I0218 09:45:03.040269 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-llj54\" (UniqueName: \"kubernetes.io/projected/052e47c9-ab25-4b23-97c1-1512ef91ebda-kube-api-access-llj54\") pod \"052e47c9-ab25-4b23-97c1-1512ef91ebda\" (UID: \"052e47c9-ab25-4b23-97c1-1512ef91ebda\") " Feb 18 09:45:03 crc kubenswrapper[4556]: I0218 09:45:03.040306 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/052e47c9-ab25-4b23-97c1-1512ef91ebda-config-volume\") pod \"052e47c9-ab25-4b23-97c1-1512ef91ebda\" (UID: \"052e47c9-ab25-4b23-97c1-1512ef91ebda\") " Feb 18 09:45:03 crc kubenswrapper[4556]: I0218 09:45:03.040974 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/052e47c9-ab25-4b23-97c1-1512ef91ebda-config-volume" (OuterVolumeSpecName: "config-volume") pod "052e47c9-ab25-4b23-97c1-1512ef91ebda" (UID: "052e47c9-ab25-4b23-97c1-1512ef91ebda"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:45:03 crc kubenswrapper[4556]: I0218 09:45:03.045784 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/052e47c9-ab25-4b23-97c1-1512ef91ebda-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "052e47c9-ab25-4b23-97c1-1512ef91ebda" (UID: "052e47c9-ab25-4b23-97c1-1512ef91ebda"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:45:03 crc kubenswrapper[4556]: I0218 09:45:03.047210 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/052e47c9-ab25-4b23-97c1-1512ef91ebda-kube-api-access-llj54" (OuterVolumeSpecName: "kube-api-access-llj54") pod "052e47c9-ab25-4b23-97c1-1512ef91ebda" (UID: "052e47c9-ab25-4b23-97c1-1512ef91ebda"). InnerVolumeSpecName "kube-api-access-llj54". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:45:03 crc kubenswrapper[4556]: I0218 09:45:03.144791 4556 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/052e47c9-ab25-4b23-97c1-1512ef91ebda-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 18 09:45:03 crc kubenswrapper[4556]: I0218 09:45:03.144835 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-llj54\" (UniqueName: \"kubernetes.io/projected/052e47c9-ab25-4b23-97c1-1512ef91ebda-kube-api-access-llj54\") on node \"crc\" DevicePath \"\"" Feb 18 09:45:03 crc kubenswrapper[4556]: I0218 09:45:03.144847 4556 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/052e47c9-ab25-4b23-97c1-1512ef91ebda-config-volume\") on node \"crc\" DevicePath \"\"" Feb 18 09:45:03 crc kubenswrapper[4556]: I0218 09:45:03.667901 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29523465-26jw4" event={"ID":"052e47c9-ab25-4b23-97c1-1512ef91ebda","Type":"ContainerDied","Data":"719fd9d13097db0acb0ba11c8639e46f15738b17ce10fa2c775e312390d66640"} Feb 18 09:45:03 crc kubenswrapper[4556]: I0218 09:45:03.668411 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="719fd9d13097db0acb0ba11c8639e46f15738b17ce10fa2c775e312390d66640" Feb 18 09:45:03 crc kubenswrapper[4556]: I0218 09:45:03.668186 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29523465-26jw4" Feb 18 09:45:04 crc kubenswrapper[4556]: I0218 09:45:04.054588 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29523420-qb78x"] Feb 18 09:45:04 crc kubenswrapper[4556]: I0218 09:45:04.061758 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29523420-qb78x"] Feb 18 09:45:05 crc kubenswrapper[4556]: I0218 09:45:05.295028 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0aa2e566-fd95-498b-8df5-b63afb290572" path="/var/lib/kubelet/pods/0aa2e566-fd95-498b-8df5-b63afb290572/volumes" Feb 18 09:45:11 crc kubenswrapper[4556]: I0218 09:45:11.133891 4556 scope.go:117] "RemoveContainer" containerID="c2201de00ba0f9c778a8b2bbece184c31365c38bd6019f97cef435b6315fc2ab" Feb 18 09:45:12 crc kubenswrapper[4556]: I0218 09:45:12.282740 4556 scope.go:117] "RemoveContainer" containerID="04a4196781cc895fa1780ee09f471180228eaaff00e4cc56bf534ca86fb1d640" Feb 18 09:45:12 crc kubenswrapper[4556]: E0218 09:45:12.283109 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:45:12 crc kubenswrapper[4556]: E0218 09:45:12.672024 4556 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod052e47c9_ab25_4b23_97c1_1512ef91ebda.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod052e47c9_ab25_4b23_97c1_1512ef91ebda.slice/crio-719fd9d13097db0acb0ba11c8639e46f15738b17ce10fa2c775e312390d66640\": RecentStats: unable to find data in memory cache]" Feb 18 09:45:22 crc kubenswrapper[4556]: E0218 09:45:22.885435 4556 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod052e47c9_ab25_4b23_97c1_1512ef91ebda.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod052e47c9_ab25_4b23_97c1_1512ef91ebda.slice/crio-719fd9d13097db0acb0ba11c8639e46f15738b17ce10fa2c775e312390d66640\": RecentStats: unable to find data in memory cache]" Feb 18 09:45:26 crc kubenswrapper[4556]: I0218 09:45:26.282270 4556 scope.go:117] "RemoveContainer" containerID="04a4196781cc895fa1780ee09f471180228eaaff00e4cc56bf534ca86fb1d640" Feb 18 09:45:26 crc kubenswrapper[4556]: E0218 09:45:26.283065 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:45:33 crc kubenswrapper[4556]: E0218 09:45:33.110643 4556 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod052e47c9_ab25_4b23_97c1_1512ef91ebda.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod052e47c9_ab25_4b23_97c1_1512ef91ebda.slice/crio-719fd9d13097db0acb0ba11c8639e46f15738b17ce10fa2c775e312390d66640\": RecentStats: unable to find data in memory cache]" Feb 18 09:45:37 crc kubenswrapper[4556]: I0218 09:45:37.282800 4556 scope.go:117] "RemoveContainer" containerID="04a4196781cc895fa1780ee09f471180228eaaff00e4cc56bf534ca86fb1d640" Feb 18 09:45:38 crc kubenswrapper[4556]: I0218 09:45:38.018476 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" event={"ID":"8dac7f27-d3d1-4778-9e54-f273035a1d37","Type":"ContainerStarted","Data":"4267588c8394174e9461033b074def2dd2f51cf44cffb9ce5d5bceee1db04806"} Feb 18 09:45:43 crc kubenswrapper[4556]: E0218 09:45:43.337712 4556 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod052e47c9_ab25_4b23_97c1_1512ef91ebda.slice/crio-719fd9d13097db0acb0ba11c8639e46f15738b17ce10fa2c775e312390d66640\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod052e47c9_ab25_4b23_97c1_1512ef91ebda.slice\": RecentStats: unable to find data in memory cache]" Feb 18 09:45:53 crc kubenswrapper[4556]: E0218 09:45:53.543510 4556 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod052e47c9_ab25_4b23_97c1_1512ef91ebda.slice/crio-719fd9d13097db0acb0ba11c8639e46f15738b17ce10fa2c775e312390d66640\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod052e47c9_ab25_4b23_97c1_1512ef91ebda.slice\": RecentStats: unable to find data in memory cache]" Feb 18 09:48:01 crc kubenswrapper[4556]: I0218 09:48:01.727612 4556 patch_prober.go:28] interesting pod/machine-config-daemon-f76hs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 09:48:01 crc kubenswrapper[4556]: I0218 09:48:01.728472 4556 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 09:48:31 crc kubenswrapper[4556]: I0218 09:48:31.727535 4556 patch_prober.go:28] interesting pod/machine-config-daemon-f76hs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 09:48:31 crc kubenswrapper[4556]: I0218 09:48:31.728205 4556 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 09:48:35 crc kubenswrapper[4556]: I0218 09:48:35.707560 4556 generic.go:334] "Generic (PLEG): container finished" podID="7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59" containerID="3cd1bb4b8d0bc13d13b1197e1a12491e89329de392c8c6c6c3c344d4dbf9b3e6" exitCode=0 Feb 18 09:48:35 crc kubenswrapper[4556]: I0218 09:48:35.707659 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59","Type":"ContainerDied","Data":"3cd1bb4b8d0bc13d13b1197e1a12491e89329de392c8c6c6c3c344d4dbf9b3e6"} Feb 18 09:48:37 crc kubenswrapper[4556]: I0218 09:48:37.058547 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Feb 18 09:48:37 crc kubenswrapper[4556]: I0218 09:48:37.199297 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6j8k\" (UniqueName: \"kubernetes.io/projected/7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59-kube-api-access-x6j8k\") pod \"7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59\" (UID: \"7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59\") " Feb 18 09:48:37 crc kubenswrapper[4556]: I0218 09:48:37.199366 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59\" (UID: \"7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59\") " Feb 18 09:48:37 crc kubenswrapper[4556]: I0218 09:48:37.199411 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59-openstack-config-secret\") pod \"7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59\" (UID: \"7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59\") " Feb 18 09:48:37 crc kubenswrapper[4556]: I0218 09:48:37.200001 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59-ssh-key\") pod \"7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59\" (UID: \"7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59\") " Feb 18 09:48:37 crc kubenswrapper[4556]: I0218 09:48:37.200035 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59-test-operator-ephemeral-temporary\") pod \"7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59\" (UID: \"7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59\") " Feb 18 09:48:37 crc kubenswrapper[4556]: I0218 09:48:37.200054 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59-openstack-config\") pod \"7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59\" (UID: \"7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59\") " Feb 18 09:48:37 crc kubenswrapper[4556]: I0218 09:48:37.200078 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59-ca-certs\") pod \"7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59\" (UID: \"7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59\") " Feb 18 09:48:37 crc kubenswrapper[4556]: I0218 09:48:37.200133 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59-test-operator-ephemeral-workdir\") pod \"7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59\" (UID: \"7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59\") " Feb 18 09:48:37 crc kubenswrapper[4556]: I0218 09:48:37.200244 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59-config-data\") pod \"7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59\" (UID: \"7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59\") " Feb 18 09:48:37 crc kubenswrapper[4556]: I0218 09:48:37.201532 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59-config-data" (OuterVolumeSpecName: "config-data") pod "7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59" (UID: "7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:48:37 crc kubenswrapper[4556]: I0218 09:48:37.202854 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59" (UID: "7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:48:37 crc kubenswrapper[4556]: I0218 09:48:37.205368 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59" (UID: "7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:48:37 crc kubenswrapper[4556]: I0218 09:48:37.206505 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "test-operator-logs") pod "7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59" (UID: "7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 18 09:48:37 crc kubenswrapper[4556]: I0218 09:48:37.213413 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59-kube-api-access-x6j8k" (OuterVolumeSpecName: "kube-api-access-x6j8k") pod "7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59" (UID: "7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59"). InnerVolumeSpecName "kube-api-access-x6j8k". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:48:37 crc kubenswrapper[4556]: I0218 09:48:37.224817 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59" (UID: "7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:48:37 crc kubenswrapper[4556]: I0218 09:48:37.225829 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59" (UID: "7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:48:37 crc kubenswrapper[4556]: I0218 09:48:37.228985 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59" (UID: "7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 09:48:37 crc kubenswrapper[4556]: I0218 09:48:37.245407 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59" (UID: "7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 09:48:37 crc kubenswrapper[4556]: I0218 09:48:37.302383 4556 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59-config-data\") on node \"crc\" DevicePath \"\"" Feb 18 09:48:37 crc kubenswrapper[4556]: I0218 09:48:37.302409 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6j8k\" (UniqueName: \"kubernetes.io/projected/7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59-kube-api-access-x6j8k\") on node \"crc\" DevicePath \"\"" Feb 18 09:48:37 crc kubenswrapper[4556]: I0218 09:48:37.302448 4556 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Feb 18 09:48:37 crc kubenswrapper[4556]: I0218 09:48:37.302459 4556 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Feb 18 09:48:37 crc kubenswrapper[4556]: I0218 09:48:37.302471 4556 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59-ssh-key\") on node \"crc\" DevicePath \"\"" Feb 18 09:48:37 crc kubenswrapper[4556]: I0218 09:48:37.302480 4556 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59-openstack-config\") on node \"crc\" DevicePath \"\"" Feb 18 09:48:37 crc kubenswrapper[4556]: I0218 09:48:37.302582 4556 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Feb 18 09:48:37 crc kubenswrapper[4556]: I0218 09:48:37.302596 4556 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59-ca-certs\") on node \"crc\" DevicePath \"\"" Feb 18 09:48:37 crc kubenswrapper[4556]: I0218 09:48:37.302609 4556 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Feb 18 09:48:37 crc kubenswrapper[4556]: I0218 09:48:37.322301 4556 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Feb 18 09:48:37 crc kubenswrapper[4556]: I0218 09:48:37.405087 4556 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Feb 18 09:48:37 crc kubenswrapper[4556]: I0218 09:48:37.730779 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59","Type":"ContainerDied","Data":"aeca925c01875991035f5d939d3e2f0d20910b37f24006ee7a735490518a300b"} Feb 18 09:48:37 crc kubenswrapper[4556]: I0218 09:48:37.730856 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aeca925c01875991035f5d939d3e2f0d20910b37f24006ee7a735490518a300b" Feb 18 09:48:37 crc kubenswrapper[4556]: I0218 09:48:37.730888 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Feb 18 09:48:42 crc kubenswrapper[4556]: I0218 09:48:42.473742 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Feb 18 09:48:42 crc kubenswrapper[4556]: E0218 09:48:42.474526 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="052e47c9-ab25-4b23-97c1-1512ef91ebda" containerName="collect-profiles" Feb 18 09:48:42 crc kubenswrapper[4556]: I0218 09:48:42.474544 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="052e47c9-ab25-4b23-97c1-1512ef91ebda" containerName="collect-profiles" Feb 18 09:48:42 crc kubenswrapper[4556]: E0218 09:48:42.474569 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59" containerName="tempest-tests-tempest-tests-runner" Feb 18 09:48:42 crc kubenswrapper[4556]: I0218 09:48:42.474576 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59" containerName="tempest-tests-tempest-tests-runner" Feb 18 09:48:42 crc kubenswrapper[4556]: I0218 09:48:42.474784 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="052e47c9-ab25-4b23-97c1-1512ef91ebda" containerName="collect-profiles" Feb 18 09:48:42 crc kubenswrapper[4556]: I0218 09:48:42.474808 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59" containerName="tempest-tests-tempest-tests-runner" Feb 18 09:48:42 crc kubenswrapper[4556]: I0218 09:48:42.475504 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 18 09:48:42 crc kubenswrapper[4556]: I0218 09:48:42.478505 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-9vt27" Feb 18 09:48:42 crc kubenswrapper[4556]: I0218 09:48:42.481014 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Feb 18 09:48:42 crc kubenswrapper[4556]: I0218 09:48:42.642019 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgqrh\" (UniqueName: \"kubernetes.io/projected/396883e0-c2e8-4698-97d2-70bbd5fe7b7f-kube-api-access-xgqrh\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"396883e0-c2e8-4698-97d2-70bbd5fe7b7f\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 18 09:48:42 crc kubenswrapper[4556]: I0218 09:48:42.642193 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"396883e0-c2e8-4698-97d2-70bbd5fe7b7f\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 18 09:48:42 crc kubenswrapper[4556]: I0218 09:48:42.744087 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgqrh\" (UniqueName: \"kubernetes.io/projected/396883e0-c2e8-4698-97d2-70bbd5fe7b7f-kube-api-access-xgqrh\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"396883e0-c2e8-4698-97d2-70bbd5fe7b7f\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 18 09:48:42 crc kubenswrapper[4556]: I0218 09:48:42.744522 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"396883e0-c2e8-4698-97d2-70bbd5fe7b7f\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 18 09:48:42 crc kubenswrapper[4556]: I0218 09:48:42.745080 4556 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"396883e0-c2e8-4698-97d2-70bbd5fe7b7f\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 18 09:48:42 crc kubenswrapper[4556]: I0218 09:48:42.779583 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgqrh\" (UniqueName: \"kubernetes.io/projected/396883e0-c2e8-4698-97d2-70bbd5fe7b7f-kube-api-access-xgqrh\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"396883e0-c2e8-4698-97d2-70bbd5fe7b7f\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 18 09:48:42 crc kubenswrapper[4556]: I0218 09:48:42.787920 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"396883e0-c2e8-4698-97d2-70bbd5fe7b7f\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 18 09:48:42 crc kubenswrapper[4556]: I0218 09:48:42.797758 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 18 09:48:43 crc kubenswrapper[4556]: I0218 09:48:43.184329 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Feb 18 09:48:43 crc kubenswrapper[4556]: I0218 09:48:43.187379 4556 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 18 09:48:43 crc kubenswrapper[4556]: I0218 09:48:43.797559 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"396883e0-c2e8-4698-97d2-70bbd5fe7b7f","Type":"ContainerStarted","Data":"f137066bd210df5420dcc6c30cf4554f5a9b39c30ede2363ab88358417e31c03"} Feb 18 09:48:44 crc kubenswrapper[4556]: I0218 09:48:44.810935 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"396883e0-c2e8-4698-97d2-70bbd5fe7b7f","Type":"ContainerStarted","Data":"0723d1498a76cf4d084692354620f878fefef4ce432f7b7f73c4508a49f38e21"} Feb 18 09:48:44 crc kubenswrapper[4556]: I0218 09:48:44.832480 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.05903535 podStartE2EDuration="2.832463432s" podCreationTimestamp="2026-02-18 09:48:42 +0000 UTC" firstStartedPulling="2026-02-18 09:48:43.187119277 +0000 UTC m=+2680.204080247" lastFinishedPulling="2026-02-18 09:48:43.960547348 +0000 UTC m=+2680.977508329" observedRunningTime="2026-02-18 09:48:44.823714558 +0000 UTC m=+2681.840675539" watchObservedRunningTime="2026-02-18 09:48:44.832463432 +0000 UTC m=+2681.849424412" Feb 18 09:49:01 crc kubenswrapper[4556]: I0218 09:49:01.727435 4556 patch_prober.go:28] interesting pod/machine-config-daemon-f76hs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 09:49:01 crc kubenswrapper[4556]: I0218 09:49:01.728193 4556 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 09:49:01 crc kubenswrapper[4556]: I0218 09:49:01.728250 4556 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" Feb 18 09:49:01 crc kubenswrapper[4556]: I0218 09:49:01.729324 4556 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4267588c8394174e9461033b074def2dd2f51cf44cffb9ce5d5bceee1db04806"} pod="openshift-machine-config-operator/machine-config-daemon-f76hs" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 18 09:49:01 crc kubenswrapper[4556]: I0218 09:49:01.729397 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerName="machine-config-daemon" containerID="cri-o://4267588c8394174e9461033b074def2dd2f51cf44cffb9ce5d5bceee1db04806" gracePeriod=600 Feb 18 09:49:01 crc kubenswrapper[4556]: I0218 09:49:01.981516 4556 generic.go:334] "Generic (PLEG): container finished" podID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerID="4267588c8394174e9461033b074def2dd2f51cf44cffb9ce5d5bceee1db04806" exitCode=0 Feb 18 09:49:01 crc kubenswrapper[4556]: I0218 09:49:01.981824 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" event={"ID":"8dac7f27-d3d1-4778-9e54-f273035a1d37","Type":"ContainerDied","Data":"4267588c8394174e9461033b074def2dd2f51cf44cffb9ce5d5bceee1db04806"} Feb 18 09:49:01 crc kubenswrapper[4556]: I0218 09:49:01.981863 4556 scope.go:117] "RemoveContainer" containerID="04a4196781cc895fa1780ee09f471180228eaaff00e4cc56bf534ca86fb1d640" Feb 18 09:49:02 crc kubenswrapper[4556]: I0218 09:49:02.956228 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-8r4vv/must-gather-2lndc"] Feb 18 09:49:02 crc kubenswrapper[4556]: I0218 09:49:02.958485 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8r4vv/must-gather-2lndc" Feb 18 09:49:02 crc kubenswrapper[4556]: I0218 09:49:02.962812 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-8r4vv"/"openshift-service-ca.crt" Feb 18 09:49:02 crc kubenswrapper[4556]: I0218 09:49:02.963382 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-8r4vv"/"kube-root-ca.crt" Feb 18 09:49:02 crc kubenswrapper[4556]: I0218 09:49:02.972550 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-8r4vv/must-gather-2lndc"] Feb 18 09:49:03 crc kubenswrapper[4556]: I0218 09:49:03.009517 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" event={"ID":"8dac7f27-d3d1-4778-9e54-f273035a1d37","Type":"ContainerStarted","Data":"9627d5bcc0cabda20a5fc92b3cc4b3dfae93225012caf89894f557937f69e6f7"} Feb 18 09:49:03 crc kubenswrapper[4556]: I0218 09:49:03.026264 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grhf9\" (UniqueName: \"kubernetes.io/projected/72cad311-e6f4-4e65-838a-59725c2e7ac8-kube-api-access-grhf9\") pod \"must-gather-2lndc\" (UID: \"72cad311-e6f4-4e65-838a-59725c2e7ac8\") " pod="openshift-must-gather-8r4vv/must-gather-2lndc" Feb 18 09:49:03 crc kubenswrapper[4556]: I0218 09:49:03.026379 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/72cad311-e6f4-4e65-838a-59725c2e7ac8-must-gather-output\") pod \"must-gather-2lndc\" (UID: \"72cad311-e6f4-4e65-838a-59725c2e7ac8\") " pod="openshift-must-gather-8r4vv/must-gather-2lndc" Feb 18 09:49:03 crc kubenswrapper[4556]: I0218 09:49:03.128804 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grhf9\" (UniqueName: \"kubernetes.io/projected/72cad311-e6f4-4e65-838a-59725c2e7ac8-kube-api-access-grhf9\") pod \"must-gather-2lndc\" (UID: \"72cad311-e6f4-4e65-838a-59725c2e7ac8\") " pod="openshift-must-gather-8r4vv/must-gather-2lndc" Feb 18 09:49:03 crc kubenswrapper[4556]: I0218 09:49:03.128985 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/72cad311-e6f4-4e65-838a-59725c2e7ac8-must-gather-output\") pod \"must-gather-2lndc\" (UID: \"72cad311-e6f4-4e65-838a-59725c2e7ac8\") " pod="openshift-must-gather-8r4vv/must-gather-2lndc" Feb 18 09:49:03 crc kubenswrapper[4556]: I0218 09:49:03.129520 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/72cad311-e6f4-4e65-838a-59725c2e7ac8-must-gather-output\") pod \"must-gather-2lndc\" (UID: \"72cad311-e6f4-4e65-838a-59725c2e7ac8\") " pod="openshift-must-gather-8r4vv/must-gather-2lndc" Feb 18 09:49:03 crc kubenswrapper[4556]: I0218 09:49:03.149303 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grhf9\" (UniqueName: \"kubernetes.io/projected/72cad311-e6f4-4e65-838a-59725c2e7ac8-kube-api-access-grhf9\") pod \"must-gather-2lndc\" (UID: \"72cad311-e6f4-4e65-838a-59725c2e7ac8\") " pod="openshift-must-gather-8r4vv/must-gather-2lndc" Feb 18 09:49:03 crc kubenswrapper[4556]: I0218 09:49:03.274520 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8r4vv/must-gather-2lndc" Feb 18 09:49:03 crc kubenswrapper[4556]: I0218 09:49:03.730114 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-8r4vv/must-gather-2lndc"] Feb 18 09:49:03 crc kubenswrapper[4556]: W0218 09:49:03.734518 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod72cad311_e6f4_4e65_838a_59725c2e7ac8.slice/crio-597ba29e475841cebacc914fc6e32e60dd71bdbce4d7d99d64a4686040a8ce83 WatchSource:0}: Error finding container 597ba29e475841cebacc914fc6e32e60dd71bdbce4d7d99d64a4686040a8ce83: Status 404 returned error can't find the container with id 597ba29e475841cebacc914fc6e32e60dd71bdbce4d7d99d64a4686040a8ce83 Feb 18 09:49:04 crc kubenswrapper[4556]: I0218 09:49:04.022565 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8r4vv/must-gather-2lndc" event={"ID":"72cad311-e6f4-4e65-838a-59725c2e7ac8","Type":"ContainerStarted","Data":"597ba29e475841cebacc914fc6e32e60dd71bdbce4d7d99d64a4686040a8ce83"} Feb 18 09:49:10 crc kubenswrapper[4556]: I0218 09:49:10.093586 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8r4vv/must-gather-2lndc" event={"ID":"72cad311-e6f4-4e65-838a-59725c2e7ac8","Type":"ContainerStarted","Data":"d93cdcf0d4fb6b9556819af66d583dc3838bea3598a0a9acc1da2b76bfbb94a4"} Feb 18 09:49:10 crc kubenswrapper[4556]: I0218 09:49:10.094304 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8r4vv/must-gather-2lndc" event={"ID":"72cad311-e6f4-4e65-838a-59725c2e7ac8","Type":"ContainerStarted","Data":"9e745c73336e6b31f33529723106acdb0d7d13100fec9a4976db3a99ec08210e"} Feb 18 09:49:10 crc kubenswrapper[4556]: I0218 09:49:10.114899 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-8r4vv/must-gather-2lndc" podStartSLOduration=2.267918605 podStartE2EDuration="8.114882691s" podCreationTimestamp="2026-02-18 09:49:02 +0000 UTC" firstStartedPulling="2026-02-18 09:49:03.736902697 +0000 UTC m=+2700.753863677" lastFinishedPulling="2026-02-18 09:49:09.583866782 +0000 UTC m=+2706.600827763" observedRunningTime="2026-02-18 09:49:10.110415104 +0000 UTC m=+2707.127376073" watchObservedRunningTime="2026-02-18 09:49:10.114882691 +0000 UTC m=+2707.131843671" Feb 18 09:49:12 crc kubenswrapper[4556]: I0218 09:49:12.791826 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-8r4vv/crc-debug-s8g6x"] Feb 18 09:49:12 crc kubenswrapper[4556]: I0218 09:49:12.794087 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8r4vv/crc-debug-s8g6x" Feb 18 09:49:12 crc kubenswrapper[4556]: I0218 09:49:12.796008 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-8r4vv"/"default-dockercfg-k49l5" Feb 18 09:49:12 crc kubenswrapper[4556]: I0218 09:49:12.943030 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f33854c8-5191-4d57-ad19-68972f69f793-host\") pod \"crc-debug-s8g6x\" (UID: \"f33854c8-5191-4d57-ad19-68972f69f793\") " pod="openshift-must-gather-8r4vv/crc-debug-s8g6x" Feb 18 09:49:12 crc kubenswrapper[4556]: I0218 09:49:12.943199 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6x5k\" (UniqueName: \"kubernetes.io/projected/f33854c8-5191-4d57-ad19-68972f69f793-kube-api-access-t6x5k\") pod \"crc-debug-s8g6x\" (UID: \"f33854c8-5191-4d57-ad19-68972f69f793\") " pod="openshift-must-gather-8r4vv/crc-debug-s8g6x" Feb 18 09:49:13 crc kubenswrapper[4556]: I0218 09:49:13.045983 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f33854c8-5191-4d57-ad19-68972f69f793-host\") pod \"crc-debug-s8g6x\" (UID: \"f33854c8-5191-4d57-ad19-68972f69f793\") " pod="openshift-must-gather-8r4vv/crc-debug-s8g6x" Feb 18 09:49:13 crc kubenswrapper[4556]: I0218 09:49:13.046142 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6x5k\" (UniqueName: \"kubernetes.io/projected/f33854c8-5191-4d57-ad19-68972f69f793-kube-api-access-t6x5k\") pod \"crc-debug-s8g6x\" (UID: \"f33854c8-5191-4d57-ad19-68972f69f793\") " pod="openshift-must-gather-8r4vv/crc-debug-s8g6x" Feb 18 09:49:13 crc kubenswrapper[4556]: I0218 09:49:13.046182 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f33854c8-5191-4d57-ad19-68972f69f793-host\") pod \"crc-debug-s8g6x\" (UID: \"f33854c8-5191-4d57-ad19-68972f69f793\") " pod="openshift-must-gather-8r4vv/crc-debug-s8g6x" Feb 18 09:49:13 crc kubenswrapper[4556]: I0218 09:49:13.069196 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6x5k\" (UniqueName: \"kubernetes.io/projected/f33854c8-5191-4d57-ad19-68972f69f793-kube-api-access-t6x5k\") pod \"crc-debug-s8g6x\" (UID: \"f33854c8-5191-4d57-ad19-68972f69f793\") " pod="openshift-must-gather-8r4vv/crc-debug-s8g6x" Feb 18 09:49:13 crc kubenswrapper[4556]: I0218 09:49:13.111469 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8r4vv/crc-debug-s8g6x" Feb 18 09:49:14 crc kubenswrapper[4556]: I0218 09:49:14.130620 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8r4vv/crc-debug-s8g6x" event={"ID":"f33854c8-5191-4d57-ad19-68972f69f793","Type":"ContainerStarted","Data":"707e09c423c339214418bf5628c61eb0c5ad5eb25b65401a88e9bf41ee3615a1"} Feb 18 09:49:23 crc kubenswrapper[4556]: I0218 09:49:23.227980 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8r4vv/crc-debug-s8g6x" event={"ID":"f33854c8-5191-4d57-ad19-68972f69f793","Type":"ContainerStarted","Data":"3a600196a112d7e302310bb03411499798e6c8abb5aff426684dd84e400ad0d3"} Feb 18 09:49:23 crc kubenswrapper[4556]: I0218 09:49:23.245987 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-8r4vv/crc-debug-s8g6x" podStartSLOduration=1.96543139 podStartE2EDuration="11.245969407s" podCreationTimestamp="2026-02-18 09:49:12 +0000 UTC" firstStartedPulling="2026-02-18 09:49:13.167007117 +0000 UTC m=+2710.183968097" lastFinishedPulling="2026-02-18 09:49:22.447545135 +0000 UTC m=+2719.464506114" observedRunningTime="2026-02-18 09:49:23.241520135 +0000 UTC m=+2720.258481115" watchObservedRunningTime="2026-02-18 09:49:23.245969407 +0000 UTC m=+2720.262930388" Feb 18 09:49:54 crc kubenswrapper[4556]: I0218 09:49:54.542561 4556 generic.go:334] "Generic (PLEG): container finished" podID="f33854c8-5191-4d57-ad19-68972f69f793" containerID="3a600196a112d7e302310bb03411499798e6c8abb5aff426684dd84e400ad0d3" exitCode=0 Feb 18 09:49:54 crc kubenswrapper[4556]: I0218 09:49:54.542652 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8r4vv/crc-debug-s8g6x" event={"ID":"f33854c8-5191-4d57-ad19-68972f69f793","Type":"ContainerDied","Data":"3a600196a112d7e302310bb03411499798e6c8abb5aff426684dd84e400ad0d3"} Feb 18 09:49:55 crc kubenswrapper[4556]: I0218 09:49:55.648537 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8r4vv/crc-debug-s8g6x" Feb 18 09:49:55 crc kubenswrapper[4556]: I0218 09:49:55.681310 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-8r4vv/crc-debug-s8g6x"] Feb 18 09:49:55 crc kubenswrapper[4556]: I0218 09:49:55.689251 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-8r4vv/crc-debug-s8g6x"] Feb 18 09:49:55 crc kubenswrapper[4556]: I0218 09:49:55.718555 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f33854c8-5191-4d57-ad19-68972f69f793-host\") pod \"f33854c8-5191-4d57-ad19-68972f69f793\" (UID: \"f33854c8-5191-4d57-ad19-68972f69f793\") " Feb 18 09:49:55 crc kubenswrapper[4556]: I0218 09:49:55.718678 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f33854c8-5191-4d57-ad19-68972f69f793-host" (OuterVolumeSpecName: "host") pod "f33854c8-5191-4d57-ad19-68972f69f793" (UID: "f33854c8-5191-4d57-ad19-68972f69f793"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 09:49:55 crc kubenswrapper[4556]: I0218 09:49:55.719146 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t6x5k\" (UniqueName: \"kubernetes.io/projected/f33854c8-5191-4d57-ad19-68972f69f793-kube-api-access-t6x5k\") pod \"f33854c8-5191-4d57-ad19-68972f69f793\" (UID: \"f33854c8-5191-4d57-ad19-68972f69f793\") " Feb 18 09:49:55 crc kubenswrapper[4556]: I0218 09:49:55.719918 4556 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f33854c8-5191-4d57-ad19-68972f69f793-host\") on node \"crc\" DevicePath \"\"" Feb 18 09:49:55 crc kubenswrapper[4556]: I0218 09:49:55.725031 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f33854c8-5191-4d57-ad19-68972f69f793-kube-api-access-t6x5k" (OuterVolumeSpecName: "kube-api-access-t6x5k") pod "f33854c8-5191-4d57-ad19-68972f69f793" (UID: "f33854c8-5191-4d57-ad19-68972f69f793"). InnerVolumeSpecName "kube-api-access-t6x5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:49:55 crc kubenswrapper[4556]: I0218 09:49:55.822180 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t6x5k\" (UniqueName: \"kubernetes.io/projected/f33854c8-5191-4d57-ad19-68972f69f793-kube-api-access-t6x5k\") on node \"crc\" DevicePath \"\"" Feb 18 09:49:56 crc kubenswrapper[4556]: I0218 09:49:56.563206 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="707e09c423c339214418bf5628c61eb0c5ad5eb25b65401a88e9bf41ee3615a1" Feb 18 09:49:56 crc kubenswrapper[4556]: I0218 09:49:56.563295 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8r4vv/crc-debug-s8g6x" Feb 18 09:49:56 crc kubenswrapper[4556]: I0218 09:49:56.840672 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-8r4vv/crc-debug-f2656"] Feb 18 09:49:56 crc kubenswrapper[4556]: E0218 09:49:56.841072 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f33854c8-5191-4d57-ad19-68972f69f793" containerName="container-00" Feb 18 09:49:56 crc kubenswrapper[4556]: I0218 09:49:56.841086 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="f33854c8-5191-4d57-ad19-68972f69f793" containerName="container-00" Feb 18 09:49:56 crc kubenswrapper[4556]: I0218 09:49:56.841287 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="f33854c8-5191-4d57-ad19-68972f69f793" containerName="container-00" Feb 18 09:49:56 crc kubenswrapper[4556]: I0218 09:49:56.841892 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8r4vv/crc-debug-f2656" Feb 18 09:49:56 crc kubenswrapper[4556]: I0218 09:49:56.846970 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-8r4vv"/"default-dockercfg-k49l5" Feb 18 09:49:56 crc kubenswrapper[4556]: I0218 09:49:56.939723 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/318c71d0-b625-4f30-9909-f157d416a123-host\") pod \"crc-debug-f2656\" (UID: \"318c71d0-b625-4f30-9909-f157d416a123\") " pod="openshift-must-gather-8r4vv/crc-debug-f2656" Feb 18 09:49:56 crc kubenswrapper[4556]: I0218 09:49:56.939804 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plhtc\" (UniqueName: \"kubernetes.io/projected/318c71d0-b625-4f30-9909-f157d416a123-kube-api-access-plhtc\") pod \"crc-debug-f2656\" (UID: \"318c71d0-b625-4f30-9909-f157d416a123\") " pod="openshift-must-gather-8r4vv/crc-debug-f2656" Feb 18 09:49:57 crc kubenswrapper[4556]: I0218 09:49:57.042213 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/318c71d0-b625-4f30-9909-f157d416a123-host\") pod \"crc-debug-f2656\" (UID: \"318c71d0-b625-4f30-9909-f157d416a123\") " pod="openshift-must-gather-8r4vv/crc-debug-f2656" Feb 18 09:49:57 crc kubenswrapper[4556]: I0218 09:49:57.042336 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plhtc\" (UniqueName: \"kubernetes.io/projected/318c71d0-b625-4f30-9909-f157d416a123-kube-api-access-plhtc\") pod \"crc-debug-f2656\" (UID: \"318c71d0-b625-4f30-9909-f157d416a123\") " pod="openshift-must-gather-8r4vv/crc-debug-f2656" Feb 18 09:49:57 crc kubenswrapper[4556]: I0218 09:49:57.042375 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/318c71d0-b625-4f30-9909-f157d416a123-host\") pod \"crc-debug-f2656\" (UID: \"318c71d0-b625-4f30-9909-f157d416a123\") " pod="openshift-must-gather-8r4vv/crc-debug-f2656" Feb 18 09:49:57 crc kubenswrapper[4556]: I0218 09:49:57.058289 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plhtc\" (UniqueName: \"kubernetes.io/projected/318c71d0-b625-4f30-9909-f157d416a123-kube-api-access-plhtc\") pod \"crc-debug-f2656\" (UID: \"318c71d0-b625-4f30-9909-f157d416a123\") " pod="openshift-must-gather-8r4vv/crc-debug-f2656" Feb 18 09:49:57 crc kubenswrapper[4556]: I0218 09:49:57.158923 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8r4vv/crc-debug-f2656" Feb 18 09:49:57 crc kubenswrapper[4556]: I0218 09:49:57.291754 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f33854c8-5191-4d57-ad19-68972f69f793" path="/var/lib/kubelet/pods/f33854c8-5191-4d57-ad19-68972f69f793/volumes" Feb 18 09:49:57 crc kubenswrapper[4556]: I0218 09:49:57.576183 4556 generic.go:334] "Generic (PLEG): container finished" podID="318c71d0-b625-4f30-9909-f157d416a123" containerID="14e2967475a39bad299d1b85d9c8661e053160fdabe3d0a619d5622a00c2193c" exitCode=0 Feb 18 09:49:57 crc kubenswrapper[4556]: I0218 09:49:57.576363 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8r4vv/crc-debug-f2656" event={"ID":"318c71d0-b625-4f30-9909-f157d416a123","Type":"ContainerDied","Data":"14e2967475a39bad299d1b85d9c8661e053160fdabe3d0a619d5622a00c2193c"} Feb 18 09:49:57 crc kubenswrapper[4556]: I0218 09:49:57.576500 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8r4vv/crc-debug-f2656" event={"ID":"318c71d0-b625-4f30-9909-f157d416a123","Type":"ContainerStarted","Data":"40ee3ad65363681a3a781e76ca6fb77bc354b89d0b4b507a85483bb2c2512000"} Feb 18 09:49:58 crc kubenswrapper[4556]: I0218 09:49:58.052139 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-8r4vv/crc-debug-f2656"] Feb 18 09:49:58 crc kubenswrapper[4556]: I0218 09:49:58.062421 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-8r4vv/crc-debug-f2656"] Feb 18 09:49:58 crc kubenswrapper[4556]: I0218 09:49:58.684992 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8r4vv/crc-debug-f2656" Feb 18 09:49:58 crc kubenswrapper[4556]: I0218 09:49:58.781011 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/318c71d0-b625-4f30-9909-f157d416a123-host\") pod \"318c71d0-b625-4f30-9909-f157d416a123\" (UID: \"318c71d0-b625-4f30-9909-f157d416a123\") " Feb 18 09:49:58 crc kubenswrapper[4556]: I0218 09:49:58.781109 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-plhtc\" (UniqueName: \"kubernetes.io/projected/318c71d0-b625-4f30-9909-f157d416a123-kube-api-access-plhtc\") pod \"318c71d0-b625-4f30-9909-f157d416a123\" (UID: \"318c71d0-b625-4f30-9909-f157d416a123\") " Feb 18 09:49:58 crc kubenswrapper[4556]: I0218 09:49:58.781175 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/318c71d0-b625-4f30-9909-f157d416a123-host" (OuterVolumeSpecName: "host") pod "318c71d0-b625-4f30-9909-f157d416a123" (UID: "318c71d0-b625-4f30-9909-f157d416a123"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 09:49:58 crc kubenswrapper[4556]: I0218 09:49:58.781851 4556 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/318c71d0-b625-4f30-9909-f157d416a123-host\") on node \"crc\" DevicePath \"\"" Feb 18 09:49:58 crc kubenswrapper[4556]: I0218 09:49:58.790041 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/318c71d0-b625-4f30-9909-f157d416a123-kube-api-access-plhtc" (OuterVolumeSpecName: "kube-api-access-plhtc") pod "318c71d0-b625-4f30-9909-f157d416a123" (UID: "318c71d0-b625-4f30-9909-f157d416a123"). InnerVolumeSpecName "kube-api-access-plhtc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:49:58 crc kubenswrapper[4556]: I0218 09:49:58.884707 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-plhtc\" (UniqueName: \"kubernetes.io/projected/318c71d0-b625-4f30-9909-f157d416a123-kube-api-access-plhtc\") on node \"crc\" DevicePath \"\"" Feb 18 09:49:59 crc kubenswrapper[4556]: I0218 09:49:59.217416 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-8r4vv/crc-debug-r257l"] Feb 18 09:49:59 crc kubenswrapper[4556]: E0218 09:49:59.218333 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="318c71d0-b625-4f30-9909-f157d416a123" containerName="container-00" Feb 18 09:49:59 crc kubenswrapper[4556]: I0218 09:49:59.218353 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="318c71d0-b625-4f30-9909-f157d416a123" containerName="container-00" Feb 18 09:49:59 crc kubenswrapper[4556]: I0218 09:49:59.218560 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="318c71d0-b625-4f30-9909-f157d416a123" containerName="container-00" Feb 18 09:49:59 crc kubenswrapper[4556]: I0218 09:49:59.219466 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8r4vv/crc-debug-r257l" Feb 18 09:49:59 crc kubenswrapper[4556]: I0218 09:49:59.296171 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4j27\" (UniqueName: \"kubernetes.io/projected/15d1c03e-9b6a-4919-9feb-3bae60abf2b3-kube-api-access-x4j27\") pod \"crc-debug-r257l\" (UID: \"15d1c03e-9b6a-4919-9feb-3bae60abf2b3\") " pod="openshift-must-gather-8r4vv/crc-debug-r257l" Feb 18 09:49:59 crc kubenswrapper[4556]: I0218 09:49:59.296491 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/15d1c03e-9b6a-4919-9feb-3bae60abf2b3-host\") pod \"crc-debug-r257l\" (UID: \"15d1c03e-9b6a-4919-9feb-3bae60abf2b3\") " pod="openshift-must-gather-8r4vv/crc-debug-r257l" Feb 18 09:49:59 crc kubenswrapper[4556]: I0218 09:49:59.297322 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="318c71d0-b625-4f30-9909-f157d416a123" path="/var/lib/kubelet/pods/318c71d0-b625-4f30-9909-f157d416a123/volumes" Feb 18 09:49:59 crc kubenswrapper[4556]: I0218 09:49:59.398920 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/15d1c03e-9b6a-4919-9feb-3bae60abf2b3-host\") pod \"crc-debug-r257l\" (UID: \"15d1c03e-9b6a-4919-9feb-3bae60abf2b3\") " pod="openshift-must-gather-8r4vv/crc-debug-r257l" Feb 18 09:49:59 crc kubenswrapper[4556]: I0218 09:49:59.399056 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/15d1c03e-9b6a-4919-9feb-3bae60abf2b3-host\") pod \"crc-debug-r257l\" (UID: \"15d1c03e-9b6a-4919-9feb-3bae60abf2b3\") " pod="openshift-must-gather-8r4vv/crc-debug-r257l" Feb 18 09:49:59 crc kubenswrapper[4556]: I0218 09:49:59.399324 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4j27\" (UniqueName: \"kubernetes.io/projected/15d1c03e-9b6a-4919-9feb-3bae60abf2b3-kube-api-access-x4j27\") pod \"crc-debug-r257l\" (UID: \"15d1c03e-9b6a-4919-9feb-3bae60abf2b3\") " pod="openshift-must-gather-8r4vv/crc-debug-r257l" Feb 18 09:49:59 crc kubenswrapper[4556]: I0218 09:49:59.414746 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4j27\" (UniqueName: \"kubernetes.io/projected/15d1c03e-9b6a-4919-9feb-3bae60abf2b3-kube-api-access-x4j27\") pod \"crc-debug-r257l\" (UID: \"15d1c03e-9b6a-4919-9feb-3bae60abf2b3\") " pod="openshift-must-gather-8r4vv/crc-debug-r257l" Feb 18 09:49:59 crc kubenswrapper[4556]: I0218 09:49:59.537969 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8r4vv/crc-debug-r257l" Feb 18 09:49:59 crc kubenswrapper[4556]: W0218 09:49:59.585925 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod15d1c03e_9b6a_4919_9feb_3bae60abf2b3.slice/crio-af059ba53c76230f22482d6ea1c173cd83bb1052abde1f0e58b0f70d11b22f77 WatchSource:0}: Error finding container af059ba53c76230f22482d6ea1c173cd83bb1052abde1f0e58b0f70d11b22f77: Status 404 returned error can't find the container with id af059ba53c76230f22482d6ea1c173cd83bb1052abde1f0e58b0f70d11b22f77 Feb 18 09:49:59 crc kubenswrapper[4556]: I0218 09:49:59.599497 4556 scope.go:117] "RemoveContainer" containerID="14e2967475a39bad299d1b85d9c8661e053160fdabe3d0a619d5622a00c2193c" Feb 18 09:49:59 crc kubenswrapper[4556]: I0218 09:49:59.599529 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8r4vv/crc-debug-f2656" Feb 18 09:50:00 crc kubenswrapper[4556]: I0218 09:50:00.608687 4556 generic.go:334] "Generic (PLEG): container finished" podID="15d1c03e-9b6a-4919-9feb-3bae60abf2b3" containerID="1b2d16739bd1067c1691bfd0a18ba68d04e82a9108abcad0b88a4cc2b55c99c3" exitCode=0 Feb 18 09:50:00 crc kubenswrapper[4556]: I0218 09:50:00.608754 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8r4vv/crc-debug-r257l" event={"ID":"15d1c03e-9b6a-4919-9feb-3bae60abf2b3","Type":"ContainerDied","Data":"1b2d16739bd1067c1691bfd0a18ba68d04e82a9108abcad0b88a4cc2b55c99c3"} Feb 18 09:50:00 crc kubenswrapper[4556]: I0218 09:50:00.609102 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8r4vv/crc-debug-r257l" event={"ID":"15d1c03e-9b6a-4919-9feb-3bae60abf2b3","Type":"ContainerStarted","Data":"af059ba53c76230f22482d6ea1c173cd83bb1052abde1f0e58b0f70d11b22f77"} Feb 18 09:50:00 crc kubenswrapper[4556]: I0218 09:50:00.652040 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-8r4vv/crc-debug-r257l"] Feb 18 09:50:00 crc kubenswrapper[4556]: I0218 09:50:00.662602 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-8r4vv/crc-debug-r257l"] Feb 18 09:50:01 crc kubenswrapper[4556]: I0218 09:50:01.705086 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8r4vv/crc-debug-r257l" Feb 18 09:50:01 crc kubenswrapper[4556]: I0218 09:50:01.857470 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/15d1c03e-9b6a-4919-9feb-3bae60abf2b3-host\") pod \"15d1c03e-9b6a-4919-9feb-3bae60abf2b3\" (UID: \"15d1c03e-9b6a-4919-9feb-3bae60abf2b3\") " Feb 18 09:50:01 crc kubenswrapper[4556]: I0218 09:50:01.857940 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4j27\" (UniqueName: \"kubernetes.io/projected/15d1c03e-9b6a-4919-9feb-3bae60abf2b3-kube-api-access-x4j27\") pod \"15d1c03e-9b6a-4919-9feb-3bae60abf2b3\" (UID: \"15d1c03e-9b6a-4919-9feb-3bae60abf2b3\") " Feb 18 09:50:01 crc kubenswrapper[4556]: I0218 09:50:01.857604 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/15d1c03e-9b6a-4919-9feb-3bae60abf2b3-host" (OuterVolumeSpecName: "host") pod "15d1c03e-9b6a-4919-9feb-3bae60abf2b3" (UID: "15d1c03e-9b6a-4919-9feb-3bae60abf2b3"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 09:50:01 crc kubenswrapper[4556]: I0218 09:50:01.858499 4556 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/15d1c03e-9b6a-4919-9feb-3bae60abf2b3-host\") on node \"crc\" DevicePath \"\"" Feb 18 09:50:01 crc kubenswrapper[4556]: I0218 09:50:01.879322 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15d1c03e-9b6a-4919-9feb-3bae60abf2b3-kube-api-access-x4j27" (OuterVolumeSpecName: "kube-api-access-x4j27") pod "15d1c03e-9b6a-4919-9feb-3bae60abf2b3" (UID: "15d1c03e-9b6a-4919-9feb-3bae60abf2b3"). InnerVolumeSpecName "kube-api-access-x4j27". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:50:01 crc kubenswrapper[4556]: I0218 09:50:01.961265 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4j27\" (UniqueName: \"kubernetes.io/projected/15d1c03e-9b6a-4919-9feb-3bae60abf2b3-kube-api-access-x4j27\") on node \"crc\" DevicePath \"\"" Feb 18 09:50:02 crc kubenswrapper[4556]: I0218 09:50:02.633432 4556 scope.go:117] "RemoveContainer" containerID="1b2d16739bd1067c1691bfd0a18ba68d04e82a9108abcad0b88a4cc2b55c99c3" Feb 18 09:50:02 crc kubenswrapper[4556]: I0218 09:50:02.633484 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8r4vv/crc-debug-r257l" Feb 18 09:50:03 crc kubenswrapper[4556]: I0218 09:50:03.303146 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15d1c03e-9b6a-4919-9feb-3bae60abf2b3" path="/var/lib/kubelet/pods/15d1c03e-9b6a-4919-9feb-3bae60abf2b3/volumes" Feb 18 09:50:16 crc kubenswrapper[4556]: I0218 09:50:16.267898 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-845cb5d46d-z5vhw_ff67ebd8-094e-4c5b-b164-d6b37fc169de/barbican-api/0.log" Feb 18 09:50:16 crc kubenswrapper[4556]: I0218 09:50:16.441965 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-845cb5d46d-z5vhw_ff67ebd8-094e-4c5b-b164-d6b37fc169de/barbican-api-log/0.log" Feb 18 09:50:16 crc kubenswrapper[4556]: I0218 09:50:16.465543 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5fc745f7dd-68srz_8be9c112-7903-4ec0-a503-fb16c676caf4/barbican-keystone-listener/0.log" Feb 18 09:50:16 crc kubenswrapper[4556]: I0218 09:50:16.502623 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5fc745f7dd-68srz_8be9c112-7903-4ec0-a503-fb16c676caf4/barbican-keystone-listener-log/0.log" Feb 18 09:50:16 crc kubenswrapper[4556]: I0218 09:50:16.617258 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-b64d8f579-7sj9h_d6ef48e2-d690-4060-b013-5ac2be288161/barbican-worker/0.log" Feb 18 09:50:16 crc kubenswrapper[4556]: I0218 09:50:16.653414 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-b64d8f579-7sj9h_d6ef48e2-d690-4060-b013-5ac2be288161/barbican-worker-log/0.log" Feb 18 09:50:16 crc kubenswrapper[4556]: I0218 09:50:16.791952 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-pqnqh_91161557-e383-4bba-81d8-d3b0ba3b6840/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Feb 18 09:50:16 crc kubenswrapper[4556]: I0218 09:50:16.851221 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_d105f935-cc14-4293-8101-17adaef31666/ceilometer-central-agent/0.log" Feb 18 09:50:16 crc kubenswrapper[4556]: I0218 09:50:16.917363 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_d105f935-cc14-4293-8101-17adaef31666/ceilometer-notification-agent/0.log" Feb 18 09:50:16 crc kubenswrapper[4556]: I0218 09:50:16.974191 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_d105f935-cc14-4293-8101-17adaef31666/proxy-httpd/0.log" Feb 18 09:50:17 crc kubenswrapper[4556]: I0218 09:50:17.007434 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_d105f935-cc14-4293-8101-17adaef31666/sg-core/0.log" Feb 18 09:50:17 crc kubenswrapper[4556]: I0218 09:50:17.124379 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d/cinder-api/0.log" Feb 18 09:50:17 crc kubenswrapper[4556]: I0218 09:50:17.129000 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d/cinder-api-log/0.log" Feb 18 09:50:17 crc kubenswrapper[4556]: I0218 09:50:17.249025 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_ca98a45d-9b95-4e00-8179-feb5a6c1ddb0/cinder-scheduler/0.log" Feb 18 09:50:17 crc kubenswrapper[4556]: I0218 09:50:17.307534 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_ca98a45d-9b95-4e00-8179-feb5a6c1ddb0/probe/0.log" Feb 18 09:50:17 crc kubenswrapper[4556]: I0218 09:50:17.435526 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-snzk4_4fdd7d3d-c538-4cb6-9d45-b94723b744b1/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Feb 18 09:50:17 crc kubenswrapper[4556]: I0218 09:50:17.495618 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-nfwmp_c71f2826-1f14-462a-8096-cace6bd934d3/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 18 09:50:17 crc kubenswrapper[4556]: I0218 09:50:17.600775 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-56455489f5-9pk6d_d6426d5e-fb37-42e3-8392-3db7d31e81e7/init/0.log" Feb 18 09:50:17 crc kubenswrapper[4556]: I0218 09:50:17.752817 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-56455489f5-9pk6d_d6426d5e-fb37-42e3-8392-3db7d31e81e7/init/0.log" Feb 18 09:50:17 crc kubenswrapper[4556]: I0218 09:50:17.806063 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-56455489f5-9pk6d_d6426d5e-fb37-42e3-8392-3db7d31e81e7/dnsmasq-dns/0.log" Feb 18 09:50:17 crc kubenswrapper[4556]: I0218 09:50:17.821554 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-qdtn4_a33e8704-dcbf-43ff-b665-cff323679c76/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Feb 18 09:50:17 crc kubenswrapper[4556]: I0218 09:50:17.976808 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_23a483a0-fda5-4bf6-bfea-06a84d8d6533/glance-log/0.log" Feb 18 09:50:17 crc kubenswrapper[4556]: I0218 09:50:17.980176 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_23a483a0-fda5-4bf6-bfea-06a84d8d6533/glance-httpd/0.log" Feb 18 09:50:18 crc kubenswrapper[4556]: I0218 09:50:18.114028 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_f50b4c03-2bce-417a-b01c-aaec09a5c8c3/glance-httpd/0.log" Feb 18 09:50:18 crc kubenswrapper[4556]: I0218 09:50:18.148046 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_f50b4c03-2bce-417a-b01c-aaec09a5c8c3/glance-log/0.log" Feb 18 09:50:18 crc kubenswrapper[4556]: I0218 09:50:18.277924 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-595566bb8b-6b5h6_30fffa0b-72a5-4bbe-a1fb-fa8b26d1decf/horizon/0.log" Feb 18 09:50:18 crc kubenswrapper[4556]: I0218 09:50:18.446839 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-t9trr_d249a625-7b1c-4a2f-a92d-c18d7fbb6142/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Feb 18 09:50:18 crc kubenswrapper[4556]: I0218 09:50:18.493521 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-595566bb8b-6b5h6_30fffa0b-72a5-4bbe-a1fb-fa8b26d1decf/horizon-log/0.log" Feb 18 09:50:18 crc kubenswrapper[4556]: I0218 09:50:18.616717 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-zmv42_d1d95573-7bea-4ee6-b0c1-13ec18b10244/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 18 09:50:18 crc kubenswrapper[4556]: I0218 09:50:18.825888 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-66f6978747-npdn5_6b87f75f-f2e8-4a26-b59c-1d0bbe9c777d/keystone-api/0.log" Feb 18 09:50:18 crc kubenswrapper[4556]: I0218 09:50:18.843544 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_e54e3af3-e2c2-4e5b-885d-0071d05acbb4/kube-state-metrics/0.log" Feb 18 09:50:18 crc kubenswrapper[4556]: I0218 09:50:18.913692 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-hhn8t_2cff0286-97e0-422e-b55c-4f28711bde2d/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Feb 18 09:50:19 crc kubenswrapper[4556]: I0218 09:50:19.222792 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-54664bdbc-cxnvs_57873d95-dd1a-4b5c-99b6-459774c90acc/neutron-api/0.log" Feb 18 09:50:19 crc kubenswrapper[4556]: I0218 09:50:19.339172 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-54664bdbc-cxnvs_57873d95-dd1a-4b5c-99b6-459774c90acc/neutron-httpd/0.log" Feb 18 09:50:19 crc kubenswrapper[4556]: I0218 09:50:19.568822 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-wcldl_3343de49-153e-43bb-9d22-703a5f3a0b0b/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Feb 18 09:50:20 crc kubenswrapper[4556]: I0218 09:50:20.046324 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_4e997362-15c5-4052-bf0b-8491c057f317/nova-api-log/0.log" Feb 18 09:50:20 crc kubenswrapper[4556]: I0218 09:50:20.085756 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_64bcfb93-2db3-4350-97ee-671806ff7d00/nova-cell0-conductor-conductor/0.log" Feb 18 09:50:20 crc kubenswrapper[4556]: I0218 09:50:20.089893 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_4e997362-15c5-4052-bf0b-8491c057f317/nova-api-api/0.log" Feb 18 09:50:20 crc kubenswrapper[4556]: I0218 09:50:20.301667 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_f0149238-212d-489d-8c94-7b0a8d86b1f5/nova-cell1-conductor-conductor/0.log" Feb 18 09:50:20 crc kubenswrapper[4556]: I0218 09:50:20.355087 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_c1c20375-9fcb-4886-a91c-3dd36651532a/nova-cell1-novncproxy-novncproxy/0.log" Feb 18 09:50:20 crc kubenswrapper[4556]: I0218 09:50:20.549145 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-snrkh_a1ab2e8c-4abd-4421-ac68-20821b8ba938/nova-edpm-deployment-openstack-edpm-ipam/0.log" Feb 18 09:50:20 crc kubenswrapper[4556]: I0218 09:50:20.619120 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_4424b924-1dd6-4323-863d-53bf1a0f5fc9/nova-metadata-log/0.log" Feb 18 09:50:20 crc kubenswrapper[4556]: I0218 09:50:20.941853 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_db4d9035-57ee-4c5f-9677-cc2174644152/nova-scheduler-scheduler/0.log" Feb 18 09:50:21 crc kubenswrapper[4556]: I0218 09:50:21.156036 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_11b77df6-67d9-4ddb-944b-2ad8b0fada78/mysql-bootstrap/0.log" Feb 18 09:50:21 crc kubenswrapper[4556]: I0218 09:50:21.299760 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_11b77df6-67d9-4ddb-944b-2ad8b0fada78/mysql-bootstrap/0.log" Feb 18 09:50:21 crc kubenswrapper[4556]: I0218 09:50:21.317493 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_11b77df6-67d9-4ddb-944b-2ad8b0fada78/galera/0.log" Feb 18 09:50:21 crc kubenswrapper[4556]: I0218 09:50:21.484431 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_2c504c68-45e0-49dc-90f3-90b91e33551c/mysql-bootstrap/0.log" Feb 18 09:50:21 crc kubenswrapper[4556]: I0218 09:50:21.493537 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_4424b924-1dd6-4323-863d-53bf1a0f5fc9/nova-metadata-metadata/0.log" Feb 18 09:50:21 crc kubenswrapper[4556]: I0218 09:50:21.737479 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_2c504c68-45e0-49dc-90f3-90b91e33551c/mysql-bootstrap/0.log" Feb 18 09:50:21 crc kubenswrapper[4556]: I0218 09:50:21.759305 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_2c504c68-45e0-49dc-90f3-90b91e33551c/galera/0.log" Feb 18 09:50:21 crc kubenswrapper[4556]: I0218 09:50:21.764462 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_97268f30-e67a-441f-93cd-0dc9771dbd42/openstackclient/0.log" Feb 18 09:50:21 crc kubenswrapper[4556]: I0218 09:50:21.980139 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-6kkrw_4906afef-c7e6-4597-a4a5-e9d758917e11/ovn-controller/0.log" Feb 18 09:50:22 crc kubenswrapper[4556]: I0218 09:50:22.035864 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-j7rpf_66575dd2-efe2-4770-a9cd-7afb99e8566c/openstack-network-exporter/0.log" Feb 18 09:50:22 crc kubenswrapper[4556]: I0218 09:50:22.223171 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-srnvn_40c51577-de87-4bd8-be03-41221bc2f415/ovsdb-server-init/0.log" Feb 18 09:50:22 crc kubenswrapper[4556]: I0218 09:50:22.396748 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-srnvn_40c51577-de87-4bd8-be03-41221bc2f415/ovs-vswitchd/0.log" Feb 18 09:50:22 crc kubenswrapper[4556]: I0218 09:50:22.400954 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-srnvn_40c51577-de87-4bd8-be03-41221bc2f415/ovsdb-server-init/0.log" Feb 18 09:50:22 crc kubenswrapper[4556]: I0218 09:50:22.471964 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-srnvn_40c51577-de87-4bd8-be03-41221bc2f415/ovsdb-server/0.log" Feb 18 09:50:22 crc kubenswrapper[4556]: I0218 09:50:22.634659 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-vgw62_7c63f939-1090-4fae-b45b-9d55f09f489d/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Feb 18 09:50:22 crc kubenswrapper[4556]: I0218 09:50:22.708120 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_ef64adb0-c929-4a3f-8aa6-d490b45ba5e1/openstack-network-exporter/0.log" Feb 18 09:50:22 crc kubenswrapper[4556]: I0218 09:50:22.767114 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_ef64adb0-c929-4a3f-8aa6-d490b45ba5e1/ovn-northd/0.log" Feb 18 09:50:22 crc kubenswrapper[4556]: I0218 09:50:22.813775 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_10b5d4cb-2beb-405d-83fc-30eb5fd4aaad/openstack-network-exporter/0.log" Feb 18 09:50:22 crc kubenswrapper[4556]: I0218 09:50:22.880505 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_10b5d4cb-2beb-405d-83fc-30eb5fd4aaad/ovsdbserver-nb/0.log" Feb 18 09:50:23 crc kubenswrapper[4556]: I0218 09:50:23.039043 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_d7840643-68ec-4a2a-b6af-8a9730729077/openstack-network-exporter/0.log" Feb 18 09:50:23 crc kubenswrapper[4556]: I0218 09:50:23.091689 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_d7840643-68ec-4a2a-b6af-8a9730729077/ovsdbserver-sb/0.log" Feb 18 09:50:23 crc kubenswrapper[4556]: I0218 09:50:23.324264 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-9855c4778-9n9f2_ac2fe5b2-09b6-440f-83db-8555ec304a27/placement-api/0.log" Feb 18 09:50:23 crc kubenswrapper[4556]: I0218 09:50:23.336427 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_8b1c0c56-94a0-4ab7-ae4c-5f4035e37359/setup-container/0.log" Feb 18 09:50:23 crc kubenswrapper[4556]: I0218 09:50:23.341738 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-9855c4778-9n9f2_ac2fe5b2-09b6-440f-83db-8555ec304a27/placement-log/0.log" Feb 18 09:50:23 crc kubenswrapper[4556]: I0218 09:50:23.525118 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_8b1c0c56-94a0-4ab7-ae4c-5f4035e37359/rabbitmq/0.log" Feb 18 09:50:23 crc kubenswrapper[4556]: I0218 09:50:23.542642 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_8b1c0c56-94a0-4ab7-ae4c-5f4035e37359/setup-container/0.log" Feb 18 09:50:23 crc kubenswrapper[4556]: I0218 09:50:23.613230 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_0f0ef5d2-a692-435f-a79d-a1af3a294e73/setup-container/0.log" Feb 18 09:50:23 crc kubenswrapper[4556]: I0218 09:50:23.743094 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_0f0ef5d2-a692-435f-a79d-a1af3a294e73/setup-container/0.log" Feb 18 09:50:23 crc kubenswrapper[4556]: I0218 09:50:23.802460 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-jmgv7_d0b159c4-dbac-4ddb-af28-523bb86639e5/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 18 09:50:23 crc kubenswrapper[4556]: I0218 09:50:23.803053 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_0f0ef5d2-a692-435f-a79d-a1af3a294e73/rabbitmq/0.log" Feb 18 09:50:24 crc kubenswrapper[4556]: I0218 09:50:24.016348 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-hjswl_3ebf502b-e4a2-4d25-8551-f1deed31b5c9/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Feb 18 09:50:24 crc kubenswrapper[4556]: I0218 09:50:24.072642 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-sqqbc_c2b34eeb-f211-4454-a74d-207de2fa9f13/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Feb 18 09:50:24 crc kubenswrapper[4556]: I0218 09:50:24.265346 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-2j6nf_1d8fa453-3277-4c76-9098-0e7838fd8d44/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 18 09:50:24 crc kubenswrapper[4556]: I0218 09:50:24.297074 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-92c7r_11adbf49-97c4-4878-96c0-4fa453f7a819/ssh-known-hosts-edpm-deployment/0.log" Feb 18 09:50:24 crc kubenswrapper[4556]: I0218 09:50:24.523948 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-864fbf8dcf-chrjx_d678be18-d610-4ea8-b248-47843cf74ea3/proxy-server/0.log" Feb 18 09:50:24 crc kubenswrapper[4556]: I0218 09:50:24.613399 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-864fbf8dcf-chrjx_d678be18-d610-4ea8-b248-47843cf74ea3/proxy-httpd/0.log" Feb 18 09:50:24 crc kubenswrapper[4556]: I0218 09:50:24.653794 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-9z5cc_03ccffdf-f11c-4788-9e22-0b0661665c5d/swift-ring-rebalance/0.log" Feb 18 09:50:24 crc kubenswrapper[4556]: I0218 09:50:24.779322 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_da163339-653b-4057-9c2f-332eb6957f40/account-reaper/0.log" Feb 18 09:50:24 crc kubenswrapper[4556]: I0218 09:50:24.811086 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_da163339-653b-4057-9c2f-332eb6957f40/account-auditor/0.log" Feb 18 09:50:24 crc kubenswrapper[4556]: I0218 09:50:24.872145 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_da163339-653b-4057-9c2f-332eb6957f40/account-replicator/0.log" Feb 18 09:50:24 crc kubenswrapper[4556]: I0218 09:50:24.929459 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_da163339-653b-4057-9c2f-332eb6957f40/account-server/0.log" Feb 18 09:50:25 crc kubenswrapper[4556]: I0218 09:50:25.010611 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_da163339-653b-4057-9c2f-332eb6957f40/container-auditor/0.log" Feb 18 09:50:25 crc kubenswrapper[4556]: I0218 09:50:25.011005 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_da163339-653b-4057-9c2f-332eb6957f40/container-replicator/0.log" Feb 18 09:50:25 crc kubenswrapper[4556]: I0218 09:50:25.083756 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_da163339-653b-4057-9c2f-332eb6957f40/container-server/0.log" Feb 18 09:50:25 crc kubenswrapper[4556]: I0218 09:50:25.119623 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_da163339-653b-4057-9c2f-332eb6957f40/container-updater/0.log" Feb 18 09:50:25 crc kubenswrapper[4556]: I0218 09:50:25.228851 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_da163339-653b-4057-9c2f-332eb6957f40/object-expirer/0.log" Feb 18 09:50:25 crc kubenswrapper[4556]: I0218 09:50:25.241258 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_da163339-653b-4057-9c2f-332eb6957f40/object-auditor/0.log" Feb 18 09:50:25 crc kubenswrapper[4556]: I0218 09:50:25.319655 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_da163339-653b-4057-9c2f-332eb6957f40/object-server/0.log" Feb 18 09:50:25 crc kubenswrapper[4556]: I0218 09:50:25.330761 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_da163339-653b-4057-9c2f-332eb6957f40/object-replicator/0.log" Feb 18 09:50:25 crc kubenswrapper[4556]: I0218 09:50:25.426461 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_da163339-653b-4057-9c2f-332eb6957f40/object-updater/0.log" Feb 18 09:50:25 crc kubenswrapper[4556]: I0218 09:50:25.480256 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_da163339-653b-4057-9c2f-332eb6957f40/swift-recon-cron/0.log" Feb 18 09:50:25 crc kubenswrapper[4556]: I0218 09:50:25.511388 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_da163339-653b-4057-9c2f-332eb6957f40/rsync/0.log" Feb 18 09:50:25 crc kubenswrapper[4556]: I0218 09:50:25.697940 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-l92vs_8ea53a2a-db68-43fb-98c4-26b9ca9f816d/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Feb 18 09:50:25 crc kubenswrapper[4556]: I0218 09:50:25.755068 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59/tempest-tests-tempest-tests-runner/0.log" Feb 18 09:50:25 crc kubenswrapper[4556]: I0218 09:50:25.886238 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_396883e0-c2e8-4698-97d2-70bbd5fe7b7f/test-operator-logs-container/0.log" Feb 18 09:50:25 crc kubenswrapper[4556]: I0218 09:50:25.988924 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-6jjjt_a5a47273-47da-41d2-a98f-dc1fba7e1102/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Feb 18 09:50:33 crc kubenswrapper[4556]: I0218 09:50:33.650534 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_68274946-b189-450c-a154-27059f411af3/memcached/0.log" Feb 18 09:50:47 crc kubenswrapper[4556]: I0218 09:50:47.769641 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967fk785_b828da70-b43b-421e-a2b2-d1a2bd562ae4/util/0.log" Feb 18 09:50:47 crc kubenswrapper[4556]: I0218 09:50:47.945041 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967fk785_b828da70-b43b-421e-a2b2-d1a2bd562ae4/pull/0.log" Feb 18 09:50:47 crc kubenswrapper[4556]: I0218 09:50:47.962903 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967fk785_b828da70-b43b-421e-a2b2-d1a2bd562ae4/util/0.log" Feb 18 09:50:47 crc kubenswrapper[4556]: I0218 09:50:47.975951 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967fk785_b828da70-b43b-421e-a2b2-d1a2bd562ae4/pull/0.log" Feb 18 09:50:48 crc kubenswrapper[4556]: I0218 09:50:48.112185 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967fk785_b828da70-b43b-421e-a2b2-d1a2bd562ae4/util/0.log" Feb 18 09:50:48 crc kubenswrapper[4556]: I0218 09:50:48.132826 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967fk785_b828da70-b43b-421e-a2b2-d1a2bd562ae4/extract/0.log" Feb 18 09:50:48 crc kubenswrapper[4556]: I0218 09:50:48.140318 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967fk785_b828da70-b43b-421e-a2b2-d1a2bd562ae4/pull/0.log" Feb 18 09:50:48 crc kubenswrapper[4556]: I0218 09:50:48.538643 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-6d8bf5c495-5pv48_fff15cfd-9701-4efa-81a6-9a482e09ca0f/manager/0.log" Feb 18 09:50:48 crc kubenswrapper[4556]: I0218 09:50:48.857671 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987464f4-mfhjc_f194b6f2-c143-4a57-8d3e-378a08147713/manager/0.log" Feb 18 09:50:48 crc kubenswrapper[4556]: I0218 09:50:48.989731 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-69f49c598c-6sck2_a980ea11-bf5f-4967-b22b-3e63454bb9ae/manager/0.log" Feb 18 09:50:49 crc kubenswrapper[4556]: I0218 09:50:49.189098 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5b9b8895d5-g6hvq_6a219296-13f5-4c55-99ec-35e34d43a341/manager/0.log" Feb 18 09:50:49 crc kubenswrapper[4556]: I0218 09:50:49.714561 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-554564d7fc-lktrc_4fdde554-179d-4128-a260-38c3de6e9d64/manager/0.log" Feb 18 09:50:49 crc kubenswrapper[4556]: I0218 09:50:49.770631 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-5d946d989d-n4x7d_463e8b44-a344-401e-bd53-22738560ca32/manager/0.log" Feb 18 09:50:49 crc kubenswrapper[4556]: I0218 09:50:49.814082 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-79d975b745-gp4rt_853bc90c-e14b-46ea-912d-5aaefa5f908a/manager/0.log" Feb 18 09:50:50 crc kubenswrapper[4556]: I0218 09:50:50.039347 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-b4d948c87-9rkqc_cd593415-1b63-4965-9865-21a1ae2e4742/manager/0.log" Feb 18 09:50:50 crc kubenswrapper[4556]: I0218 09:50:50.110660 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-54f6768c69-r5kw5_6e7470ce-ab97-4534-84e8-e951071eb6ee/manager/0.log" Feb 18 09:50:50 crc kubenswrapper[4556]: I0218 09:50:50.276416 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6994f66f48-t2ng7_fdd8c829-9082-40ae-8b65-46a8bd293c4d/manager/0.log" Feb 18 09:50:50 crc kubenswrapper[4556]: I0218 09:50:50.430871 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-64ddbf8bb-wdp2t_9d670c80-f1d0-4b08-b83b-086977d49380/manager/0.log" Feb 18 09:50:50 crc kubenswrapper[4556]: I0218 09:50:50.588542 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-567668f5cf-t9gtc_d672d582-25f7-4654-bc52-5f07631f5e9e/manager/0.log" Feb 18 09:50:50 crc kubenswrapper[4556]: I0218 09:50:50.839374 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-fb5fcc5b8-9mlxh_48ca8d44-b660-43ee-8a9e-6489da3cee3d/manager/0.log" Feb 18 09:50:51 crc kubenswrapper[4556]: I0218 09:50:51.260503 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-6679bf9b57-r8zc2_acd47be1-151c-4fd9-9cc6-f49e1cc3f403/operator/0.log" Feb 18 09:50:51 crc kubenswrapper[4556]: I0218 09:50:51.483505 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-tg2z4_85052b3f-274e-437c-af72-fdd608552f7c/registry-server/0.log" Feb 18 09:50:51 crc kubenswrapper[4556]: I0218 09:50:51.720842 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-d44cf6b75-qr5np_2c577857-1c7f-4d16-9c13-d3d83eddb3de/manager/0.log" Feb 18 09:50:51 crc kubenswrapper[4556]: I0218 09:50:51.917351 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-8497b45c89-5vgnl_f2bd3ad9-919a-40ec-b111-f9655a6650a6/manager/0.log" Feb 18 09:50:52 crc kubenswrapper[4556]: I0218 09:50:52.163052 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-vt6jv_28452a75-3b84-4add-9952-c5f0193e0954/operator/0.log" Feb 18 09:50:52 crc kubenswrapper[4556]: I0218 09:50:52.202869 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-69f8888797-rpk9w_96be73f9-04a8-4ec6-9b75-ae0dd4ae84c8/manager/0.log" Feb 18 09:50:52 crc kubenswrapper[4556]: I0218 09:50:52.403857 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-68f46476f-jz577_26bd53bd-69fd-4b14-afd5-5ffb7e557346/manager/0.log" Feb 18 09:50:52 crc kubenswrapper[4556]: I0218 09:50:52.471996 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-7f45b4ff68-4f5nl_acad90fd-fc19-4b74-bb00-192d0e6061a6/manager/0.log" Feb 18 09:50:52 crc kubenswrapper[4556]: I0218 09:50:52.673630 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-7866795846-9bsld_bc370f09-39b1-4e31-bbc6-7756c63f0c30/manager/0.log" Feb 18 09:50:52 crc kubenswrapper[4556]: I0218 09:50:52.776659 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-5db88f68c-98pqn_8e02c422-7a17-4d4f-9ac1-7fe86c5dd472/manager/0.log" Feb 18 09:50:53 crc kubenswrapper[4556]: I0218 09:50:53.020600 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-69ff7bc449-jjnjd_4fb1fa58-065b-4412-99e1-e9d77cdf4b41/manager/0.log" Feb 18 09:50:54 crc kubenswrapper[4556]: I0218 09:50:54.838269 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-868647ff47-x74c8_9cf84ffa-7a16-4e6a-aa7a-c90d6c1635fe/manager/0.log" Feb 18 09:51:09 crc kubenswrapper[4556]: I0218 09:51:09.449300 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bk6mp"] Feb 18 09:51:09 crc kubenswrapper[4556]: E0218 09:51:09.451190 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15d1c03e-9b6a-4919-9feb-3bae60abf2b3" containerName="container-00" Feb 18 09:51:09 crc kubenswrapper[4556]: I0218 09:51:09.451438 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="15d1c03e-9b6a-4919-9feb-3bae60abf2b3" containerName="container-00" Feb 18 09:51:09 crc kubenswrapper[4556]: I0218 09:51:09.451755 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="15d1c03e-9b6a-4919-9feb-3bae60abf2b3" containerName="container-00" Feb 18 09:51:09 crc kubenswrapper[4556]: I0218 09:51:09.453129 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bk6mp" Feb 18 09:51:09 crc kubenswrapper[4556]: I0218 09:51:09.474448 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bk6mp"] Feb 18 09:51:09 crc kubenswrapper[4556]: I0218 09:51:09.507219 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckrm8\" (UniqueName: \"kubernetes.io/projected/7baa1e0e-e912-4e07-8e0b-2fa63fc6d2df-kube-api-access-ckrm8\") pod \"redhat-marketplace-bk6mp\" (UID: \"7baa1e0e-e912-4e07-8e0b-2fa63fc6d2df\") " pod="openshift-marketplace/redhat-marketplace-bk6mp" Feb 18 09:51:09 crc kubenswrapper[4556]: I0218 09:51:09.507329 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7baa1e0e-e912-4e07-8e0b-2fa63fc6d2df-catalog-content\") pod \"redhat-marketplace-bk6mp\" (UID: \"7baa1e0e-e912-4e07-8e0b-2fa63fc6d2df\") " pod="openshift-marketplace/redhat-marketplace-bk6mp" Feb 18 09:51:09 crc kubenswrapper[4556]: I0218 09:51:09.507694 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7baa1e0e-e912-4e07-8e0b-2fa63fc6d2df-utilities\") pod \"redhat-marketplace-bk6mp\" (UID: \"7baa1e0e-e912-4e07-8e0b-2fa63fc6d2df\") " pod="openshift-marketplace/redhat-marketplace-bk6mp" Feb 18 09:51:09 crc kubenswrapper[4556]: I0218 09:51:09.609993 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7baa1e0e-e912-4e07-8e0b-2fa63fc6d2df-catalog-content\") pod \"redhat-marketplace-bk6mp\" (UID: \"7baa1e0e-e912-4e07-8e0b-2fa63fc6d2df\") " pod="openshift-marketplace/redhat-marketplace-bk6mp" Feb 18 09:51:09 crc kubenswrapper[4556]: I0218 09:51:09.610401 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7baa1e0e-e912-4e07-8e0b-2fa63fc6d2df-utilities\") pod \"redhat-marketplace-bk6mp\" (UID: \"7baa1e0e-e912-4e07-8e0b-2fa63fc6d2df\") " pod="openshift-marketplace/redhat-marketplace-bk6mp" Feb 18 09:51:09 crc kubenswrapper[4556]: I0218 09:51:09.610684 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckrm8\" (UniqueName: \"kubernetes.io/projected/7baa1e0e-e912-4e07-8e0b-2fa63fc6d2df-kube-api-access-ckrm8\") pod \"redhat-marketplace-bk6mp\" (UID: \"7baa1e0e-e912-4e07-8e0b-2fa63fc6d2df\") " pod="openshift-marketplace/redhat-marketplace-bk6mp" Feb 18 09:51:09 crc kubenswrapper[4556]: I0218 09:51:09.610770 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7baa1e0e-e912-4e07-8e0b-2fa63fc6d2df-catalog-content\") pod \"redhat-marketplace-bk6mp\" (UID: \"7baa1e0e-e912-4e07-8e0b-2fa63fc6d2df\") " pod="openshift-marketplace/redhat-marketplace-bk6mp" Feb 18 09:51:09 crc kubenswrapper[4556]: I0218 09:51:09.610868 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7baa1e0e-e912-4e07-8e0b-2fa63fc6d2df-utilities\") pod \"redhat-marketplace-bk6mp\" (UID: \"7baa1e0e-e912-4e07-8e0b-2fa63fc6d2df\") " pod="openshift-marketplace/redhat-marketplace-bk6mp" Feb 18 09:51:09 crc kubenswrapper[4556]: I0218 09:51:09.645128 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckrm8\" (UniqueName: \"kubernetes.io/projected/7baa1e0e-e912-4e07-8e0b-2fa63fc6d2df-kube-api-access-ckrm8\") pod \"redhat-marketplace-bk6mp\" (UID: \"7baa1e0e-e912-4e07-8e0b-2fa63fc6d2df\") " pod="openshift-marketplace/redhat-marketplace-bk6mp" Feb 18 09:51:09 crc kubenswrapper[4556]: I0218 09:51:09.647837 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-j5fdh_b3ca6b25-cb73-47a9-867c-8d1f6f628077/control-plane-machine-set-operator/0.log" Feb 18 09:51:09 crc kubenswrapper[4556]: I0218 09:51:09.782510 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bk6mp" Feb 18 09:51:09 crc kubenswrapper[4556]: I0218 09:51:09.860378 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-zhtqs_b5f35d95-f6b7-4068-8036-dffcbf955272/machine-api-operator/0.log" Feb 18 09:51:09 crc kubenswrapper[4556]: I0218 09:51:09.921373 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-zhtqs_b5f35d95-f6b7-4068-8036-dffcbf955272/kube-rbac-proxy/0.log" Feb 18 09:51:10 crc kubenswrapper[4556]: I0218 09:51:10.247021 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bk6mp"] Feb 18 09:51:11 crc kubenswrapper[4556]: I0218 09:51:11.241691 4556 generic.go:334] "Generic (PLEG): container finished" podID="7baa1e0e-e912-4e07-8e0b-2fa63fc6d2df" containerID="eb8d9d6b6b6049cc45f005bf587f42483d2e577b529949a59dcec476cc617bed" exitCode=0 Feb 18 09:51:11 crc kubenswrapper[4556]: I0218 09:51:11.241786 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bk6mp" event={"ID":"7baa1e0e-e912-4e07-8e0b-2fa63fc6d2df","Type":"ContainerDied","Data":"eb8d9d6b6b6049cc45f005bf587f42483d2e577b529949a59dcec476cc617bed"} Feb 18 09:51:11 crc kubenswrapper[4556]: I0218 09:51:11.243013 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bk6mp" event={"ID":"7baa1e0e-e912-4e07-8e0b-2fa63fc6d2df","Type":"ContainerStarted","Data":"1eb4ff368db0c3e837a96074db9a2948bea0478bcaadf45f705a843ee8e9339c"} Feb 18 09:51:12 crc kubenswrapper[4556]: I0218 09:51:12.253473 4556 generic.go:334] "Generic (PLEG): container finished" podID="7baa1e0e-e912-4e07-8e0b-2fa63fc6d2df" containerID="2085a1c77064ac6829c3ddb4bc3d5e771a544a9bfea9e9109761848030518f4b" exitCode=0 Feb 18 09:51:12 crc kubenswrapper[4556]: I0218 09:51:12.253559 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bk6mp" event={"ID":"7baa1e0e-e912-4e07-8e0b-2fa63fc6d2df","Type":"ContainerDied","Data":"2085a1c77064ac6829c3ddb4bc3d5e771a544a9bfea9e9109761848030518f4b"} Feb 18 09:51:13 crc kubenswrapper[4556]: I0218 09:51:13.270027 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bk6mp" event={"ID":"7baa1e0e-e912-4e07-8e0b-2fa63fc6d2df","Type":"ContainerStarted","Data":"a95c751826ee352908fbe98d82f9be57e14e98302282745acf5fea6ebc5463a6"} Feb 18 09:51:13 crc kubenswrapper[4556]: I0218 09:51:13.292047 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bk6mp" podStartSLOduration=2.605566331 podStartE2EDuration="4.291990407s" podCreationTimestamp="2026-02-18 09:51:09 +0000 UTC" firstStartedPulling="2026-02-18 09:51:11.246084593 +0000 UTC m=+2828.263045574" lastFinishedPulling="2026-02-18 09:51:12.93250867 +0000 UTC m=+2829.949469650" observedRunningTime="2026-02-18 09:51:13.288854159 +0000 UTC m=+2830.305815139" watchObservedRunningTime="2026-02-18 09:51:13.291990407 +0000 UTC m=+2830.308951386" Feb 18 09:51:19 crc kubenswrapper[4556]: I0218 09:51:19.782638 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bk6mp" Feb 18 09:51:19 crc kubenswrapper[4556]: I0218 09:51:19.783876 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bk6mp" Feb 18 09:51:19 crc kubenswrapper[4556]: I0218 09:51:19.826637 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bk6mp" Feb 18 09:51:20 crc kubenswrapper[4556]: I0218 09:51:20.399412 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bk6mp" Feb 18 09:51:20 crc kubenswrapper[4556]: I0218 09:51:20.446663 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bk6mp"] Feb 18 09:51:21 crc kubenswrapper[4556]: I0218 09:51:21.310225 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-bw685_257b9c79-fc7b-4bc5-88eb-d767e4762f77/cert-manager-controller/0.log" Feb 18 09:51:21 crc kubenswrapper[4556]: I0218 09:51:21.436808 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-wfnn4_428e282a-fe3e-47f5-9d19-deb95236a4c9/cert-manager-cainjector/0.log" Feb 18 09:51:21 crc kubenswrapper[4556]: I0218 09:51:21.479264 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-j67wz_58f7a628-b1a3-4fd5-a730-0a80524f751c/cert-manager-webhook/0.log" Feb 18 09:51:22 crc kubenswrapper[4556]: I0218 09:51:22.376842 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bk6mp" podUID="7baa1e0e-e912-4e07-8e0b-2fa63fc6d2df" containerName="registry-server" containerID="cri-o://a95c751826ee352908fbe98d82f9be57e14e98302282745acf5fea6ebc5463a6" gracePeriod=2 Feb 18 09:51:22 crc kubenswrapper[4556]: I0218 09:51:22.773405 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bk6mp" Feb 18 09:51:22 crc kubenswrapper[4556]: I0218 09:51:22.910855 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7baa1e0e-e912-4e07-8e0b-2fa63fc6d2df-utilities\") pod \"7baa1e0e-e912-4e07-8e0b-2fa63fc6d2df\" (UID: \"7baa1e0e-e912-4e07-8e0b-2fa63fc6d2df\") " Feb 18 09:51:22 crc kubenswrapper[4556]: I0218 09:51:22.911034 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7baa1e0e-e912-4e07-8e0b-2fa63fc6d2df-catalog-content\") pod \"7baa1e0e-e912-4e07-8e0b-2fa63fc6d2df\" (UID: \"7baa1e0e-e912-4e07-8e0b-2fa63fc6d2df\") " Feb 18 09:51:22 crc kubenswrapper[4556]: I0218 09:51:22.911207 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ckrm8\" (UniqueName: \"kubernetes.io/projected/7baa1e0e-e912-4e07-8e0b-2fa63fc6d2df-kube-api-access-ckrm8\") pod \"7baa1e0e-e912-4e07-8e0b-2fa63fc6d2df\" (UID: \"7baa1e0e-e912-4e07-8e0b-2fa63fc6d2df\") " Feb 18 09:51:22 crc kubenswrapper[4556]: I0218 09:51:22.918853 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7baa1e0e-e912-4e07-8e0b-2fa63fc6d2df-utilities" (OuterVolumeSpecName: "utilities") pod "7baa1e0e-e912-4e07-8e0b-2fa63fc6d2df" (UID: "7baa1e0e-e912-4e07-8e0b-2fa63fc6d2df"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:51:22 crc kubenswrapper[4556]: I0218 09:51:22.920821 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7baa1e0e-e912-4e07-8e0b-2fa63fc6d2df-kube-api-access-ckrm8" (OuterVolumeSpecName: "kube-api-access-ckrm8") pod "7baa1e0e-e912-4e07-8e0b-2fa63fc6d2df" (UID: "7baa1e0e-e912-4e07-8e0b-2fa63fc6d2df"). InnerVolumeSpecName "kube-api-access-ckrm8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:51:22 crc kubenswrapper[4556]: I0218 09:51:22.933034 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7baa1e0e-e912-4e07-8e0b-2fa63fc6d2df-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7baa1e0e-e912-4e07-8e0b-2fa63fc6d2df" (UID: "7baa1e0e-e912-4e07-8e0b-2fa63fc6d2df"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:51:23 crc kubenswrapper[4556]: I0218 09:51:23.013534 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ckrm8\" (UniqueName: \"kubernetes.io/projected/7baa1e0e-e912-4e07-8e0b-2fa63fc6d2df-kube-api-access-ckrm8\") on node \"crc\" DevicePath \"\"" Feb 18 09:51:23 crc kubenswrapper[4556]: I0218 09:51:23.013727 4556 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7baa1e0e-e912-4e07-8e0b-2fa63fc6d2df-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 09:51:23 crc kubenswrapper[4556]: I0218 09:51:23.013809 4556 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7baa1e0e-e912-4e07-8e0b-2fa63fc6d2df-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 09:51:23 crc kubenswrapper[4556]: I0218 09:51:23.390084 4556 generic.go:334] "Generic (PLEG): container finished" podID="7baa1e0e-e912-4e07-8e0b-2fa63fc6d2df" containerID="a95c751826ee352908fbe98d82f9be57e14e98302282745acf5fea6ebc5463a6" exitCode=0 Feb 18 09:51:23 crc kubenswrapper[4556]: I0218 09:51:23.390130 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bk6mp" event={"ID":"7baa1e0e-e912-4e07-8e0b-2fa63fc6d2df","Type":"ContainerDied","Data":"a95c751826ee352908fbe98d82f9be57e14e98302282745acf5fea6ebc5463a6"} Feb 18 09:51:23 crc kubenswrapper[4556]: I0218 09:51:23.390178 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bk6mp" event={"ID":"7baa1e0e-e912-4e07-8e0b-2fa63fc6d2df","Type":"ContainerDied","Data":"1eb4ff368db0c3e837a96074db9a2948bea0478bcaadf45f705a843ee8e9339c"} Feb 18 09:51:23 crc kubenswrapper[4556]: I0218 09:51:23.390200 4556 scope.go:117] "RemoveContainer" containerID="a95c751826ee352908fbe98d82f9be57e14e98302282745acf5fea6ebc5463a6" Feb 18 09:51:23 crc kubenswrapper[4556]: I0218 09:51:23.390338 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bk6mp" Feb 18 09:51:23 crc kubenswrapper[4556]: I0218 09:51:23.415018 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bk6mp"] Feb 18 09:51:23 crc kubenswrapper[4556]: I0218 09:51:23.419688 4556 scope.go:117] "RemoveContainer" containerID="2085a1c77064ac6829c3ddb4bc3d5e771a544a9bfea9e9109761848030518f4b" Feb 18 09:51:23 crc kubenswrapper[4556]: I0218 09:51:23.423735 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bk6mp"] Feb 18 09:51:23 crc kubenswrapper[4556]: I0218 09:51:23.441131 4556 scope.go:117] "RemoveContainer" containerID="eb8d9d6b6b6049cc45f005bf587f42483d2e577b529949a59dcec476cc617bed" Feb 18 09:51:23 crc kubenswrapper[4556]: I0218 09:51:23.476555 4556 scope.go:117] "RemoveContainer" containerID="a95c751826ee352908fbe98d82f9be57e14e98302282745acf5fea6ebc5463a6" Feb 18 09:51:23 crc kubenswrapper[4556]: E0218 09:51:23.477223 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a95c751826ee352908fbe98d82f9be57e14e98302282745acf5fea6ebc5463a6\": container with ID starting with a95c751826ee352908fbe98d82f9be57e14e98302282745acf5fea6ebc5463a6 not found: ID does not exist" containerID="a95c751826ee352908fbe98d82f9be57e14e98302282745acf5fea6ebc5463a6" Feb 18 09:51:23 crc kubenswrapper[4556]: I0218 09:51:23.477255 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a95c751826ee352908fbe98d82f9be57e14e98302282745acf5fea6ebc5463a6"} err="failed to get container status \"a95c751826ee352908fbe98d82f9be57e14e98302282745acf5fea6ebc5463a6\": rpc error: code = NotFound desc = could not find container \"a95c751826ee352908fbe98d82f9be57e14e98302282745acf5fea6ebc5463a6\": container with ID starting with a95c751826ee352908fbe98d82f9be57e14e98302282745acf5fea6ebc5463a6 not found: ID does not exist" Feb 18 09:51:23 crc kubenswrapper[4556]: I0218 09:51:23.477279 4556 scope.go:117] "RemoveContainer" containerID="2085a1c77064ac6829c3ddb4bc3d5e771a544a9bfea9e9109761848030518f4b" Feb 18 09:51:23 crc kubenswrapper[4556]: E0218 09:51:23.477597 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2085a1c77064ac6829c3ddb4bc3d5e771a544a9bfea9e9109761848030518f4b\": container with ID starting with 2085a1c77064ac6829c3ddb4bc3d5e771a544a9bfea9e9109761848030518f4b not found: ID does not exist" containerID="2085a1c77064ac6829c3ddb4bc3d5e771a544a9bfea9e9109761848030518f4b" Feb 18 09:51:23 crc kubenswrapper[4556]: I0218 09:51:23.477617 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2085a1c77064ac6829c3ddb4bc3d5e771a544a9bfea9e9109761848030518f4b"} err="failed to get container status \"2085a1c77064ac6829c3ddb4bc3d5e771a544a9bfea9e9109761848030518f4b\": rpc error: code = NotFound desc = could not find container \"2085a1c77064ac6829c3ddb4bc3d5e771a544a9bfea9e9109761848030518f4b\": container with ID starting with 2085a1c77064ac6829c3ddb4bc3d5e771a544a9bfea9e9109761848030518f4b not found: ID does not exist" Feb 18 09:51:23 crc kubenswrapper[4556]: I0218 09:51:23.477630 4556 scope.go:117] "RemoveContainer" containerID="eb8d9d6b6b6049cc45f005bf587f42483d2e577b529949a59dcec476cc617bed" Feb 18 09:51:23 crc kubenswrapper[4556]: E0218 09:51:23.477863 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb8d9d6b6b6049cc45f005bf587f42483d2e577b529949a59dcec476cc617bed\": container with ID starting with eb8d9d6b6b6049cc45f005bf587f42483d2e577b529949a59dcec476cc617bed not found: ID does not exist" containerID="eb8d9d6b6b6049cc45f005bf587f42483d2e577b529949a59dcec476cc617bed" Feb 18 09:51:23 crc kubenswrapper[4556]: I0218 09:51:23.477884 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb8d9d6b6b6049cc45f005bf587f42483d2e577b529949a59dcec476cc617bed"} err="failed to get container status \"eb8d9d6b6b6049cc45f005bf587f42483d2e577b529949a59dcec476cc617bed\": rpc error: code = NotFound desc = could not find container \"eb8d9d6b6b6049cc45f005bf587f42483d2e577b529949a59dcec476cc617bed\": container with ID starting with eb8d9d6b6b6049cc45f005bf587f42483d2e577b529949a59dcec476cc617bed not found: ID does not exist" Feb 18 09:51:25 crc kubenswrapper[4556]: I0218 09:51:25.291465 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7baa1e0e-e912-4e07-8e0b-2fa63fc6d2df" path="/var/lib/kubelet/pods/7baa1e0e-e912-4e07-8e0b-2fa63fc6d2df/volumes" Feb 18 09:51:31 crc kubenswrapper[4556]: I0218 09:51:31.727297 4556 patch_prober.go:28] interesting pod/machine-config-daemon-f76hs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 09:51:31 crc kubenswrapper[4556]: I0218 09:51:31.727897 4556 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 09:51:32 crc kubenswrapper[4556]: I0218 09:51:32.000500 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5c78fc5d65-lzgcp_bccd4d34-9d89-40ef-9d4e-4c6a58d2571b/nmstate-console-plugin/0.log" Feb 18 09:51:32 crc kubenswrapper[4556]: I0218 09:51:32.158220 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-7dmwd_9a59cd87-e29a-4cf4-a407-3de0680bc1dc/nmstate-handler/0.log" Feb 18 09:51:32 crc kubenswrapper[4556]: I0218 09:51:32.202400 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58c85c668d-nhd7k_0933417f-4853-4022-b87e-7f3584341e8d/kube-rbac-proxy/0.log" Feb 18 09:51:32 crc kubenswrapper[4556]: I0218 09:51:32.208355 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58c85c668d-nhd7k_0933417f-4853-4022-b87e-7f3584341e8d/nmstate-metrics/0.log" Feb 18 09:51:32 crc kubenswrapper[4556]: I0218 09:51:32.365135 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-866bcb46dc-cm8rm_df346433-edc0-4b09-b028-ef60228567d1/nmstate-webhook/0.log" Feb 18 09:51:32 crc kubenswrapper[4556]: I0218 09:51:32.393468 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-694c9596b7-mrrmv_b2ae390e-7b87-4f49-b95b-c39e89bbf523/nmstate-operator/0.log" Feb 18 09:51:55 crc kubenswrapper[4556]: I0218 09:51:55.686523 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-69bbfbf88f-l7dg7_cb943421-d850-4010-817d-15920051ce04/kube-rbac-proxy/0.log" Feb 18 09:51:55 crc kubenswrapper[4556]: I0218 09:51:55.825638 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-69bbfbf88f-l7dg7_cb943421-d850-4010-817d-15920051ce04/controller/0.log" Feb 18 09:51:55 crc kubenswrapper[4556]: I0218 09:51:55.910764 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qltwk_7cb5a650-ed14-42c7-8655-c5dfcaeb9d40/cp-frr-files/0.log" Feb 18 09:51:56 crc kubenswrapper[4556]: I0218 09:51:56.027474 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qltwk_7cb5a650-ed14-42c7-8655-c5dfcaeb9d40/cp-reloader/0.log" Feb 18 09:51:56 crc kubenswrapper[4556]: I0218 09:51:56.040957 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qltwk_7cb5a650-ed14-42c7-8655-c5dfcaeb9d40/cp-metrics/0.log" Feb 18 09:51:56 crc kubenswrapper[4556]: I0218 09:51:56.079558 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qltwk_7cb5a650-ed14-42c7-8655-c5dfcaeb9d40/cp-frr-files/0.log" Feb 18 09:51:56 crc kubenswrapper[4556]: I0218 09:51:56.096333 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qltwk_7cb5a650-ed14-42c7-8655-c5dfcaeb9d40/cp-reloader/0.log" Feb 18 09:51:56 crc kubenswrapper[4556]: I0218 09:51:56.254234 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qltwk_7cb5a650-ed14-42c7-8655-c5dfcaeb9d40/cp-frr-files/0.log" Feb 18 09:51:56 crc kubenswrapper[4556]: I0218 09:51:56.263369 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qltwk_7cb5a650-ed14-42c7-8655-c5dfcaeb9d40/cp-metrics/0.log" Feb 18 09:51:56 crc kubenswrapper[4556]: I0218 09:51:56.292017 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qltwk_7cb5a650-ed14-42c7-8655-c5dfcaeb9d40/cp-reloader/0.log" Feb 18 09:51:56 crc kubenswrapper[4556]: I0218 09:51:56.297887 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qltwk_7cb5a650-ed14-42c7-8655-c5dfcaeb9d40/cp-metrics/0.log" Feb 18 09:51:56 crc kubenswrapper[4556]: I0218 09:51:56.485741 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qltwk_7cb5a650-ed14-42c7-8655-c5dfcaeb9d40/cp-reloader/0.log" Feb 18 09:51:56 crc kubenswrapper[4556]: I0218 09:51:56.496773 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qltwk_7cb5a650-ed14-42c7-8655-c5dfcaeb9d40/controller/0.log" Feb 18 09:51:56 crc kubenswrapper[4556]: I0218 09:51:56.505795 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qltwk_7cb5a650-ed14-42c7-8655-c5dfcaeb9d40/cp-frr-files/0.log" Feb 18 09:51:56 crc kubenswrapper[4556]: I0218 09:51:56.551953 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qltwk_7cb5a650-ed14-42c7-8655-c5dfcaeb9d40/cp-metrics/0.log" Feb 18 09:51:56 crc kubenswrapper[4556]: I0218 09:51:56.648537 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qltwk_7cb5a650-ed14-42c7-8655-c5dfcaeb9d40/frr-metrics/0.log" Feb 18 09:51:56 crc kubenswrapper[4556]: I0218 09:51:56.672131 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qltwk_7cb5a650-ed14-42c7-8655-c5dfcaeb9d40/kube-rbac-proxy/0.log" Feb 18 09:51:56 crc kubenswrapper[4556]: I0218 09:51:56.756668 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qltwk_7cb5a650-ed14-42c7-8655-c5dfcaeb9d40/kube-rbac-proxy-frr/0.log" Feb 18 09:51:56 crc kubenswrapper[4556]: I0218 09:51:56.836855 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qltwk_7cb5a650-ed14-42c7-8655-c5dfcaeb9d40/reloader/0.log" Feb 18 09:51:56 crc kubenswrapper[4556]: I0218 09:51:56.926313 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-78b44bf5bb-spppg_04e46782-3a29-4a3d-8b82-ff9f8d34090c/frr-k8s-webhook-server/0.log" Feb 18 09:51:57 crc kubenswrapper[4556]: I0218 09:51:57.104546 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-7cccb48f48-gbb9s_6750b0ea-8f96-404d-bd38-e04e8be95127/manager/0.log" Feb 18 09:51:57 crc kubenswrapper[4556]: I0218 09:51:57.220390 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-749d56bf4-c8bg9_414515f8-ef03-41cd-a1c8-7811efcca959/webhook-server/0.log" Feb 18 09:51:57 crc kubenswrapper[4556]: I0218 09:51:57.338553 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-cjpmm_90e14ccf-fe48-4f1a-85af-5e7aa2cc1874/kube-rbac-proxy/0.log" Feb 18 09:51:57 crc kubenswrapper[4556]: I0218 09:51:57.813932 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-cjpmm_90e14ccf-fe48-4f1a-85af-5e7aa2cc1874/speaker/0.log" Feb 18 09:51:57 crc kubenswrapper[4556]: I0218 09:51:57.845362 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qltwk_7cb5a650-ed14-42c7-8655-c5dfcaeb9d40/frr/0.log" Feb 18 09:52:01 crc kubenswrapper[4556]: I0218 09:52:01.727868 4556 patch_prober.go:28] interesting pod/machine-config-daemon-f76hs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 09:52:01 crc kubenswrapper[4556]: I0218 09:52:01.728627 4556 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 09:52:08 crc kubenswrapper[4556]: I0218 09:52:08.882436 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gp95f_3adcbc7e-17be-4546-bd1b-38b6ddee82b6/util/0.log" Feb 18 09:52:09 crc kubenswrapper[4556]: I0218 09:52:09.045818 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gp95f_3adcbc7e-17be-4546-bd1b-38b6ddee82b6/util/0.log" Feb 18 09:52:09 crc kubenswrapper[4556]: I0218 09:52:09.075121 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gp95f_3adcbc7e-17be-4546-bd1b-38b6ddee82b6/pull/0.log" Feb 18 09:52:09 crc kubenswrapper[4556]: I0218 09:52:09.093445 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gp95f_3adcbc7e-17be-4546-bd1b-38b6ddee82b6/pull/0.log" Feb 18 09:52:09 crc kubenswrapper[4556]: I0218 09:52:09.225545 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gp95f_3adcbc7e-17be-4546-bd1b-38b6ddee82b6/util/0.log" Feb 18 09:52:09 crc kubenswrapper[4556]: I0218 09:52:09.231206 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gp95f_3adcbc7e-17be-4546-bd1b-38b6ddee82b6/extract/0.log" Feb 18 09:52:09 crc kubenswrapper[4556]: I0218 09:52:09.247994 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gp95f_3adcbc7e-17be-4546-bd1b-38b6ddee82b6/pull/0.log" Feb 18 09:52:09 crc kubenswrapper[4556]: I0218 09:52:09.396237 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bfrcb_6ba6b57c-a8a5-4692-955d-e851c1014fa4/extract-utilities/0.log" Feb 18 09:52:09 crc kubenswrapper[4556]: I0218 09:52:09.524906 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bfrcb_6ba6b57c-a8a5-4692-955d-e851c1014fa4/extract-utilities/0.log" Feb 18 09:52:09 crc kubenswrapper[4556]: I0218 09:52:09.525036 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bfrcb_6ba6b57c-a8a5-4692-955d-e851c1014fa4/extract-content/0.log" Feb 18 09:52:09 crc kubenswrapper[4556]: I0218 09:52:09.542044 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bfrcb_6ba6b57c-a8a5-4692-955d-e851c1014fa4/extract-content/0.log" Feb 18 09:52:09 crc kubenswrapper[4556]: I0218 09:52:09.711471 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bfrcb_6ba6b57c-a8a5-4692-955d-e851c1014fa4/extract-content/0.log" Feb 18 09:52:09 crc kubenswrapper[4556]: I0218 09:52:09.773396 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bfrcb_6ba6b57c-a8a5-4692-955d-e851c1014fa4/extract-utilities/0.log" Feb 18 09:52:09 crc kubenswrapper[4556]: I0218 09:52:09.907976 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ldrzh_745b3a46-b0b9-443b-95d2-6bfd635d1243/extract-utilities/0.log" Feb 18 09:52:10 crc kubenswrapper[4556]: I0218 09:52:10.037378 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bfrcb_6ba6b57c-a8a5-4692-955d-e851c1014fa4/registry-server/0.log" Feb 18 09:52:10 crc kubenswrapper[4556]: I0218 09:52:10.055051 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ldrzh_745b3a46-b0b9-443b-95d2-6bfd635d1243/extract-utilities/0.log" Feb 18 09:52:10 crc kubenswrapper[4556]: I0218 09:52:10.075879 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ldrzh_745b3a46-b0b9-443b-95d2-6bfd635d1243/extract-content/0.log" Feb 18 09:52:10 crc kubenswrapper[4556]: I0218 09:52:10.104415 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ldrzh_745b3a46-b0b9-443b-95d2-6bfd635d1243/extract-content/0.log" Feb 18 09:52:10 crc kubenswrapper[4556]: I0218 09:52:10.329051 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ldrzh_745b3a46-b0b9-443b-95d2-6bfd635d1243/extract-content/0.log" Feb 18 09:52:10 crc kubenswrapper[4556]: I0218 09:52:10.370299 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ldrzh_745b3a46-b0b9-443b-95d2-6bfd635d1243/extract-utilities/0.log" Feb 18 09:52:10 crc kubenswrapper[4556]: I0218 09:52:10.489145 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecal8bm8_6b5ce21b-7d77-4d8c-b2d6-272ab408d929/util/0.log" Feb 18 09:52:10 crc kubenswrapper[4556]: I0218 09:52:10.714670 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ldrzh_745b3a46-b0b9-443b-95d2-6bfd635d1243/registry-server/0.log" Feb 18 09:52:10 crc kubenswrapper[4556]: I0218 09:52:10.732702 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecal8bm8_6b5ce21b-7d77-4d8c-b2d6-272ab408d929/util/0.log" Feb 18 09:52:10 crc kubenswrapper[4556]: I0218 09:52:10.760294 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecal8bm8_6b5ce21b-7d77-4d8c-b2d6-272ab408d929/pull/0.log" Feb 18 09:52:10 crc kubenswrapper[4556]: I0218 09:52:10.764647 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecal8bm8_6b5ce21b-7d77-4d8c-b2d6-272ab408d929/pull/0.log" Feb 18 09:52:10 crc kubenswrapper[4556]: I0218 09:52:10.897755 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecal8bm8_6b5ce21b-7d77-4d8c-b2d6-272ab408d929/util/0.log" Feb 18 09:52:10 crc kubenswrapper[4556]: I0218 09:52:10.939882 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecal8bm8_6b5ce21b-7d77-4d8c-b2d6-272ab408d929/pull/0.log" Feb 18 09:52:10 crc kubenswrapper[4556]: I0218 09:52:10.947689 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecal8bm8_6b5ce21b-7d77-4d8c-b2d6-272ab408d929/extract/0.log" Feb 18 09:52:11 crc kubenswrapper[4556]: I0218 09:52:11.092640 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-vh264_070f724e-944f-48d2-bb39-67b98f7667ce/marketplace-operator/0.log" Feb 18 09:52:11 crc kubenswrapper[4556]: I0218 09:52:11.131925 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m8g6v_7e5017d0-70be-4177-9c9c-827f97ed470b/extract-utilities/0.log" Feb 18 09:52:11 crc kubenswrapper[4556]: I0218 09:52:11.266313 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m8g6v_7e5017d0-70be-4177-9c9c-827f97ed470b/extract-utilities/0.log" Feb 18 09:52:11 crc kubenswrapper[4556]: I0218 09:52:11.271882 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m8g6v_7e5017d0-70be-4177-9c9c-827f97ed470b/extract-content/0.log" Feb 18 09:52:11 crc kubenswrapper[4556]: I0218 09:52:11.319292 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m8g6v_7e5017d0-70be-4177-9c9c-827f97ed470b/extract-content/0.log" Feb 18 09:52:11 crc kubenswrapper[4556]: I0218 09:52:11.499992 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m8g6v_7e5017d0-70be-4177-9c9c-827f97ed470b/extract-utilities/0.log" Feb 18 09:52:11 crc kubenswrapper[4556]: I0218 09:52:11.517408 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m8g6v_7e5017d0-70be-4177-9c9c-827f97ed470b/extract-content/0.log" Feb 18 09:52:11 crc kubenswrapper[4556]: I0218 09:52:11.602687 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m8g6v_7e5017d0-70be-4177-9c9c-827f97ed470b/registry-server/0.log" Feb 18 09:52:11 crc kubenswrapper[4556]: I0218 09:52:11.700044 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-w6td5_e82c70e2-ba17-4919-a836-a6e252af6380/extract-utilities/0.log" Feb 18 09:52:11 crc kubenswrapper[4556]: I0218 09:52:11.823773 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-w6td5_e82c70e2-ba17-4919-a836-a6e252af6380/extract-utilities/0.log" Feb 18 09:52:11 crc kubenswrapper[4556]: I0218 09:52:11.860589 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-w6td5_e82c70e2-ba17-4919-a836-a6e252af6380/extract-content/0.log" Feb 18 09:52:11 crc kubenswrapper[4556]: I0218 09:52:11.862371 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-w6td5_e82c70e2-ba17-4919-a836-a6e252af6380/extract-content/0.log" Feb 18 09:52:12 crc kubenswrapper[4556]: I0218 09:52:12.021044 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-w6td5_e82c70e2-ba17-4919-a836-a6e252af6380/extract-content/0.log" Feb 18 09:52:12 crc kubenswrapper[4556]: I0218 09:52:12.027807 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-w6td5_e82c70e2-ba17-4919-a836-a6e252af6380/extract-utilities/0.log" Feb 18 09:52:12 crc kubenswrapper[4556]: I0218 09:52:12.363090 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-w6td5_e82c70e2-ba17-4919-a836-a6e252af6380/registry-server/0.log" Feb 18 09:52:16 crc kubenswrapper[4556]: I0218 09:52:16.854390 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4ndmh"] Feb 18 09:52:16 crc kubenswrapper[4556]: E0218 09:52:16.858025 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7baa1e0e-e912-4e07-8e0b-2fa63fc6d2df" containerName="registry-server" Feb 18 09:52:16 crc kubenswrapper[4556]: I0218 09:52:16.858171 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="7baa1e0e-e912-4e07-8e0b-2fa63fc6d2df" containerName="registry-server" Feb 18 09:52:16 crc kubenswrapper[4556]: E0218 09:52:16.858254 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7baa1e0e-e912-4e07-8e0b-2fa63fc6d2df" containerName="extract-content" Feb 18 09:52:16 crc kubenswrapper[4556]: I0218 09:52:16.858314 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="7baa1e0e-e912-4e07-8e0b-2fa63fc6d2df" containerName="extract-content" Feb 18 09:52:16 crc kubenswrapper[4556]: E0218 09:52:16.858418 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7baa1e0e-e912-4e07-8e0b-2fa63fc6d2df" containerName="extract-utilities" Feb 18 09:52:16 crc kubenswrapper[4556]: I0218 09:52:16.858478 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="7baa1e0e-e912-4e07-8e0b-2fa63fc6d2df" containerName="extract-utilities" Feb 18 09:52:16 crc kubenswrapper[4556]: I0218 09:52:16.858818 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="7baa1e0e-e912-4e07-8e0b-2fa63fc6d2df" containerName="registry-server" Feb 18 09:52:16 crc kubenswrapper[4556]: I0218 09:52:16.860673 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4ndmh" Feb 18 09:52:16 crc kubenswrapper[4556]: I0218 09:52:16.871395 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4ndmh"] Feb 18 09:52:16 crc kubenswrapper[4556]: I0218 09:52:16.911659 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45zh6\" (UniqueName: \"kubernetes.io/projected/02bdbd38-9b5c-49e5-933e-46363b4e6c1a-kube-api-access-45zh6\") pod \"redhat-operators-4ndmh\" (UID: \"02bdbd38-9b5c-49e5-933e-46363b4e6c1a\") " pod="openshift-marketplace/redhat-operators-4ndmh" Feb 18 09:52:16 crc kubenswrapper[4556]: I0218 09:52:16.911851 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02bdbd38-9b5c-49e5-933e-46363b4e6c1a-catalog-content\") pod \"redhat-operators-4ndmh\" (UID: \"02bdbd38-9b5c-49e5-933e-46363b4e6c1a\") " pod="openshift-marketplace/redhat-operators-4ndmh" Feb 18 09:52:16 crc kubenswrapper[4556]: I0218 09:52:16.912079 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02bdbd38-9b5c-49e5-933e-46363b4e6c1a-utilities\") pod \"redhat-operators-4ndmh\" (UID: \"02bdbd38-9b5c-49e5-933e-46363b4e6c1a\") " pod="openshift-marketplace/redhat-operators-4ndmh" Feb 18 09:52:17 crc kubenswrapper[4556]: I0218 09:52:17.013968 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45zh6\" (UniqueName: \"kubernetes.io/projected/02bdbd38-9b5c-49e5-933e-46363b4e6c1a-kube-api-access-45zh6\") pod \"redhat-operators-4ndmh\" (UID: \"02bdbd38-9b5c-49e5-933e-46363b4e6c1a\") " pod="openshift-marketplace/redhat-operators-4ndmh" Feb 18 09:52:17 crc kubenswrapper[4556]: I0218 09:52:17.014092 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02bdbd38-9b5c-49e5-933e-46363b4e6c1a-catalog-content\") pod \"redhat-operators-4ndmh\" (UID: \"02bdbd38-9b5c-49e5-933e-46363b4e6c1a\") " pod="openshift-marketplace/redhat-operators-4ndmh" Feb 18 09:52:17 crc kubenswrapper[4556]: I0218 09:52:17.014274 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02bdbd38-9b5c-49e5-933e-46363b4e6c1a-utilities\") pod \"redhat-operators-4ndmh\" (UID: \"02bdbd38-9b5c-49e5-933e-46363b4e6c1a\") " pod="openshift-marketplace/redhat-operators-4ndmh" Feb 18 09:52:17 crc kubenswrapper[4556]: I0218 09:52:17.014549 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02bdbd38-9b5c-49e5-933e-46363b4e6c1a-catalog-content\") pod \"redhat-operators-4ndmh\" (UID: \"02bdbd38-9b5c-49e5-933e-46363b4e6c1a\") " pod="openshift-marketplace/redhat-operators-4ndmh" Feb 18 09:52:17 crc kubenswrapper[4556]: I0218 09:52:17.014746 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02bdbd38-9b5c-49e5-933e-46363b4e6c1a-utilities\") pod \"redhat-operators-4ndmh\" (UID: \"02bdbd38-9b5c-49e5-933e-46363b4e6c1a\") " pod="openshift-marketplace/redhat-operators-4ndmh" Feb 18 09:52:17 crc kubenswrapper[4556]: I0218 09:52:17.038593 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45zh6\" (UniqueName: \"kubernetes.io/projected/02bdbd38-9b5c-49e5-933e-46363b4e6c1a-kube-api-access-45zh6\") pod \"redhat-operators-4ndmh\" (UID: \"02bdbd38-9b5c-49e5-933e-46363b4e6c1a\") " pod="openshift-marketplace/redhat-operators-4ndmh" Feb 18 09:52:17 crc kubenswrapper[4556]: I0218 09:52:17.183037 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4ndmh" Feb 18 09:52:17 crc kubenswrapper[4556]: I0218 09:52:17.619328 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4ndmh"] Feb 18 09:52:17 crc kubenswrapper[4556]: I0218 09:52:17.883526 4556 generic.go:334] "Generic (PLEG): container finished" podID="02bdbd38-9b5c-49e5-933e-46363b4e6c1a" containerID="5fd97cfc440ea2d2ba29132e87115afed6842e3fd95f80aad272dfda1ce77061" exitCode=0 Feb 18 09:52:17 crc kubenswrapper[4556]: I0218 09:52:17.883771 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4ndmh" event={"ID":"02bdbd38-9b5c-49e5-933e-46363b4e6c1a","Type":"ContainerDied","Data":"5fd97cfc440ea2d2ba29132e87115afed6842e3fd95f80aad272dfda1ce77061"} Feb 18 09:52:17 crc kubenswrapper[4556]: I0218 09:52:17.883860 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4ndmh" event={"ID":"02bdbd38-9b5c-49e5-933e-46363b4e6c1a","Type":"ContainerStarted","Data":"2115d5ef8e8310d4465ed7fc2408013765093d85fafcc14242a388fad84eab8e"} Feb 18 09:52:18 crc kubenswrapper[4556]: I0218 09:52:18.895167 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4ndmh" event={"ID":"02bdbd38-9b5c-49e5-933e-46363b4e6c1a","Type":"ContainerStarted","Data":"ffffd8972e5a652d44498f4b4894fe56fd8831e3d57faaa1d4c3ab44e67bd5f5"} Feb 18 09:52:19 crc kubenswrapper[4556]: I0218 09:52:19.906904 4556 generic.go:334] "Generic (PLEG): container finished" podID="02bdbd38-9b5c-49e5-933e-46363b4e6c1a" containerID="ffffd8972e5a652d44498f4b4894fe56fd8831e3d57faaa1d4c3ab44e67bd5f5" exitCode=0 Feb 18 09:52:19 crc kubenswrapper[4556]: I0218 09:52:19.907005 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4ndmh" event={"ID":"02bdbd38-9b5c-49e5-933e-46363b4e6c1a","Type":"ContainerDied","Data":"ffffd8972e5a652d44498f4b4894fe56fd8831e3d57faaa1d4c3ab44e67bd5f5"} Feb 18 09:52:20 crc kubenswrapper[4556]: I0218 09:52:20.917545 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4ndmh" event={"ID":"02bdbd38-9b5c-49e5-933e-46363b4e6c1a","Type":"ContainerStarted","Data":"09aca20feb9abc18dd975b03af6e98a38e2301affcf7a6bd41715083ebaf2f4f"} Feb 18 09:52:20 crc kubenswrapper[4556]: I0218 09:52:20.936246 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4ndmh" podStartSLOduration=2.470677169 podStartE2EDuration="4.936225761s" podCreationTimestamp="2026-02-18 09:52:16 +0000 UTC" firstStartedPulling="2026-02-18 09:52:17.885499372 +0000 UTC m=+2894.902460352" lastFinishedPulling="2026-02-18 09:52:20.351047974 +0000 UTC m=+2897.368008944" observedRunningTime="2026-02-18 09:52:20.930720746 +0000 UTC m=+2897.947681726" watchObservedRunningTime="2026-02-18 09:52:20.936225761 +0000 UTC m=+2897.953186741" Feb 18 09:52:27 crc kubenswrapper[4556]: I0218 09:52:27.183753 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4ndmh" Feb 18 09:52:27 crc kubenswrapper[4556]: I0218 09:52:27.185314 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4ndmh" Feb 18 09:52:27 crc kubenswrapper[4556]: I0218 09:52:27.229137 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4ndmh" Feb 18 09:52:28 crc kubenswrapper[4556]: I0218 09:52:28.018234 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4ndmh" Feb 18 09:52:28 crc kubenswrapper[4556]: I0218 09:52:28.060619 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4ndmh"] Feb 18 09:52:29 crc kubenswrapper[4556]: I0218 09:52:29.988507 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-4ndmh" podUID="02bdbd38-9b5c-49e5-933e-46363b4e6c1a" containerName="registry-server" containerID="cri-o://09aca20feb9abc18dd975b03af6e98a38e2301affcf7a6bd41715083ebaf2f4f" gracePeriod=2 Feb 18 09:52:30 crc kubenswrapper[4556]: I0218 09:52:30.451089 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4ndmh" Feb 18 09:52:30 crc kubenswrapper[4556]: I0218 09:52:30.606867 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02bdbd38-9b5c-49e5-933e-46363b4e6c1a-catalog-content\") pod \"02bdbd38-9b5c-49e5-933e-46363b4e6c1a\" (UID: \"02bdbd38-9b5c-49e5-933e-46363b4e6c1a\") " Feb 18 09:52:30 crc kubenswrapper[4556]: I0218 09:52:30.607073 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02bdbd38-9b5c-49e5-933e-46363b4e6c1a-utilities\") pod \"02bdbd38-9b5c-49e5-933e-46363b4e6c1a\" (UID: \"02bdbd38-9b5c-49e5-933e-46363b4e6c1a\") " Feb 18 09:52:30 crc kubenswrapper[4556]: I0218 09:52:30.607189 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-45zh6\" (UniqueName: \"kubernetes.io/projected/02bdbd38-9b5c-49e5-933e-46363b4e6c1a-kube-api-access-45zh6\") pod \"02bdbd38-9b5c-49e5-933e-46363b4e6c1a\" (UID: \"02bdbd38-9b5c-49e5-933e-46363b4e6c1a\") " Feb 18 09:52:30 crc kubenswrapper[4556]: I0218 09:52:30.608643 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/02bdbd38-9b5c-49e5-933e-46363b4e6c1a-utilities" (OuterVolumeSpecName: "utilities") pod "02bdbd38-9b5c-49e5-933e-46363b4e6c1a" (UID: "02bdbd38-9b5c-49e5-933e-46363b4e6c1a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:52:30 crc kubenswrapper[4556]: I0218 09:52:30.627901 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02bdbd38-9b5c-49e5-933e-46363b4e6c1a-kube-api-access-45zh6" (OuterVolumeSpecName: "kube-api-access-45zh6") pod "02bdbd38-9b5c-49e5-933e-46363b4e6c1a" (UID: "02bdbd38-9b5c-49e5-933e-46363b4e6c1a"). InnerVolumeSpecName "kube-api-access-45zh6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:52:30 crc kubenswrapper[4556]: I0218 09:52:30.709851 4556 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02bdbd38-9b5c-49e5-933e-46363b4e6c1a-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 09:52:30 crc kubenswrapper[4556]: I0218 09:52:30.709884 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-45zh6\" (UniqueName: \"kubernetes.io/projected/02bdbd38-9b5c-49e5-933e-46363b4e6c1a-kube-api-access-45zh6\") on node \"crc\" DevicePath \"\"" Feb 18 09:52:30 crc kubenswrapper[4556]: I0218 09:52:30.721799 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/02bdbd38-9b5c-49e5-933e-46363b4e6c1a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "02bdbd38-9b5c-49e5-933e-46363b4e6c1a" (UID: "02bdbd38-9b5c-49e5-933e-46363b4e6c1a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:52:30 crc kubenswrapper[4556]: I0218 09:52:30.812526 4556 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02bdbd38-9b5c-49e5-933e-46363b4e6c1a-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 09:52:31 crc kubenswrapper[4556]: I0218 09:52:31.001121 4556 generic.go:334] "Generic (PLEG): container finished" podID="02bdbd38-9b5c-49e5-933e-46363b4e6c1a" containerID="09aca20feb9abc18dd975b03af6e98a38e2301affcf7a6bd41715083ebaf2f4f" exitCode=0 Feb 18 09:52:31 crc kubenswrapper[4556]: I0218 09:52:31.001189 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4ndmh" Feb 18 09:52:31 crc kubenswrapper[4556]: I0218 09:52:31.001190 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4ndmh" event={"ID":"02bdbd38-9b5c-49e5-933e-46363b4e6c1a","Type":"ContainerDied","Data":"09aca20feb9abc18dd975b03af6e98a38e2301affcf7a6bd41715083ebaf2f4f"} Feb 18 09:52:31 crc kubenswrapper[4556]: I0218 09:52:31.001533 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4ndmh" event={"ID":"02bdbd38-9b5c-49e5-933e-46363b4e6c1a","Type":"ContainerDied","Data":"2115d5ef8e8310d4465ed7fc2408013765093d85fafcc14242a388fad84eab8e"} Feb 18 09:52:31 crc kubenswrapper[4556]: I0218 09:52:31.001554 4556 scope.go:117] "RemoveContainer" containerID="09aca20feb9abc18dd975b03af6e98a38e2301affcf7a6bd41715083ebaf2f4f" Feb 18 09:52:31 crc kubenswrapper[4556]: I0218 09:52:31.038934 4556 scope.go:117] "RemoveContainer" containerID="ffffd8972e5a652d44498f4b4894fe56fd8831e3d57faaa1d4c3ab44e67bd5f5" Feb 18 09:52:31 crc kubenswrapper[4556]: I0218 09:52:31.039642 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4ndmh"] Feb 18 09:52:31 crc kubenswrapper[4556]: I0218 09:52:31.046080 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-4ndmh"] Feb 18 09:52:31 crc kubenswrapper[4556]: I0218 09:52:31.063571 4556 scope.go:117] "RemoveContainer" containerID="5fd97cfc440ea2d2ba29132e87115afed6842e3fd95f80aad272dfda1ce77061" Feb 18 09:52:31 crc kubenswrapper[4556]: I0218 09:52:31.092911 4556 scope.go:117] "RemoveContainer" containerID="09aca20feb9abc18dd975b03af6e98a38e2301affcf7a6bd41715083ebaf2f4f" Feb 18 09:52:31 crc kubenswrapper[4556]: E0218 09:52:31.093365 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09aca20feb9abc18dd975b03af6e98a38e2301affcf7a6bd41715083ebaf2f4f\": container with ID starting with 09aca20feb9abc18dd975b03af6e98a38e2301affcf7a6bd41715083ebaf2f4f not found: ID does not exist" containerID="09aca20feb9abc18dd975b03af6e98a38e2301affcf7a6bd41715083ebaf2f4f" Feb 18 09:52:31 crc kubenswrapper[4556]: I0218 09:52:31.093417 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09aca20feb9abc18dd975b03af6e98a38e2301affcf7a6bd41715083ebaf2f4f"} err="failed to get container status \"09aca20feb9abc18dd975b03af6e98a38e2301affcf7a6bd41715083ebaf2f4f\": rpc error: code = NotFound desc = could not find container \"09aca20feb9abc18dd975b03af6e98a38e2301affcf7a6bd41715083ebaf2f4f\": container with ID starting with 09aca20feb9abc18dd975b03af6e98a38e2301affcf7a6bd41715083ebaf2f4f not found: ID does not exist" Feb 18 09:52:31 crc kubenswrapper[4556]: I0218 09:52:31.093448 4556 scope.go:117] "RemoveContainer" containerID="ffffd8972e5a652d44498f4b4894fe56fd8831e3d57faaa1d4c3ab44e67bd5f5" Feb 18 09:52:31 crc kubenswrapper[4556]: E0218 09:52:31.093895 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ffffd8972e5a652d44498f4b4894fe56fd8831e3d57faaa1d4c3ab44e67bd5f5\": container with ID starting with ffffd8972e5a652d44498f4b4894fe56fd8831e3d57faaa1d4c3ab44e67bd5f5 not found: ID does not exist" containerID="ffffd8972e5a652d44498f4b4894fe56fd8831e3d57faaa1d4c3ab44e67bd5f5" Feb 18 09:52:31 crc kubenswrapper[4556]: I0218 09:52:31.093949 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffffd8972e5a652d44498f4b4894fe56fd8831e3d57faaa1d4c3ab44e67bd5f5"} err="failed to get container status \"ffffd8972e5a652d44498f4b4894fe56fd8831e3d57faaa1d4c3ab44e67bd5f5\": rpc error: code = NotFound desc = could not find container \"ffffd8972e5a652d44498f4b4894fe56fd8831e3d57faaa1d4c3ab44e67bd5f5\": container with ID starting with ffffd8972e5a652d44498f4b4894fe56fd8831e3d57faaa1d4c3ab44e67bd5f5 not found: ID does not exist" Feb 18 09:52:31 crc kubenswrapper[4556]: I0218 09:52:31.093991 4556 scope.go:117] "RemoveContainer" containerID="5fd97cfc440ea2d2ba29132e87115afed6842e3fd95f80aad272dfda1ce77061" Feb 18 09:52:31 crc kubenswrapper[4556]: E0218 09:52:31.094434 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5fd97cfc440ea2d2ba29132e87115afed6842e3fd95f80aad272dfda1ce77061\": container with ID starting with 5fd97cfc440ea2d2ba29132e87115afed6842e3fd95f80aad272dfda1ce77061 not found: ID does not exist" containerID="5fd97cfc440ea2d2ba29132e87115afed6842e3fd95f80aad272dfda1ce77061" Feb 18 09:52:31 crc kubenswrapper[4556]: I0218 09:52:31.094473 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5fd97cfc440ea2d2ba29132e87115afed6842e3fd95f80aad272dfda1ce77061"} err="failed to get container status \"5fd97cfc440ea2d2ba29132e87115afed6842e3fd95f80aad272dfda1ce77061\": rpc error: code = NotFound desc = could not find container \"5fd97cfc440ea2d2ba29132e87115afed6842e3fd95f80aad272dfda1ce77061\": container with ID starting with 5fd97cfc440ea2d2ba29132e87115afed6842e3fd95f80aad272dfda1ce77061 not found: ID does not exist" Feb 18 09:52:31 crc kubenswrapper[4556]: I0218 09:52:31.291541 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02bdbd38-9b5c-49e5-933e-46363b4e6c1a" path="/var/lib/kubelet/pods/02bdbd38-9b5c-49e5-933e-46363b4e6c1a/volumes" Feb 18 09:52:31 crc kubenswrapper[4556]: I0218 09:52:31.730189 4556 patch_prober.go:28] interesting pod/machine-config-daemon-f76hs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 09:52:31 crc kubenswrapper[4556]: I0218 09:52:31.730275 4556 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 09:52:31 crc kubenswrapper[4556]: I0218 09:52:31.730343 4556 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" Feb 18 09:52:31 crc kubenswrapper[4556]: I0218 09:52:31.731523 4556 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9627d5bcc0cabda20a5fc92b3cc4b3dfae93225012caf89894f557937f69e6f7"} pod="openshift-machine-config-operator/machine-config-daemon-f76hs" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 18 09:52:31 crc kubenswrapper[4556]: I0218 09:52:31.731595 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerName="machine-config-daemon" containerID="cri-o://9627d5bcc0cabda20a5fc92b3cc4b3dfae93225012caf89894f557937f69e6f7" gracePeriod=600 Feb 18 09:52:31 crc kubenswrapper[4556]: E0218 09:52:31.857082 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:52:32 crc kubenswrapper[4556]: I0218 09:52:32.027680 4556 generic.go:334] "Generic (PLEG): container finished" podID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerID="9627d5bcc0cabda20a5fc92b3cc4b3dfae93225012caf89894f557937f69e6f7" exitCode=0 Feb 18 09:52:32 crc kubenswrapper[4556]: I0218 09:52:32.027723 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" event={"ID":"8dac7f27-d3d1-4778-9e54-f273035a1d37","Type":"ContainerDied","Data":"9627d5bcc0cabda20a5fc92b3cc4b3dfae93225012caf89894f557937f69e6f7"} Feb 18 09:52:32 crc kubenswrapper[4556]: I0218 09:52:32.027770 4556 scope.go:117] "RemoveContainer" containerID="4267588c8394174e9461033b074def2dd2f51cf44cffb9ce5d5bceee1db04806" Feb 18 09:52:32 crc kubenswrapper[4556]: I0218 09:52:32.028547 4556 scope.go:117] "RemoveContainer" containerID="9627d5bcc0cabda20a5fc92b3cc4b3dfae93225012caf89894f557937f69e6f7" Feb 18 09:52:32 crc kubenswrapper[4556]: E0218 09:52:32.028943 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:52:43 crc kubenswrapper[4556]: I0218 09:52:43.289352 4556 scope.go:117] "RemoveContainer" containerID="9627d5bcc0cabda20a5fc92b3cc4b3dfae93225012caf89894f557937f69e6f7" Feb 18 09:52:43 crc kubenswrapper[4556]: E0218 09:52:43.290230 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:52:58 crc kubenswrapper[4556]: I0218 09:52:58.282267 4556 scope.go:117] "RemoveContainer" containerID="9627d5bcc0cabda20a5fc92b3cc4b3dfae93225012caf89894f557937f69e6f7" Feb 18 09:52:58 crc kubenswrapper[4556]: E0218 09:52:58.283221 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:53:10 crc kubenswrapper[4556]: I0218 09:53:10.282412 4556 scope.go:117] "RemoveContainer" containerID="9627d5bcc0cabda20a5fc92b3cc4b3dfae93225012caf89894f557937f69e6f7" Feb 18 09:53:10 crc kubenswrapper[4556]: E0218 09:53:10.283229 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:53:23 crc kubenswrapper[4556]: I0218 09:53:23.293863 4556 scope.go:117] "RemoveContainer" containerID="9627d5bcc0cabda20a5fc92b3cc4b3dfae93225012caf89894f557937f69e6f7" Feb 18 09:53:23 crc kubenswrapper[4556]: E0218 09:53:23.294829 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:53:38 crc kubenswrapper[4556]: I0218 09:53:38.283776 4556 scope.go:117] "RemoveContainer" containerID="9627d5bcc0cabda20a5fc92b3cc4b3dfae93225012caf89894f557937f69e6f7" Feb 18 09:53:38 crc kubenswrapper[4556]: E0218 09:53:38.284825 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:53:43 crc kubenswrapper[4556]: I0218 09:53:43.720291 4556 generic.go:334] "Generic (PLEG): container finished" podID="72cad311-e6f4-4e65-838a-59725c2e7ac8" containerID="9e745c73336e6b31f33529723106acdb0d7d13100fec9a4976db3a99ec08210e" exitCode=0 Feb 18 09:53:43 crc kubenswrapper[4556]: I0218 09:53:43.720382 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8r4vv/must-gather-2lndc" event={"ID":"72cad311-e6f4-4e65-838a-59725c2e7ac8","Type":"ContainerDied","Data":"9e745c73336e6b31f33529723106acdb0d7d13100fec9a4976db3a99ec08210e"} Feb 18 09:53:43 crc kubenswrapper[4556]: I0218 09:53:43.721905 4556 scope.go:117] "RemoveContainer" containerID="9e745c73336e6b31f33529723106acdb0d7d13100fec9a4976db3a99ec08210e" Feb 18 09:53:44 crc kubenswrapper[4556]: I0218 09:53:44.706518 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-8r4vv_must-gather-2lndc_72cad311-e6f4-4e65-838a-59725c2e7ac8/gather/0.log" Feb 18 09:53:52 crc kubenswrapper[4556]: I0218 09:53:52.283413 4556 scope.go:117] "RemoveContainer" containerID="9627d5bcc0cabda20a5fc92b3cc4b3dfae93225012caf89894f557937f69e6f7" Feb 18 09:53:52 crc kubenswrapper[4556]: E0218 09:53:52.284370 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:53:52 crc kubenswrapper[4556]: I0218 09:53:52.730993 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-8r4vv/must-gather-2lndc"] Feb 18 09:53:52 crc kubenswrapper[4556]: I0218 09:53:52.731322 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-8r4vv/must-gather-2lndc" podUID="72cad311-e6f4-4e65-838a-59725c2e7ac8" containerName="copy" containerID="cri-o://d93cdcf0d4fb6b9556819af66d583dc3838bea3598a0a9acc1da2b76bfbb94a4" gracePeriod=2 Feb 18 09:53:52 crc kubenswrapper[4556]: I0218 09:53:52.736999 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-8r4vv/must-gather-2lndc"] Feb 18 09:53:53 crc kubenswrapper[4556]: I0218 09:53:53.099800 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-8r4vv_must-gather-2lndc_72cad311-e6f4-4e65-838a-59725c2e7ac8/copy/0.log" Feb 18 09:53:53 crc kubenswrapper[4556]: I0218 09:53:53.100426 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8r4vv/must-gather-2lndc" Feb 18 09:53:53 crc kubenswrapper[4556]: I0218 09:53:53.176805 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-grhf9\" (UniqueName: \"kubernetes.io/projected/72cad311-e6f4-4e65-838a-59725c2e7ac8-kube-api-access-grhf9\") pod \"72cad311-e6f4-4e65-838a-59725c2e7ac8\" (UID: \"72cad311-e6f4-4e65-838a-59725c2e7ac8\") " Feb 18 09:53:53 crc kubenswrapper[4556]: I0218 09:53:53.177083 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/72cad311-e6f4-4e65-838a-59725c2e7ac8-must-gather-output\") pod \"72cad311-e6f4-4e65-838a-59725c2e7ac8\" (UID: \"72cad311-e6f4-4e65-838a-59725c2e7ac8\") " Feb 18 09:53:53 crc kubenswrapper[4556]: I0218 09:53:53.186295 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72cad311-e6f4-4e65-838a-59725c2e7ac8-kube-api-access-grhf9" (OuterVolumeSpecName: "kube-api-access-grhf9") pod "72cad311-e6f4-4e65-838a-59725c2e7ac8" (UID: "72cad311-e6f4-4e65-838a-59725c2e7ac8"). InnerVolumeSpecName "kube-api-access-grhf9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:53:53 crc kubenswrapper[4556]: I0218 09:53:53.279998 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-grhf9\" (UniqueName: \"kubernetes.io/projected/72cad311-e6f4-4e65-838a-59725c2e7ac8-kube-api-access-grhf9\") on node \"crc\" DevicePath \"\"" Feb 18 09:53:53 crc kubenswrapper[4556]: I0218 09:53:53.312059 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72cad311-e6f4-4e65-838a-59725c2e7ac8-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "72cad311-e6f4-4e65-838a-59725c2e7ac8" (UID: "72cad311-e6f4-4e65-838a-59725c2e7ac8"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:53:53 crc kubenswrapper[4556]: I0218 09:53:53.383596 4556 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/72cad311-e6f4-4e65-838a-59725c2e7ac8-must-gather-output\") on node \"crc\" DevicePath \"\"" Feb 18 09:53:53 crc kubenswrapper[4556]: I0218 09:53:53.827524 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-8r4vv_must-gather-2lndc_72cad311-e6f4-4e65-838a-59725c2e7ac8/copy/0.log" Feb 18 09:53:53 crc kubenswrapper[4556]: I0218 09:53:53.828128 4556 generic.go:334] "Generic (PLEG): container finished" podID="72cad311-e6f4-4e65-838a-59725c2e7ac8" containerID="d93cdcf0d4fb6b9556819af66d583dc3838bea3598a0a9acc1da2b76bfbb94a4" exitCode=143 Feb 18 09:53:53 crc kubenswrapper[4556]: I0218 09:53:53.828217 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8r4vv/must-gather-2lndc" Feb 18 09:53:53 crc kubenswrapper[4556]: I0218 09:53:53.828369 4556 scope.go:117] "RemoveContainer" containerID="d93cdcf0d4fb6b9556819af66d583dc3838bea3598a0a9acc1da2b76bfbb94a4" Feb 18 09:53:53 crc kubenswrapper[4556]: I0218 09:53:53.848589 4556 scope.go:117] "RemoveContainer" containerID="9e745c73336e6b31f33529723106acdb0d7d13100fec9a4976db3a99ec08210e" Feb 18 09:53:53 crc kubenswrapper[4556]: I0218 09:53:53.912624 4556 scope.go:117] "RemoveContainer" containerID="d93cdcf0d4fb6b9556819af66d583dc3838bea3598a0a9acc1da2b76bfbb94a4" Feb 18 09:53:53 crc kubenswrapper[4556]: E0218 09:53:53.913318 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d93cdcf0d4fb6b9556819af66d583dc3838bea3598a0a9acc1da2b76bfbb94a4\": container with ID starting with d93cdcf0d4fb6b9556819af66d583dc3838bea3598a0a9acc1da2b76bfbb94a4 not found: ID does not exist" containerID="d93cdcf0d4fb6b9556819af66d583dc3838bea3598a0a9acc1da2b76bfbb94a4" Feb 18 09:53:53 crc kubenswrapper[4556]: I0218 09:53:53.913365 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d93cdcf0d4fb6b9556819af66d583dc3838bea3598a0a9acc1da2b76bfbb94a4"} err="failed to get container status \"d93cdcf0d4fb6b9556819af66d583dc3838bea3598a0a9acc1da2b76bfbb94a4\": rpc error: code = NotFound desc = could not find container \"d93cdcf0d4fb6b9556819af66d583dc3838bea3598a0a9acc1da2b76bfbb94a4\": container with ID starting with d93cdcf0d4fb6b9556819af66d583dc3838bea3598a0a9acc1da2b76bfbb94a4 not found: ID does not exist" Feb 18 09:53:53 crc kubenswrapper[4556]: I0218 09:53:53.913397 4556 scope.go:117] "RemoveContainer" containerID="9e745c73336e6b31f33529723106acdb0d7d13100fec9a4976db3a99ec08210e" Feb 18 09:53:53 crc kubenswrapper[4556]: E0218 09:53:53.913822 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e745c73336e6b31f33529723106acdb0d7d13100fec9a4976db3a99ec08210e\": container with ID starting with 9e745c73336e6b31f33529723106acdb0d7d13100fec9a4976db3a99ec08210e not found: ID does not exist" containerID="9e745c73336e6b31f33529723106acdb0d7d13100fec9a4976db3a99ec08210e" Feb 18 09:53:53 crc kubenswrapper[4556]: I0218 09:53:53.913874 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e745c73336e6b31f33529723106acdb0d7d13100fec9a4976db3a99ec08210e"} err="failed to get container status \"9e745c73336e6b31f33529723106acdb0d7d13100fec9a4976db3a99ec08210e\": rpc error: code = NotFound desc = could not find container \"9e745c73336e6b31f33529723106acdb0d7d13100fec9a4976db3a99ec08210e\": container with ID starting with 9e745c73336e6b31f33529723106acdb0d7d13100fec9a4976db3a99ec08210e not found: ID does not exist" Feb 18 09:53:55 crc kubenswrapper[4556]: I0218 09:53:55.296848 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72cad311-e6f4-4e65-838a-59725c2e7ac8" path="/var/lib/kubelet/pods/72cad311-e6f4-4e65-838a-59725c2e7ac8/volumes" Feb 18 09:54:06 crc kubenswrapper[4556]: I0218 09:54:06.282556 4556 scope.go:117] "RemoveContainer" containerID="9627d5bcc0cabda20a5fc92b3cc4b3dfae93225012caf89894f557937f69e6f7" Feb 18 09:54:06 crc kubenswrapper[4556]: E0218 09:54:06.283480 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:54:18 crc kubenswrapper[4556]: I0218 09:54:18.282106 4556 scope.go:117] "RemoveContainer" containerID="9627d5bcc0cabda20a5fc92b3cc4b3dfae93225012caf89894f557937f69e6f7" Feb 18 09:54:18 crc kubenswrapper[4556]: E0218 09:54:18.283034 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:54:32 crc kubenswrapper[4556]: I0218 09:54:32.282738 4556 scope.go:117] "RemoveContainer" containerID="9627d5bcc0cabda20a5fc92b3cc4b3dfae93225012caf89894f557937f69e6f7" Feb 18 09:54:32 crc kubenswrapper[4556]: E0218 09:54:32.283727 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:54:47 crc kubenswrapper[4556]: I0218 09:54:47.283472 4556 scope.go:117] "RemoveContainer" containerID="9627d5bcc0cabda20a5fc92b3cc4b3dfae93225012caf89894f557937f69e6f7" Feb 18 09:54:47 crc kubenswrapper[4556]: E0218 09:54:47.284454 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:54:58 crc kubenswrapper[4556]: I0218 09:54:58.282858 4556 scope.go:117] "RemoveContainer" containerID="9627d5bcc0cabda20a5fc92b3cc4b3dfae93225012caf89894f557937f69e6f7" Feb 18 09:54:58 crc kubenswrapper[4556]: E0218 09:54:58.283877 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:55:09 crc kubenswrapper[4556]: I0218 09:55:09.282209 4556 scope.go:117] "RemoveContainer" containerID="9627d5bcc0cabda20a5fc92b3cc4b3dfae93225012caf89894f557937f69e6f7" Feb 18 09:55:09 crc kubenswrapper[4556]: E0218 09:55:09.284048 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:55:20 crc kubenswrapper[4556]: I0218 09:55:20.282590 4556 scope.go:117] "RemoveContainer" containerID="9627d5bcc0cabda20a5fc92b3cc4b3dfae93225012caf89894f557937f69e6f7" Feb 18 09:55:20 crc kubenswrapper[4556]: E0218 09:55:20.283497 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:55:34 crc kubenswrapper[4556]: I0218 09:55:34.282698 4556 scope.go:117] "RemoveContainer" containerID="9627d5bcc0cabda20a5fc92b3cc4b3dfae93225012caf89894f557937f69e6f7" Feb 18 09:55:34 crc kubenswrapper[4556]: E0218 09:55:34.283948 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:55:47 crc kubenswrapper[4556]: I0218 09:55:47.283336 4556 scope.go:117] "RemoveContainer" containerID="9627d5bcc0cabda20a5fc92b3cc4b3dfae93225012caf89894f557937f69e6f7" Feb 18 09:55:47 crc kubenswrapper[4556]: E0218 09:55:47.285126 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:55:59 crc kubenswrapper[4556]: I0218 09:55:59.282588 4556 scope.go:117] "RemoveContainer" containerID="9627d5bcc0cabda20a5fc92b3cc4b3dfae93225012caf89894f557937f69e6f7" Feb 18 09:55:59 crc kubenswrapper[4556]: E0218 09:55:59.283662 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:56:11 crc kubenswrapper[4556]: I0218 09:56:11.469487 4556 scope.go:117] "RemoveContainer" containerID="3a600196a112d7e302310bb03411499798e6c8abb5aff426684dd84e400ad0d3" Feb 18 09:56:13 crc kubenswrapper[4556]: I0218 09:56:13.288511 4556 scope.go:117] "RemoveContainer" containerID="9627d5bcc0cabda20a5fc92b3cc4b3dfae93225012caf89894f557937f69e6f7" Feb 18 09:56:13 crc kubenswrapper[4556]: E0218 09:56:13.289215 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:56:26 crc kubenswrapper[4556]: I0218 09:56:26.279692 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-rnjw6/must-gather-vhc8n"] Feb 18 09:56:26 crc kubenswrapper[4556]: E0218 09:56:26.280540 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02bdbd38-9b5c-49e5-933e-46363b4e6c1a" containerName="extract-content" Feb 18 09:56:26 crc kubenswrapper[4556]: I0218 09:56:26.280554 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="02bdbd38-9b5c-49e5-933e-46363b4e6c1a" containerName="extract-content" Feb 18 09:56:26 crc kubenswrapper[4556]: E0218 09:56:26.280579 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72cad311-e6f4-4e65-838a-59725c2e7ac8" containerName="gather" Feb 18 09:56:26 crc kubenswrapper[4556]: I0218 09:56:26.280584 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="72cad311-e6f4-4e65-838a-59725c2e7ac8" containerName="gather" Feb 18 09:56:26 crc kubenswrapper[4556]: E0218 09:56:26.280596 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72cad311-e6f4-4e65-838a-59725c2e7ac8" containerName="copy" Feb 18 09:56:26 crc kubenswrapper[4556]: I0218 09:56:26.280603 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="72cad311-e6f4-4e65-838a-59725c2e7ac8" containerName="copy" Feb 18 09:56:26 crc kubenswrapper[4556]: E0218 09:56:26.280619 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02bdbd38-9b5c-49e5-933e-46363b4e6c1a" containerName="extract-utilities" Feb 18 09:56:26 crc kubenswrapper[4556]: I0218 09:56:26.280625 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="02bdbd38-9b5c-49e5-933e-46363b4e6c1a" containerName="extract-utilities" Feb 18 09:56:26 crc kubenswrapper[4556]: E0218 09:56:26.280638 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02bdbd38-9b5c-49e5-933e-46363b4e6c1a" containerName="registry-server" Feb 18 09:56:26 crc kubenswrapper[4556]: I0218 09:56:26.280643 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="02bdbd38-9b5c-49e5-933e-46363b4e6c1a" containerName="registry-server" Feb 18 09:56:26 crc kubenswrapper[4556]: I0218 09:56:26.280801 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="72cad311-e6f4-4e65-838a-59725c2e7ac8" containerName="copy" Feb 18 09:56:26 crc kubenswrapper[4556]: I0218 09:56:26.280811 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="72cad311-e6f4-4e65-838a-59725c2e7ac8" containerName="gather" Feb 18 09:56:26 crc kubenswrapper[4556]: I0218 09:56:26.280821 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="02bdbd38-9b5c-49e5-933e-46363b4e6c1a" containerName="registry-server" Feb 18 09:56:26 crc kubenswrapper[4556]: I0218 09:56:26.281719 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rnjw6/must-gather-vhc8n" Feb 18 09:56:26 crc kubenswrapper[4556]: I0218 09:56:26.287356 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-rnjw6"/"openshift-service-ca.crt" Feb 18 09:56:26 crc kubenswrapper[4556]: I0218 09:56:26.300597 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-rnjw6"/"kube-root-ca.crt" Feb 18 09:56:26 crc kubenswrapper[4556]: I0218 09:56:26.301546 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-rnjw6/must-gather-vhc8n"] Feb 18 09:56:26 crc kubenswrapper[4556]: I0218 09:56:26.378204 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/106c448a-8930-4be7-aa0e-970ab3d8b33e-must-gather-output\") pod \"must-gather-vhc8n\" (UID: \"106c448a-8930-4be7-aa0e-970ab3d8b33e\") " pod="openshift-must-gather-rnjw6/must-gather-vhc8n" Feb 18 09:56:26 crc kubenswrapper[4556]: I0218 09:56:26.378652 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2pjv\" (UniqueName: \"kubernetes.io/projected/106c448a-8930-4be7-aa0e-970ab3d8b33e-kube-api-access-d2pjv\") pod \"must-gather-vhc8n\" (UID: \"106c448a-8930-4be7-aa0e-970ab3d8b33e\") " pod="openshift-must-gather-rnjw6/must-gather-vhc8n" Feb 18 09:56:26 crc kubenswrapper[4556]: I0218 09:56:26.480274 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2pjv\" (UniqueName: \"kubernetes.io/projected/106c448a-8930-4be7-aa0e-970ab3d8b33e-kube-api-access-d2pjv\") pod \"must-gather-vhc8n\" (UID: \"106c448a-8930-4be7-aa0e-970ab3d8b33e\") " pod="openshift-must-gather-rnjw6/must-gather-vhc8n" Feb 18 09:56:26 crc kubenswrapper[4556]: I0218 09:56:26.480755 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/106c448a-8930-4be7-aa0e-970ab3d8b33e-must-gather-output\") pod \"must-gather-vhc8n\" (UID: \"106c448a-8930-4be7-aa0e-970ab3d8b33e\") " pod="openshift-must-gather-rnjw6/must-gather-vhc8n" Feb 18 09:56:26 crc kubenswrapper[4556]: I0218 09:56:26.481198 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/106c448a-8930-4be7-aa0e-970ab3d8b33e-must-gather-output\") pod \"must-gather-vhc8n\" (UID: \"106c448a-8930-4be7-aa0e-970ab3d8b33e\") " pod="openshift-must-gather-rnjw6/must-gather-vhc8n" Feb 18 09:56:26 crc kubenswrapper[4556]: I0218 09:56:26.502699 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2pjv\" (UniqueName: \"kubernetes.io/projected/106c448a-8930-4be7-aa0e-970ab3d8b33e-kube-api-access-d2pjv\") pod \"must-gather-vhc8n\" (UID: \"106c448a-8930-4be7-aa0e-970ab3d8b33e\") " pod="openshift-must-gather-rnjw6/must-gather-vhc8n" Feb 18 09:56:26 crc kubenswrapper[4556]: I0218 09:56:26.599634 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rnjw6/must-gather-vhc8n" Feb 18 09:56:27 crc kubenswrapper[4556]: I0218 09:56:27.038540 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-rnjw6/must-gather-vhc8n"] Feb 18 09:56:27 crc kubenswrapper[4556]: I0218 09:56:27.283288 4556 scope.go:117] "RemoveContainer" containerID="9627d5bcc0cabda20a5fc92b3cc4b3dfae93225012caf89894f557937f69e6f7" Feb 18 09:56:27 crc kubenswrapper[4556]: E0218 09:56:27.283774 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:56:27 crc kubenswrapper[4556]: I0218 09:56:27.294982 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rnjw6/must-gather-vhc8n" event={"ID":"106c448a-8930-4be7-aa0e-970ab3d8b33e","Type":"ContainerStarted","Data":"ab98b80451544ff53c43502227fd54e98ead544320b3850fccc3876aa72bd333"} Feb 18 09:56:27 crc kubenswrapper[4556]: I0218 09:56:27.295018 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rnjw6/must-gather-vhc8n" event={"ID":"106c448a-8930-4be7-aa0e-970ab3d8b33e","Type":"ContainerStarted","Data":"cb2b81fd9bf67d93bfbc03d3c8377be220d9ea3c89ce0f25ca2687b639210b10"} Feb 18 09:56:28 crc kubenswrapper[4556]: I0218 09:56:28.300022 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rnjw6/must-gather-vhc8n" event={"ID":"106c448a-8930-4be7-aa0e-970ab3d8b33e","Type":"ContainerStarted","Data":"8ffa55cc5a0ecee01dde67ff58189f72bcb9b29e97e75c66ab7db9e99bff8698"} Feb 18 09:56:28 crc kubenswrapper[4556]: I0218 09:56:28.327410 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-rnjw6/must-gather-vhc8n" podStartSLOduration=2.327386717 podStartE2EDuration="2.327386717s" podCreationTimestamp="2026-02-18 09:56:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:56:28.31820591 +0000 UTC m=+3145.335166890" watchObservedRunningTime="2026-02-18 09:56:28.327386717 +0000 UTC m=+3145.344347698" Feb 18 09:56:30 crc kubenswrapper[4556]: I0218 09:56:30.368906 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-rnjw6/crc-debug-xsgrc"] Feb 18 09:56:30 crc kubenswrapper[4556]: I0218 09:56:30.371621 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rnjw6/crc-debug-xsgrc" Feb 18 09:56:30 crc kubenswrapper[4556]: I0218 09:56:30.374299 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-rnjw6"/"default-dockercfg-8vkjz" Feb 18 09:56:30 crc kubenswrapper[4556]: I0218 09:56:30.485296 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1a25618c-9684-457f-9ea9-488ef38a0e23-host\") pod \"crc-debug-xsgrc\" (UID: \"1a25618c-9684-457f-9ea9-488ef38a0e23\") " pod="openshift-must-gather-rnjw6/crc-debug-xsgrc" Feb 18 09:56:30 crc kubenswrapper[4556]: I0218 09:56:30.485389 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvffm\" (UniqueName: \"kubernetes.io/projected/1a25618c-9684-457f-9ea9-488ef38a0e23-kube-api-access-hvffm\") pod \"crc-debug-xsgrc\" (UID: \"1a25618c-9684-457f-9ea9-488ef38a0e23\") " pod="openshift-must-gather-rnjw6/crc-debug-xsgrc" Feb 18 09:56:30 crc kubenswrapper[4556]: I0218 09:56:30.587399 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1a25618c-9684-457f-9ea9-488ef38a0e23-host\") pod \"crc-debug-xsgrc\" (UID: \"1a25618c-9684-457f-9ea9-488ef38a0e23\") " pod="openshift-must-gather-rnjw6/crc-debug-xsgrc" Feb 18 09:56:30 crc kubenswrapper[4556]: I0218 09:56:30.587458 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvffm\" (UniqueName: \"kubernetes.io/projected/1a25618c-9684-457f-9ea9-488ef38a0e23-kube-api-access-hvffm\") pod \"crc-debug-xsgrc\" (UID: \"1a25618c-9684-457f-9ea9-488ef38a0e23\") " pod="openshift-must-gather-rnjw6/crc-debug-xsgrc" Feb 18 09:56:30 crc kubenswrapper[4556]: I0218 09:56:30.587577 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1a25618c-9684-457f-9ea9-488ef38a0e23-host\") pod \"crc-debug-xsgrc\" (UID: \"1a25618c-9684-457f-9ea9-488ef38a0e23\") " pod="openshift-must-gather-rnjw6/crc-debug-xsgrc" Feb 18 09:56:30 crc kubenswrapper[4556]: I0218 09:56:30.607370 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvffm\" (UniqueName: \"kubernetes.io/projected/1a25618c-9684-457f-9ea9-488ef38a0e23-kube-api-access-hvffm\") pod \"crc-debug-xsgrc\" (UID: \"1a25618c-9684-457f-9ea9-488ef38a0e23\") " pod="openshift-must-gather-rnjw6/crc-debug-xsgrc" Feb 18 09:56:30 crc kubenswrapper[4556]: I0218 09:56:30.698175 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rnjw6/crc-debug-xsgrc" Feb 18 09:56:30 crc kubenswrapper[4556]: W0218 09:56:30.738806 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1a25618c_9684_457f_9ea9_488ef38a0e23.slice/crio-4c25900c00c2504f0ffc62a007821070d2cadd355d97f452d5ed9fd4f2a95b7a WatchSource:0}: Error finding container 4c25900c00c2504f0ffc62a007821070d2cadd355d97f452d5ed9fd4f2a95b7a: Status 404 returned error can't find the container with id 4c25900c00c2504f0ffc62a007821070d2cadd355d97f452d5ed9fd4f2a95b7a Feb 18 09:56:31 crc kubenswrapper[4556]: I0218 09:56:31.327137 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rnjw6/crc-debug-xsgrc" event={"ID":"1a25618c-9684-457f-9ea9-488ef38a0e23","Type":"ContainerStarted","Data":"6b05317fa9359cf529529b893f3580663d1cba39dddbed9ef83b550afce96760"} Feb 18 09:56:31 crc kubenswrapper[4556]: I0218 09:56:31.327638 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rnjw6/crc-debug-xsgrc" event={"ID":"1a25618c-9684-457f-9ea9-488ef38a0e23","Type":"ContainerStarted","Data":"4c25900c00c2504f0ffc62a007821070d2cadd355d97f452d5ed9fd4f2a95b7a"} Feb 18 09:56:31 crc kubenswrapper[4556]: I0218 09:56:31.352220 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-rnjw6/crc-debug-xsgrc" podStartSLOduration=1.352205345 podStartE2EDuration="1.352205345s" podCreationTimestamp="2026-02-18 09:56:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 09:56:31.344949718 +0000 UTC m=+3148.361910699" watchObservedRunningTime="2026-02-18 09:56:31.352205345 +0000 UTC m=+3148.369166324" Feb 18 09:56:42 crc kubenswrapper[4556]: I0218 09:56:42.283514 4556 scope.go:117] "RemoveContainer" containerID="9627d5bcc0cabda20a5fc92b3cc4b3dfae93225012caf89894f557937f69e6f7" Feb 18 09:56:42 crc kubenswrapper[4556]: E0218 09:56:42.285282 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:56:55 crc kubenswrapper[4556]: I0218 09:56:55.281950 4556 scope.go:117] "RemoveContainer" containerID="9627d5bcc0cabda20a5fc92b3cc4b3dfae93225012caf89894f557937f69e6f7" Feb 18 09:56:55 crc kubenswrapper[4556]: E0218 09:56:55.282758 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:56:56 crc kubenswrapper[4556]: I0218 09:56:56.550584 4556 generic.go:334] "Generic (PLEG): container finished" podID="1a25618c-9684-457f-9ea9-488ef38a0e23" containerID="6b05317fa9359cf529529b893f3580663d1cba39dddbed9ef83b550afce96760" exitCode=0 Feb 18 09:56:56 crc kubenswrapper[4556]: I0218 09:56:56.550658 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rnjw6/crc-debug-xsgrc" event={"ID":"1a25618c-9684-457f-9ea9-488ef38a0e23","Type":"ContainerDied","Data":"6b05317fa9359cf529529b893f3580663d1cba39dddbed9ef83b550afce96760"} Feb 18 09:56:57 crc kubenswrapper[4556]: I0218 09:56:57.644748 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rnjw6/crc-debug-xsgrc" Feb 18 09:56:57 crc kubenswrapper[4556]: I0218 09:56:57.679058 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-rnjw6/crc-debug-xsgrc"] Feb 18 09:56:57 crc kubenswrapper[4556]: I0218 09:56:57.688656 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-rnjw6/crc-debug-xsgrc"] Feb 18 09:56:57 crc kubenswrapper[4556]: I0218 09:56:57.799764 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1a25618c-9684-457f-9ea9-488ef38a0e23-host\") pod \"1a25618c-9684-457f-9ea9-488ef38a0e23\" (UID: \"1a25618c-9684-457f-9ea9-488ef38a0e23\") " Feb 18 09:56:57 crc kubenswrapper[4556]: I0218 09:56:57.799913 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1a25618c-9684-457f-9ea9-488ef38a0e23-host" (OuterVolumeSpecName: "host") pod "1a25618c-9684-457f-9ea9-488ef38a0e23" (UID: "1a25618c-9684-457f-9ea9-488ef38a0e23"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 09:56:57 crc kubenswrapper[4556]: I0218 09:56:57.800470 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hvffm\" (UniqueName: \"kubernetes.io/projected/1a25618c-9684-457f-9ea9-488ef38a0e23-kube-api-access-hvffm\") pod \"1a25618c-9684-457f-9ea9-488ef38a0e23\" (UID: \"1a25618c-9684-457f-9ea9-488ef38a0e23\") " Feb 18 09:56:57 crc kubenswrapper[4556]: I0218 09:56:57.801391 4556 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1a25618c-9684-457f-9ea9-488ef38a0e23-host\") on node \"crc\" DevicePath \"\"" Feb 18 09:56:57 crc kubenswrapper[4556]: I0218 09:56:57.806282 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a25618c-9684-457f-9ea9-488ef38a0e23-kube-api-access-hvffm" (OuterVolumeSpecName: "kube-api-access-hvffm") pod "1a25618c-9684-457f-9ea9-488ef38a0e23" (UID: "1a25618c-9684-457f-9ea9-488ef38a0e23"). InnerVolumeSpecName "kube-api-access-hvffm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:56:57 crc kubenswrapper[4556]: I0218 09:56:57.904297 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hvffm\" (UniqueName: \"kubernetes.io/projected/1a25618c-9684-457f-9ea9-488ef38a0e23-kube-api-access-hvffm\") on node \"crc\" DevicePath \"\"" Feb 18 09:56:58 crc kubenswrapper[4556]: I0218 09:56:58.565952 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4c25900c00c2504f0ffc62a007821070d2cadd355d97f452d5ed9fd4f2a95b7a" Feb 18 09:56:58 crc kubenswrapper[4556]: I0218 09:56:58.565992 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rnjw6/crc-debug-xsgrc" Feb 18 09:56:58 crc kubenswrapper[4556]: I0218 09:56:58.841048 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-rnjw6/crc-debug-zgs52"] Feb 18 09:56:58 crc kubenswrapper[4556]: E0218 09:56:58.841549 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a25618c-9684-457f-9ea9-488ef38a0e23" containerName="container-00" Feb 18 09:56:58 crc kubenswrapper[4556]: I0218 09:56:58.841569 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a25618c-9684-457f-9ea9-488ef38a0e23" containerName="container-00" Feb 18 09:56:58 crc kubenswrapper[4556]: I0218 09:56:58.841791 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a25618c-9684-457f-9ea9-488ef38a0e23" containerName="container-00" Feb 18 09:56:58 crc kubenswrapper[4556]: I0218 09:56:58.842433 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rnjw6/crc-debug-zgs52" Feb 18 09:56:58 crc kubenswrapper[4556]: I0218 09:56:58.844858 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-rnjw6"/"default-dockercfg-8vkjz" Feb 18 09:56:58 crc kubenswrapper[4556]: I0218 09:56:58.923936 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1031b8ce-db4e-4977-b1a4-7a8d384cde49-host\") pod \"crc-debug-zgs52\" (UID: \"1031b8ce-db4e-4977-b1a4-7a8d384cde49\") " pod="openshift-must-gather-rnjw6/crc-debug-zgs52" Feb 18 09:56:58 crc kubenswrapper[4556]: I0218 09:56:58.924171 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwm9m\" (UniqueName: \"kubernetes.io/projected/1031b8ce-db4e-4977-b1a4-7a8d384cde49-kube-api-access-xwm9m\") pod \"crc-debug-zgs52\" (UID: \"1031b8ce-db4e-4977-b1a4-7a8d384cde49\") " pod="openshift-must-gather-rnjw6/crc-debug-zgs52" Feb 18 09:56:59 crc kubenswrapper[4556]: I0218 09:56:59.026240 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1031b8ce-db4e-4977-b1a4-7a8d384cde49-host\") pod \"crc-debug-zgs52\" (UID: \"1031b8ce-db4e-4977-b1a4-7a8d384cde49\") " pod="openshift-must-gather-rnjw6/crc-debug-zgs52" Feb 18 09:56:59 crc kubenswrapper[4556]: I0218 09:56:59.026381 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1031b8ce-db4e-4977-b1a4-7a8d384cde49-host\") pod \"crc-debug-zgs52\" (UID: \"1031b8ce-db4e-4977-b1a4-7a8d384cde49\") " pod="openshift-must-gather-rnjw6/crc-debug-zgs52" Feb 18 09:56:59 crc kubenswrapper[4556]: I0218 09:56:59.026466 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwm9m\" (UniqueName: \"kubernetes.io/projected/1031b8ce-db4e-4977-b1a4-7a8d384cde49-kube-api-access-xwm9m\") pod \"crc-debug-zgs52\" (UID: \"1031b8ce-db4e-4977-b1a4-7a8d384cde49\") " pod="openshift-must-gather-rnjw6/crc-debug-zgs52" Feb 18 09:56:59 crc kubenswrapper[4556]: I0218 09:56:59.042681 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwm9m\" (UniqueName: \"kubernetes.io/projected/1031b8ce-db4e-4977-b1a4-7a8d384cde49-kube-api-access-xwm9m\") pod \"crc-debug-zgs52\" (UID: \"1031b8ce-db4e-4977-b1a4-7a8d384cde49\") " pod="openshift-must-gather-rnjw6/crc-debug-zgs52" Feb 18 09:56:59 crc kubenswrapper[4556]: I0218 09:56:59.156965 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rnjw6/crc-debug-zgs52" Feb 18 09:56:59 crc kubenswrapper[4556]: I0218 09:56:59.293734 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a25618c-9684-457f-9ea9-488ef38a0e23" path="/var/lib/kubelet/pods/1a25618c-9684-457f-9ea9-488ef38a0e23/volumes" Feb 18 09:56:59 crc kubenswrapper[4556]: I0218 09:56:59.577957 4556 generic.go:334] "Generic (PLEG): container finished" podID="1031b8ce-db4e-4977-b1a4-7a8d384cde49" containerID="9395a493f61f74e79c2333427ee8cdcbfc937ed5d87538a6575d0e8f262bf0c6" exitCode=0 Feb 18 09:56:59 crc kubenswrapper[4556]: I0218 09:56:59.578021 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rnjw6/crc-debug-zgs52" event={"ID":"1031b8ce-db4e-4977-b1a4-7a8d384cde49","Type":"ContainerDied","Data":"9395a493f61f74e79c2333427ee8cdcbfc937ed5d87538a6575d0e8f262bf0c6"} Feb 18 09:56:59 crc kubenswrapper[4556]: I0218 09:56:59.578071 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rnjw6/crc-debug-zgs52" event={"ID":"1031b8ce-db4e-4977-b1a4-7a8d384cde49","Type":"ContainerStarted","Data":"0353dd33e14e8f0b50bd2649a86394fe0bd71c28656f344f20cded9fb6d2a16c"} Feb 18 09:57:00 crc kubenswrapper[4556]: I0218 09:57:00.029502 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-rnjw6/crc-debug-zgs52"] Feb 18 09:57:00 crc kubenswrapper[4556]: I0218 09:57:00.035548 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-rnjw6/crc-debug-zgs52"] Feb 18 09:57:00 crc kubenswrapper[4556]: I0218 09:57:00.667389 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rnjw6/crc-debug-zgs52" Feb 18 09:57:00 crc kubenswrapper[4556]: I0218 09:57:00.866525 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xwm9m\" (UniqueName: \"kubernetes.io/projected/1031b8ce-db4e-4977-b1a4-7a8d384cde49-kube-api-access-xwm9m\") pod \"1031b8ce-db4e-4977-b1a4-7a8d384cde49\" (UID: \"1031b8ce-db4e-4977-b1a4-7a8d384cde49\") " Feb 18 09:57:00 crc kubenswrapper[4556]: I0218 09:57:00.866578 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1031b8ce-db4e-4977-b1a4-7a8d384cde49-host\") pod \"1031b8ce-db4e-4977-b1a4-7a8d384cde49\" (UID: \"1031b8ce-db4e-4977-b1a4-7a8d384cde49\") " Feb 18 09:57:00 crc kubenswrapper[4556]: I0218 09:57:00.867513 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1031b8ce-db4e-4977-b1a4-7a8d384cde49-host" (OuterVolumeSpecName: "host") pod "1031b8ce-db4e-4977-b1a4-7a8d384cde49" (UID: "1031b8ce-db4e-4977-b1a4-7a8d384cde49"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 09:57:00 crc kubenswrapper[4556]: I0218 09:57:00.872934 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1031b8ce-db4e-4977-b1a4-7a8d384cde49-kube-api-access-xwm9m" (OuterVolumeSpecName: "kube-api-access-xwm9m") pod "1031b8ce-db4e-4977-b1a4-7a8d384cde49" (UID: "1031b8ce-db4e-4977-b1a4-7a8d384cde49"). InnerVolumeSpecName "kube-api-access-xwm9m". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:57:00 crc kubenswrapper[4556]: I0218 09:57:00.970224 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xwm9m\" (UniqueName: \"kubernetes.io/projected/1031b8ce-db4e-4977-b1a4-7a8d384cde49-kube-api-access-xwm9m\") on node \"crc\" DevicePath \"\"" Feb 18 09:57:00 crc kubenswrapper[4556]: I0218 09:57:00.970264 4556 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1031b8ce-db4e-4977-b1a4-7a8d384cde49-host\") on node \"crc\" DevicePath \"\"" Feb 18 09:57:01 crc kubenswrapper[4556]: I0218 09:57:01.208495 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-rnjw6/crc-debug-lr92h"] Feb 18 09:57:01 crc kubenswrapper[4556]: E0218 09:57:01.208883 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1031b8ce-db4e-4977-b1a4-7a8d384cde49" containerName="container-00" Feb 18 09:57:01 crc kubenswrapper[4556]: I0218 09:57:01.208897 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="1031b8ce-db4e-4977-b1a4-7a8d384cde49" containerName="container-00" Feb 18 09:57:01 crc kubenswrapper[4556]: I0218 09:57:01.209098 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="1031b8ce-db4e-4977-b1a4-7a8d384cde49" containerName="container-00" Feb 18 09:57:01 crc kubenswrapper[4556]: I0218 09:57:01.209775 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rnjw6/crc-debug-lr92h" Feb 18 09:57:01 crc kubenswrapper[4556]: I0218 09:57:01.291574 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1031b8ce-db4e-4977-b1a4-7a8d384cde49" path="/var/lib/kubelet/pods/1031b8ce-db4e-4977-b1a4-7a8d384cde49/volumes" Feb 18 09:57:01 crc kubenswrapper[4556]: I0218 09:57:01.378941 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cj92j\" (UniqueName: \"kubernetes.io/projected/19a44573-13ff-4677-9c4d-ab5dc378fff1-kube-api-access-cj92j\") pod \"crc-debug-lr92h\" (UID: \"19a44573-13ff-4677-9c4d-ab5dc378fff1\") " pod="openshift-must-gather-rnjw6/crc-debug-lr92h" Feb 18 09:57:01 crc kubenswrapper[4556]: I0218 09:57:01.379093 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/19a44573-13ff-4677-9c4d-ab5dc378fff1-host\") pod \"crc-debug-lr92h\" (UID: \"19a44573-13ff-4677-9c4d-ab5dc378fff1\") " pod="openshift-must-gather-rnjw6/crc-debug-lr92h" Feb 18 09:57:01 crc kubenswrapper[4556]: I0218 09:57:01.481165 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cj92j\" (UniqueName: \"kubernetes.io/projected/19a44573-13ff-4677-9c4d-ab5dc378fff1-kube-api-access-cj92j\") pod \"crc-debug-lr92h\" (UID: \"19a44573-13ff-4677-9c4d-ab5dc378fff1\") " pod="openshift-must-gather-rnjw6/crc-debug-lr92h" Feb 18 09:57:01 crc kubenswrapper[4556]: I0218 09:57:01.481278 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/19a44573-13ff-4677-9c4d-ab5dc378fff1-host\") pod \"crc-debug-lr92h\" (UID: \"19a44573-13ff-4677-9c4d-ab5dc378fff1\") " pod="openshift-must-gather-rnjw6/crc-debug-lr92h" Feb 18 09:57:01 crc kubenswrapper[4556]: I0218 09:57:01.481711 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/19a44573-13ff-4677-9c4d-ab5dc378fff1-host\") pod \"crc-debug-lr92h\" (UID: \"19a44573-13ff-4677-9c4d-ab5dc378fff1\") " pod="openshift-must-gather-rnjw6/crc-debug-lr92h" Feb 18 09:57:01 crc kubenswrapper[4556]: I0218 09:57:01.496989 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cj92j\" (UniqueName: \"kubernetes.io/projected/19a44573-13ff-4677-9c4d-ab5dc378fff1-kube-api-access-cj92j\") pod \"crc-debug-lr92h\" (UID: \"19a44573-13ff-4677-9c4d-ab5dc378fff1\") " pod="openshift-must-gather-rnjw6/crc-debug-lr92h" Feb 18 09:57:01 crc kubenswrapper[4556]: I0218 09:57:01.525459 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rnjw6/crc-debug-lr92h" Feb 18 09:57:01 crc kubenswrapper[4556]: W0218 09:57:01.553058 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod19a44573_13ff_4677_9c4d_ab5dc378fff1.slice/crio-6bf4704b36e16e58074a8548027d4b7bba330cb503a9b4e099ce3c2665afba3c WatchSource:0}: Error finding container 6bf4704b36e16e58074a8548027d4b7bba330cb503a9b4e099ce3c2665afba3c: Status 404 returned error can't find the container with id 6bf4704b36e16e58074a8548027d4b7bba330cb503a9b4e099ce3c2665afba3c Feb 18 09:57:01 crc kubenswrapper[4556]: I0218 09:57:01.596442 4556 scope.go:117] "RemoveContainer" containerID="9395a493f61f74e79c2333427ee8cdcbfc937ed5d87538a6575d0e8f262bf0c6" Feb 18 09:57:01 crc kubenswrapper[4556]: I0218 09:57:01.596574 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rnjw6/crc-debug-zgs52" Feb 18 09:57:01 crc kubenswrapper[4556]: I0218 09:57:01.598656 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rnjw6/crc-debug-lr92h" event={"ID":"19a44573-13ff-4677-9c4d-ab5dc378fff1","Type":"ContainerStarted","Data":"6bf4704b36e16e58074a8548027d4b7bba330cb503a9b4e099ce3c2665afba3c"} Feb 18 09:57:02 crc kubenswrapper[4556]: I0218 09:57:02.614451 4556 generic.go:334] "Generic (PLEG): container finished" podID="19a44573-13ff-4677-9c4d-ab5dc378fff1" containerID="1c07e5dff6e5e14d86849919cf277d9d8cdb0c55b249723bc4263db4cf5cb82f" exitCode=0 Feb 18 09:57:02 crc kubenswrapper[4556]: I0218 09:57:02.614627 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rnjw6/crc-debug-lr92h" event={"ID":"19a44573-13ff-4677-9c4d-ab5dc378fff1","Type":"ContainerDied","Data":"1c07e5dff6e5e14d86849919cf277d9d8cdb0c55b249723bc4263db4cf5cb82f"} Feb 18 09:57:02 crc kubenswrapper[4556]: I0218 09:57:02.657649 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-rnjw6/crc-debug-lr92h"] Feb 18 09:57:02 crc kubenswrapper[4556]: I0218 09:57:02.667576 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-rnjw6/crc-debug-lr92h"] Feb 18 09:57:03 crc kubenswrapper[4556]: I0218 09:57:03.706779 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rnjw6/crc-debug-lr92h" Feb 18 09:57:03 crc kubenswrapper[4556]: I0218 09:57:03.829370 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/19a44573-13ff-4677-9c4d-ab5dc378fff1-host\") pod \"19a44573-13ff-4677-9c4d-ab5dc378fff1\" (UID: \"19a44573-13ff-4677-9c4d-ab5dc378fff1\") " Feb 18 09:57:03 crc kubenswrapper[4556]: I0218 09:57:03.829494 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/19a44573-13ff-4677-9c4d-ab5dc378fff1-host" (OuterVolumeSpecName: "host") pod "19a44573-13ff-4677-9c4d-ab5dc378fff1" (UID: "19a44573-13ff-4677-9c4d-ab5dc378fff1"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 09:57:03 crc kubenswrapper[4556]: I0218 09:57:03.829515 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cj92j\" (UniqueName: \"kubernetes.io/projected/19a44573-13ff-4677-9c4d-ab5dc378fff1-kube-api-access-cj92j\") pod \"19a44573-13ff-4677-9c4d-ab5dc378fff1\" (UID: \"19a44573-13ff-4677-9c4d-ab5dc378fff1\") " Feb 18 09:57:03 crc kubenswrapper[4556]: I0218 09:57:03.830065 4556 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/19a44573-13ff-4677-9c4d-ab5dc378fff1-host\") on node \"crc\" DevicePath \"\"" Feb 18 09:57:03 crc kubenswrapper[4556]: I0218 09:57:03.835070 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19a44573-13ff-4677-9c4d-ab5dc378fff1-kube-api-access-cj92j" (OuterVolumeSpecName: "kube-api-access-cj92j") pod "19a44573-13ff-4677-9c4d-ab5dc378fff1" (UID: "19a44573-13ff-4677-9c4d-ab5dc378fff1"). InnerVolumeSpecName "kube-api-access-cj92j". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:57:03 crc kubenswrapper[4556]: I0218 09:57:03.932708 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cj92j\" (UniqueName: \"kubernetes.io/projected/19a44573-13ff-4677-9c4d-ab5dc378fff1-kube-api-access-cj92j\") on node \"crc\" DevicePath \"\"" Feb 18 09:57:04 crc kubenswrapper[4556]: I0218 09:57:04.634519 4556 scope.go:117] "RemoveContainer" containerID="1c07e5dff6e5e14d86849919cf277d9d8cdb0c55b249723bc4263db4cf5cb82f" Feb 18 09:57:04 crc kubenswrapper[4556]: I0218 09:57:04.634600 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rnjw6/crc-debug-lr92h" Feb 18 09:57:05 crc kubenswrapper[4556]: I0218 09:57:05.291316 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19a44573-13ff-4677-9c4d-ab5dc378fff1" path="/var/lib/kubelet/pods/19a44573-13ff-4677-9c4d-ab5dc378fff1/volumes" Feb 18 09:57:07 crc kubenswrapper[4556]: I0218 09:57:07.282472 4556 scope.go:117] "RemoveContainer" containerID="9627d5bcc0cabda20a5fc92b3cc4b3dfae93225012caf89894f557937f69e6f7" Feb 18 09:57:07 crc kubenswrapper[4556]: E0218 09:57:07.283019 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:57:22 crc kubenswrapper[4556]: I0218 09:57:22.284231 4556 scope.go:117] "RemoveContainer" containerID="9627d5bcc0cabda20a5fc92b3cc4b3dfae93225012caf89894f557937f69e6f7" Feb 18 09:57:22 crc kubenswrapper[4556]: E0218 09:57:22.285092 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 09:57:29 crc kubenswrapper[4556]: I0218 09:57:29.524683 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-845cb5d46d-z5vhw_ff67ebd8-094e-4c5b-b164-d6b37fc169de/barbican-api/0.log" Feb 18 09:57:29 crc kubenswrapper[4556]: I0218 09:57:29.613489 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-845cb5d46d-z5vhw_ff67ebd8-094e-4c5b-b164-d6b37fc169de/barbican-api-log/0.log" Feb 18 09:57:29 crc kubenswrapper[4556]: I0218 09:57:29.692430 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5fc745f7dd-68srz_8be9c112-7903-4ec0-a503-fb16c676caf4/barbican-keystone-listener/0.log" Feb 18 09:57:29 crc kubenswrapper[4556]: I0218 09:57:29.740770 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5fc745f7dd-68srz_8be9c112-7903-4ec0-a503-fb16c676caf4/barbican-keystone-listener-log/0.log" Feb 18 09:57:29 crc kubenswrapper[4556]: I0218 09:57:29.868931 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-b64d8f579-7sj9h_d6ef48e2-d690-4060-b013-5ac2be288161/barbican-worker-log/0.log" Feb 18 09:57:29 crc kubenswrapper[4556]: I0218 09:57:29.892428 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-b64d8f579-7sj9h_d6ef48e2-d690-4060-b013-5ac2be288161/barbican-worker/0.log" Feb 18 09:57:29 crc kubenswrapper[4556]: I0218 09:57:29.994302 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-pqnqh_91161557-e383-4bba-81d8-d3b0ba3b6840/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Feb 18 09:57:30 crc kubenswrapper[4556]: I0218 09:57:30.043009 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_d105f935-cc14-4293-8101-17adaef31666/ceilometer-central-agent/0.log" Feb 18 09:57:30 crc kubenswrapper[4556]: I0218 09:57:30.101293 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_d105f935-cc14-4293-8101-17adaef31666/ceilometer-notification-agent/0.log" Feb 18 09:57:30 crc kubenswrapper[4556]: I0218 09:57:30.160472 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_d105f935-cc14-4293-8101-17adaef31666/proxy-httpd/0.log" Feb 18 09:57:30 crc kubenswrapper[4556]: I0218 09:57:30.174059 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_d105f935-cc14-4293-8101-17adaef31666/sg-core/0.log" Feb 18 09:57:30 crc kubenswrapper[4556]: I0218 09:57:30.288401 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d/cinder-api/0.log" Feb 18 09:57:30 crc kubenswrapper[4556]: I0218 09:57:30.307320 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_e9f6ea4f-0d2a-4e91-b23c-301bf7ba2a4d/cinder-api-log/0.log" Feb 18 09:57:30 crc kubenswrapper[4556]: I0218 09:57:30.468018 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_ca98a45d-9b95-4e00-8179-feb5a6c1ddb0/cinder-scheduler/0.log" Feb 18 09:57:30 crc kubenswrapper[4556]: I0218 09:57:30.497881 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_ca98a45d-9b95-4e00-8179-feb5a6c1ddb0/probe/0.log" Feb 18 09:57:30 crc kubenswrapper[4556]: I0218 09:57:30.529300 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-snzk4_4fdd7d3d-c538-4cb6-9d45-b94723b744b1/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Feb 18 09:57:30 crc kubenswrapper[4556]: I0218 09:57:30.661329 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-nfwmp_c71f2826-1f14-462a-8096-cace6bd934d3/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 18 09:57:30 crc kubenswrapper[4556]: I0218 09:57:30.721375 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-56455489f5-9pk6d_d6426d5e-fb37-42e3-8392-3db7d31e81e7/init/0.log" Feb 18 09:57:30 crc kubenswrapper[4556]: I0218 09:57:30.854773 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-56455489f5-9pk6d_d6426d5e-fb37-42e3-8392-3db7d31e81e7/init/0.log" Feb 18 09:57:30 crc kubenswrapper[4556]: I0218 09:57:30.932411 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-qdtn4_a33e8704-dcbf-43ff-b665-cff323679c76/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Feb 18 09:57:30 crc kubenswrapper[4556]: I0218 09:57:30.957206 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-56455489f5-9pk6d_d6426d5e-fb37-42e3-8392-3db7d31e81e7/dnsmasq-dns/0.log" Feb 18 09:57:31 crc kubenswrapper[4556]: I0218 09:57:31.103072 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_23a483a0-fda5-4bf6-bfea-06a84d8d6533/glance-httpd/0.log" Feb 18 09:57:31 crc kubenswrapper[4556]: I0218 09:57:31.117873 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_23a483a0-fda5-4bf6-bfea-06a84d8d6533/glance-log/0.log" Feb 18 09:57:31 crc kubenswrapper[4556]: I0218 09:57:31.244329 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_f50b4c03-2bce-417a-b01c-aaec09a5c8c3/glance-httpd/0.log" Feb 18 09:57:31 crc kubenswrapper[4556]: I0218 09:57:31.258418 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_f50b4c03-2bce-417a-b01c-aaec09a5c8c3/glance-log/0.log" Feb 18 09:57:31 crc kubenswrapper[4556]: I0218 09:57:31.446464 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-595566bb8b-6b5h6_30fffa0b-72a5-4bbe-a1fb-fa8b26d1decf/horizon/0.log" Feb 18 09:57:31 crc kubenswrapper[4556]: I0218 09:57:31.515126 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-t9trr_d249a625-7b1c-4a2f-a92d-c18d7fbb6142/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Feb 18 09:57:31 crc kubenswrapper[4556]: I0218 09:57:31.725705 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-zmv42_d1d95573-7bea-4ee6-b0c1-13ec18b10244/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 18 09:57:31 crc kubenswrapper[4556]: I0218 09:57:31.762344 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-595566bb8b-6b5h6_30fffa0b-72a5-4bbe-a1fb-fa8b26d1decf/horizon-log/0.log" Feb 18 09:57:31 crc kubenswrapper[4556]: I0218 09:57:31.985590 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_e54e3af3-e2c2-4e5b-885d-0071d05acbb4/kube-state-metrics/0.log" Feb 18 09:57:32 crc kubenswrapper[4556]: I0218 09:57:32.020495 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-66f6978747-npdn5_6b87f75f-f2e8-4a26-b59c-1d0bbe9c777d/keystone-api/0.log" Feb 18 09:57:32 crc kubenswrapper[4556]: I0218 09:57:32.156395 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-hhn8t_2cff0286-97e0-422e-b55c-4f28711bde2d/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Feb 18 09:57:32 crc kubenswrapper[4556]: I0218 09:57:32.445619 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-54664bdbc-cxnvs_57873d95-dd1a-4b5c-99b6-459774c90acc/neutron-httpd/0.log" Feb 18 09:57:32 crc kubenswrapper[4556]: I0218 09:57:32.471164 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-54664bdbc-cxnvs_57873d95-dd1a-4b5c-99b6-459774c90acc/neutron-api/0.log" Feb 18 09:57:32 crc kubenswrapper[4556]: I0218 09:57:32.643262 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-wcldl_3343de49-153e-43bb-9d22-703a5f3a0b0b/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Feb 18 09:57:33 crc kubenswrapper[4556]: I0218 09:57:33.064101 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_4e997362-15c5-4052-bf0b-8491c057f317/nova-api-log/0.log" Feb 18 09:57:33 crc kubenswrapper[4556]: I0218 09:57:33.156981 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_64bcfb93-2db3-4350-97ee-671806ff7d00/nova-cell0-conductor-conductor/0.log" Feb 18 09:57:33 crc kubenswrapper[4556]: I0218 09:57:33.383602 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_4e997362-15c5-4052-bf0b-8491c057f317/nova-api-api/0.log" Feb 18 09:57:33 crc kubenswrapper[4556]: I0218 09:57:33.410575 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_f0149238-212d-489d-8c94-7b0a8d86b1f5/nova-cell1-conductor-conductor/0.log" Feb 18 09:57:33 crc kubenswrapper[4556]: I0218 09:57:33.647436 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_c1c20375-9fcb-4886-a91c-3dd36651532a/nova-cell1-novncproxy-novncproxy/0.log" Feb 18 09:57:33 crc kubenswrapper[4556]: I0218 09:57:33.677535 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-snrkh_a1ab2e8c-4abd-4421-ac68-20821b8ba938/nova-edpm-deployment-openstack-edpm-ipam/0.log" Feb 18 09:57:33 crc kubenswrapper[4556]: I0218 09:57:33.904055 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_4424b924-1dd6-4323-863d-53bf1a0f5fc9/nova-metadata-log/0.log" Feb 18 09:57:34 crc kubenswrapper[4556]: I0218 09:57:34.059974 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_11b77df6-67d9-4ddb-944b-2ad8b0fada78/mysql-bootstrap/0.log" Feb 18 09:57:34 crc kubenswrapper[4556]: I0218 09:57:34.078417 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_db4d9035-57ee-4c5f-9677-cc2174644152/nova-scheduler-scheduler/0.log" Feb 18 09:57:34 crc kubenswrapper[4556]: I0218 09:57:34.255483 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_11b77df6-67d9-4ddb-944b-2ad8b0fada78/mysql-bootstrap/0.log" Feb 18 09:57:34 crc kubenswrapper[4556]: I0218 09:57:34.283208 4556 scope.go:117] "RemoveContainer" containerID="9627d5bcc0cabda20a5fc92b3cc4b3dfae93225012caf89894f557937f69e6f7" Feb 18 09:57:34 crc kubenswrapper[4556]: I0218 09:57:34.306434 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_11b77df6-67d9-4ddb-944b-2ad8b0fada78/galera/0.log" Feb 18 09:57:34 crc kubenswrapper[4556]: I0218 09:57:34.462221 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_2c504c68-45e0-49dc-90f3-90b91e33551c/mysql-bootstrap/0.log" Feb 18 09:57:34 crc kubenswrapper[4556]: I0218 09:57:34.668410 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_2c504c68-45e0-49dc-90f3-90b91e33551c/mysql-bootstrap/0.log" Feb 18 09:57:34 crc kubenswrapper[4556]: I0218 09:57:34.687945 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_2c504c68-45e0-49dc-90f3-90b91e33551c/galera/0.log" Feb 18 09:57:34 crc kubenswrapper[4556]: I0218 09:57:34.865447 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_97268f30-e67a-441f-93cd-0dc9771dbd42/openstackclient/0.log" Feb 18 09:57:34 crc kubenswrapper[4556]: I0218 09:57:34.894319 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" event={"ID":"8dac7f27-d3d1-4778-9e54-f273035a1d37","Type":"ContainerStarted","Data":"34f3a3491bd005389624fe9685306e5b96d58407e240c9903fadf6abf52b3afd"} Feb 18 09:57:34 crc kubenswrapper[4556]: I0218 09:57:34.968876 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_4424b924-1dd6-4323-863d-53bf1a0f5fc9/nova-metadata-metadata/0.log" Feb 18 09:57:34 crc kubenswrapper[4556]: I0218 09:57:34.984920 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-6kkrw_4906afef-c7e6-4597-a4a5-e9d758917e11/ovn-controller/0.log" Feb 18 09:57:35 crc kubenswrapper[4556]: I0218 09:57:35.113316 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-j7rpf_66575dd2-efe2-4770-a9cd-7afb99e8566c/openstack-network-exporter/0.log" Feb 18 09:57:35 crc kubenswrapper[4556]: I0218 09:57:35.162168 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-srnvn_40c51577-de87-4bd8-be03-41221bc2f415/ovsdb-server-init/0.log" Feb 18 09:57:35 crc kubenswrapper[4556]: I0218 09:57:35.374740 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-srnvn_40c51577-de87-4bd8-be03-41221bc2f415/ovsdb-server-init/0.log" Feb 18 09:57:35 crc kubenswrapper[4556]: I0218 09:57:35.375830 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-srnvn_40c51577-de87-4bd8-be03-41221bc2f415/ovs-vswitchd/0.log" Feb 18 09:57:35 crc kubenswrapper[4556]: I0218 09:57:35.383288 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-srnvn_40c51577-de87-4bd8-be03-41221bc2f415/ovsdb-server/0.log" Feb 18 09:57:35 crc kubenswrapper[4556]: I0218 09:57:35.561993 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-vgw62_7c63f939-1090-4fae-b45b-9d55f09f489d/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Feb 18 09:57:35 crc kubenswrapper[4556]: I0218 09:57:35.581581 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_ef64adb0-c929-4a3f-8aa6-d490b45ba5e1/openstack-network-exporter/0.log" Feb 18 09:57:35 crc kubenswrapper[4556]: I0218 09:57:35.647982 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_ef64adb0-c929-4a3f-8aa6-d490b45ba5e1/ovn-northd/0.log" Feb 18 09:57:35 crc kubenswrapper[4556]: I0218 09:57:35.756095 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_10b5d4cb-2beb-405d-83fc-30eb5fd4aaad/openstack-network-exporter/0.log" Feb 18 09:57:35 crc kubenswrapper[4556]: I0218 09:57:35.809798 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_10b5d4cb-2beb-405d-83fc-30eb5fd4aaad/ovsdbserver-nb/0.log" Feb 18 09:57:35 crc kubenswrapper[4556]: I0218 09:57:35.945352 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_d7840643-68ec-4a2a-b6af-8a9730729077/openstack-network-exporter/0.log" Feb 18 09:57:35 crc kubenswrapper[4556]: I0218 09:57:35.968693 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_d7840643-68ec-4a2a-b6af-8a9730729077/ovsdbserver-sb/0.log" Feb 18 09:57:36 crc kubenswrapper[4556]: I0218 09:57:36.186270 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-9855c4778-9n9f2_ac2fe5b2-09b6-440f-83db-8555ec304a27/placement-api/0.log" Feb 18 09:57:36 crc kubenswrapper[4556]: I0218 09:57:36.210612 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-9855c4778-9n9f2_ac2fe5b2-09b6-440f-83db-8555ec304a27/placement-log/0.log" Feb 18 09:57:36 crc kubenswrapper[4556]: I0218 09:57:36.294360 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_8b1c0c56-94a0-4ab7-ae4c-5f4035e37359/setup-container/0.log" Feb 18 09:57:36 crc kubenswrapper[4556]: I0218 09:57:36.498257 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_8b1c0c56-94a0-4ab7-ae4c-5f4035e37359/setup-container/0.log" Feb 18 09:57:36 crc kubenswrapper[4556]: I0218 09:57:36.516743 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_8b1c0c56-94a0-4ab7-ae4c-5f4035e37359/rabbitmq/0.log" Feb 18 09:57:36 crc kubenswrapper[4556]: I0218 09:57:36.553226 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_0f0ef5d2-a692-435f-a79d-a1af3a294e73/setup-container/0.log" Feb 18 09:57:36 crc kubenswrapper[4556]: I0218 09:57:36.699811 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_0f0ef5d2-a692-435f-a79d-a1af3a294e73/rabbitmq/0.log" Feb 18 09:57:36 crc kubenswrapper[4556]: I0218 09:57:36.710329 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_0f0ef5d2-a692-435f-a79d-a1af3a294e73/setup-container/0.log" Feb 18 09:57:36 crc kubenswrapper[4556]: I0218 09:57:36.728452 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-jmgv7_d0b159c4-dbac-4ddb-af28-523bb86639e5/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 18 09:57:36 crc kubenswrapper[4556]: I0218 09:57:36.898305 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-hjswl_3ebf502b-e4a2-4d25-8551-f1deed31b5c9/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Feb 18 09:57:36 crc kubenswrapper[4556]: I0218 09:57:36.935181 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-sqqbc_c2b34eeb-f211-4454-a74d-207de2fa9f13/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Feb 18 09:57:37 crc kubenswrapper[4556]: I0218 09:57:37.116692 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-2j6nf_1d8fa453-3277-4c76-9098-0e7838fd8d44/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 18 09:57:37 crc kubenswrapper[4556]: I0218 09:57:37.165661 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-92c7r_11adbf49-97c4-4878-96c0-4fa453f7a819/ssh-known-hosts-edpm-deployment/0.log" Feb 18 09:57:37 crc kubenswrapper[4556]: I0218 09:57:37.364479 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-864fbf8dcf-chrjx_d678be18-d610-4ea8-b248-47843cf74ea3/proxy-server/0.log" Feb 18 09:57:37 crc kubenswrapper[4556]: I0218 09:57:37.454195 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-864fbf8dcf-chrjx_d678be18-d610-4ea8-b248-47843cf74ea3/proxy-httpd/0.log" Feb 18 09:57:37 crc kubenswrapper[4556]: I0218 09:57:37.565582 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-9z5cc_03ccffdf-f11c-4788-9e22-0b0661665c5d/swift-ring-rebalance/0.log" Feb 18 09:57:37 crc kubenswrapper[4556]: I0218 09:57:37.610045 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_da163339-653b-4057-9c2f-332eb6957f40/account-auditor/0.log" Feb 18 09:57:37 crc kubenswrapper[4556]: I0218 09:57:37.673244 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_da163339-653b-4057-9c2f-332eb6957f40/account-reaper/0.log" Feb 18 09:57:37 crc kubenswrapper[4556]: I0218 09:57:37.772060 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_da163339-653b-4057-9c2f-332eb6957f40/account-replicator/0.log" Feb 18 09:57:37 crc kubenswrapper[4556]: I0218 09:57:37.778038 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_da163339-653b-4057-9c2f-332eb6957f40/account-server/0.log" Feb 18 09:57:37 crc kubenswrapper[4556]: I0218 09:57:37.791847 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_da163339-653b-4057-9c2f-332eb6957f40/container-auditor/0.log" Feb 18 09:57:37 crc kubenswrapper[4556]: I0218 09:57:37.933262 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_da163339-653b-4057-9c2f-332eb6957f40/container-replicator/0.log" Feb 18 09:57:37 crc kubenswrapper[4556]: I0218 09:57:37.953109 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_da163339-653b-4057-9c2f-332eb6957f40/container-server/0.log" Feb 18 09:57:37 crc kubenswrapper[4556]: I0218 09:57:37.970341 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_da163339-653b-4057-9c2f-332eb6957f40/container-updater/0.log" Feb 18 09:57:38 crc kubenswrapper[4556]: I0218 09:57:38.011239 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_da163339-653b-4057-9c2f-332eb6957f40/object-auditor/0.log" Feb 18 09:57:38 crc kubenswrapper[4556]: I0218 09:57:38.138519 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_da163339-653b-4057-9c2f-332eb6957f40/object-server/0.log" Feb 18 09:57:38 crc kubenswrapper[4556]: I0218 09:57:38.180438 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_da163339-653b-4057-9c2f-332eb6957f40/object-replicator/0.log" Feb 18 09:57:38 crc kubenswrapper[4556]: I0218 09:57:38.207463 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_da163339-653b-4057-9c2f-332eb6957f40/object-expirer/0.log" Feb 18 09:57:38 crc kubenswrapper[4556]: I0218 09:57:38.210221 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_da163339-653b-4057-9c2f-332eb6957f40/object-updater/0.log" Feb 18 09:57:38 crc kubenswrapper[4556]: I0218 09:57:38.322194 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_da163339-653b-4057-9c2f-332eb6957f40/rsync/0.log" Feb 18 09:57:38 crc kubenswrapper[4556]: I0218 09:57:38.381569 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_da163339-653b-4057-9c2f-332eb6957f40/swift-recon-cron/0.log" Feb 18 09:57:38 crc kubenswrapper[4556]: I0218 09:57:38.466395 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-l92vs_8ea53a2a-db68-43fb-98c4-26b9ca9f816d/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Feb 18 09:57:38 crc kubenswrapper[4556]: I0218 09:57:38.575328 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_7fa2fae4-d6f4-4e8b-98f6-34f8ccd47b59/tempest-tests-tempest-tests-runner/0.log" Feb 18 09:57:38 crc kubenswrapper[4556]: I0218 09:57:38.729768 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_396883e0-c2e8-4698-97d2-70bbd5fe7b7f/test-operator-logs-container/0.log" Feb 18 09:57:38 crc kubenswrapper[4556]: I0218 09:57:38.763134 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-6jjjt_a5a47273-47da-41d2-a98f-dc1fba7e1102/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Feb 18 09:57:48 crc kubenswrapper[4556]: I0218 09:57:48.068757 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_68274946-b189-450c-a154-27059f411af3/memcached/0.log" Feb 18 09:58:01 crc kubenswrapper[4556]: I0218 09:58:01.357456 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967fk785_b828da70-b43b-421e-a2b2-d1a2bd562ae4/util/0.log" Feb 18 09:58:01 crc kubenswrapper[4556]: I0218 09:58:01.546698 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967fk785_b828da70-b43b-421e-a2b2-d1a2bd562ae4/util/0.log" Feb 18 09:58:01 crc kubenswrapper[4556]: I0218 09:58:01.573490 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967fk785_b828da70-b43b-421e-a2b2-d1a2bd562ae4/pull/0.log" Feb 18 09:58:01 crc kubenswrapper[4556]: I0218 09:58:01.583007 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967fk785_b828da70-b43b-421e-a2b2-d1a2bd562ae4/pull/0.log" Feb 18 09:58:01 crc kubenswrapper[4556]: I0218 09:58:01.714038 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967fk785_b828da70-b43b-421e-a2b2-d1a2bd562ae4/util/0.log" Feb 18 09:58:01 crc kubenswrapper[4556]: I0218 09:58:01.719512 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967fk785_b828da70-b43b-421e-a2b2-d1a2bd562ae4/pull/0.log" Feb 18 09:58:01 crc kubenswrapper[4556]: I0218 09:58:01.762346 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967fk785_b828da70-b43b-421e-a2b2-d1a2bd562ae4/extract/0.log" Feb 18 09:58:02 crc kubenswrapper[4556]: I0218 09:58:02.087266 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-6d8bf5c495-5pv48_fff15cfd-9701-4efa-81a6-9a482e09ca0f/manager/0.log" Feb 18 09:58:02 crc kubenswrapper[4556]: I0218 09:58:02.416225 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987464f4-mfhjc_f194b6f2-c143-4a57-8d3e-378a08147713/manager/0.log" Feb 18 09:58:02 crc kubenswrapper[4556]: I0218 09:58:02.532974 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-69f49c598c-6sck2_a980ea11-bf5f-4967-b22b-3e63454bb9ae/manager/0.log" Feb 18 09:58:02 crc kubenswrapper[4556]: I0218 09:58:02.729760 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5b9b8895d5-g6hvq_6a219296-13f5-4c55-99ec-35e34d43a341/manager/0.log" Feb 18 09:58:03 crc kubenswrapper[4556]: I0218 09:58:03.172250 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-554564d7fc-lktrc_4fdde554-179d-4128-a260-38c3de6e9d64/manager/0.log" Feb 18 09:58:03 crc kubenswrapper[4556]: I0218 09:58:03.190322 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-5d946d989d-n4x7d_463e8b44-a344-401e-bd53-22738560ca32/manager/0.log" Feb 18 09:58:03 crc kubenswrapper[4556]: I0218 09:58:03.331033 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-79d975b745-gp4rt_853bc90c-e14b-46ea-912d-5aaefa5f908a/manager/0.log" Feb 18 09:58:03 crc kubenswrapper[4556]: I0218 09:58:03.481532 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-b4d948c87-9rkqc_cd593415-1b63-4965-9865-21a1ae2e4742/manager/0.log" Feb 18 09:58:03 crc kubenswrapper[4556]: I0218 09:58:03.795010 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-54f6768c69-r5kw5_6e7470ce-ab97-4534-84e8-e951071eb6ee/manager/0.log" Feb 18 09:58:03 crc kubenswrapper[4556]: I0218 09:58:03.946926 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6994f66f48-t2ng7_fdd8c829-9082-40ae-8b65-46a8bd293c4d/manager/0.log" Feb 18 09:58:04 crc kubenswrapper[4556]: I0218 09:58:04.088321 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-64ddbf8bb-wdp2t_9d670c80-f1d0-4b08-b83b-086977d49380/manager/0.log" Feb 18 09:58:04 crc kubenswrapper[4556]: I0218 09:58:04.335725 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-567668f5cf-t9gtc_d672d582-25f7-4654-bc52-5f07631f5e9e/manager/0.log" Feb 18 09:58:04 crc kubenswrapper[4556]: I0218 09:58:04.519847 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-fb5fcc5b8-9mlxh_48ca8d44-b660-43ee-8a9e-6489da3cee3d/manager/0.log" Feb 18 09:58:05 crc kubenswrapper[4556]: I0218 09:58:05.042283 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-6679bf9b57-r8zc2_acd47be1-151c-4fd9-9cc6-f49e1cc3f403/operator/0.log" Feb 18 09:58:05 crc kubenswrapper[4556]: I0218 09:58:05.268310 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-tg2z4_85052b3f-274e-437c-af72-fdd608552f7c/registry-server/0.log" Feb 18 09:58:05 crc kubenswrapper[4556]: I0218 09:58:05.552797 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-d44cf6b75-qr5np_2c577857-1c7f-4d16-9c13-d3d83eddb3de/manager/0.log" Feb 18 09:58:05 crc kubenswrapper[4556]: I0218 09:58:05.741207 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-8497b45c89-5vgnl_f2bd3ad9-919a-40ec-b111-f9655a6650a6/manager/0.log" Feb 18 09:58:05 crc kubenswrapper[4556]: I0218 09:58:05.944713 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-vt6jv_28452a75-3b84-4add-9952-c5f0193e0954/operator/0.log" Feb 18 09:58:05 crc kubenswrapper[4556]: I0218 09:58:05.993310 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-69f8888797-rpk9w_96be73f9-04a8-4ec6-9b75-ae0dd4ae84c8/manager/0.log" Feb 18 09:58:06 crc kubenswrapper[4556]: I0218 09:58:06.170332 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-68f46476f-jz577_26bd53bd-69fd-4b14-afd5-5ffb7e557346/manager/0.log" Feb 18 09:58:06 crc kubenswrapper[4556]: I0218 09:58:06.272604 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-7f45b4ff68-4f5nl_acad90fd-fc19-4b74-bb00-192d0e6061a6/manager/0.log" Feb 18 09:58:06 crc kubenswrapper[4556]: I0218 09:58:06.314383 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-7866795846-9bsld_bc370f09-39b1-4e31-bbc6-7756c63f0c30/manager/0.log" Feb 18 09:58:06 crc kubenswrapper[4556]: I0218 09:58:06.524579 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-5db88f68c-98pqn_8e02c422-7a17-4d4f-9ac1-7fe86c5dd472/manager/0.log" Feb 18 09:58:06 crc kubenswrapper[4556]: I0218 09:58:06.607004 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-69ff7bc449-jjnjd_4fb1fa58-065b-4412-99e1-e9d77cdf4b41/manager/0.log" Feb 18 09:58:08 crc kubenswrapper[4556]: I0218 09:58:08.210267 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-868647ff47-x74c8_9cf84ffa-7a16-4e6a-aa7a-c90d6c1635fe/manager/0.log" Feb 18 09:58:24 crc kubenswrapper[4556]: I0218 09:58:24.208230 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-j5fdh_b3ca6b25-cb73-47a9-867c-8d1f6f628077/control-plane-machine-set-operator/0.log" Feb 18 09:58:24 crc kubenswrapper[4556]: I0218 09:58:24.363829 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-zhtqs_b5f35d95-f6b7-4068-8036-dffcbf955272/machine-api-operator/0.log" Feb 18 09:58:24 crc kubenswrapper[4556]: I0218 09:58:24.369703 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-zhtqs_b5f35d95-f6b7-4068-8036-dffcbf955272/kube-rbac-proxy/0.log" Feb 18 09:58:36 crc kubenswrapper[4556]: I0218 09:58:36.008559 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-bw685_257b9c79-fc7b-4bc5-88eb-d767e4762f77/cert-manager-controller/0.log" Feb 18 09:58:36 crc kubenswrapper[4556]: I0218 09:58:36.149136 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-wfnn4_428e282a-fe3e-47f5-9d19-deb95236a4c9/cert-manager-cainjector/0.log" Feb 18 09:58:36 crc kubenswrapper[4556]: I0218 09:58:36.160072 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-j67wz_58f7a628-b1a3-4fd5-a730-0a80524f751c/cert-manager-webhook/0.log" Feb 18 09:58:47 crc kubenswrapper[4556]: I0218 09:58:47.118901 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5c78fc5d65-lzgcp_bccd4d34-9d89-40ef-9d4e-4c6a58d2571b/nmstate-console-plugin/0.log" Feb 18 09:58:47 crc kubenswrapper[4556]: I0218 09:58:47.281833 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-7dmwd_9a59cd87-e29a-4cf4-a407-3de0680bc1dc/nmstate-handler/0.log" Feb 18 09:58:47 crc kubenswrapper[4556]: I0218 09:58:47.351511 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58c85c668d-nhd7k_0933417f-4853-4022-b87e-7f3584341e8d/kube-rbac-proxy/0.log" Feb 18 09:58:47 crc kubenswrapper[4556]: I0218 09:58:47.362243 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58c85c668d-nhd7k_0933417f-4853-4022-b87e-7f3584341e8d/nmstate-metrics/0.log" Feb 18 09:58:47 crc kubenswrapper[4556]: I0218 09:58:47.458852 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-694c9596b7-mrrmv_b2ae390e-7b87-4f49-b95b-c39e89bbf523/nmstate-operator/0.log" Feb 18 09:58:47 crc kubenswrapper[4556]: I0218 09:58:47.531104 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-866bcb46dc-cm8rm_df346433-edc0-4b09-b028-ef60228567d1/nmstate-webhook/0.log" Feb 18 09:58:56 crc kubenswrapper[4556]: I0218 09:58:56.359059 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-24qsq"] Feb 18 09:58:56 crc kubenswrapper[4556]: E0218 09:58:56.359997 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19a44573-13ff-4677-9c4d-ab5dc378fff1" containerName="container-00" Feb 18 09:58:56 crc kubenswrapper[4556]: I0218 09:58:56.360014 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="19a44573-13ff-4677-9c4d-ab5dc378fff1" containerName="container-00" Feb 18 09:58:56 crc kubenswrapper[4556]: I0218 09:58:56.360214 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="19a44573-13ff-4677-9c4d-ab5dc378fff1" containerName="container-00" Feb 18 09:58:56 crc kubenswrapper[4556]: I0218 09:58:56.365684 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-24qsq" Feb 18 09:58:56 crc kubenswrapper[4556]: I0218 09:58:56.371587 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-24qsq"] Feb 18 09:58:56 crc kubenswrapper[4556]: I0218 09:58:56.417261 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8eebcd1f-e95a-46d3-a01d-e599eafc5c57-catalog-content\") pod \"certified-operators-24qsq\" (UID: \"8eebcd1f-e95a-46d3-a01d-e599eafc5c57\") " pod="openshift-marketplace/certified-operators-24qsq" Feb 18 09:58:56 crc kubenswrapper[4556]: I0218 09:58:56.417331 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8eebcd1f-e95a-46d3-a01d-e599eafc5c57-utilities\") pod \"certified-operators-24qsq\" (UID: \"8eebcd1f-e95a-46d3-a01d-e599eafc5c57\") " pod="openshift-marketplace/certified-operators-24qsq" Feb 18 09:58:56 crc kubenswrapper[4556]: I0218 09:58:56.417469 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-shswh\" (UniqueName: \"kubernetes.io/projected/8eebcd1f-e95a-46d3-a01d-e599eafc5c57-kube-api-access-shswh\") pod \"certified-operators-24qsq\" (UID: \"8eebcd1f-e95a-46d3-a01d-e599eafc5c57\") " pod="openshift-marketplace/certified-operators-24qsq" Feb 18 09:58:56 crc kubenswrapper[4556]: I0218 09:58:56.520209 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8eebcd1f-e95a-46d3-a01d-e599eafc5c57-catalog-content\") pod \"certified-operators-24qsq\" (UID: \"8eebcd1f-e95a-46d3-a01d-e599eafc5c57\") " pod="openshift-marketplace/certified-operators-24qsq" Feb 18 09:58:56 crc kubenswrapper[4556]: I0218 09:58:56.520279 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8eebcd1f-e95a-46d3-a01d-e599eafc5c57-utilities\") pod \"certified-operators-24qsq\" (UID: \"8eebcd1f-e95a-46d3-a01d-e599eafc5c57\") " pod="openshift-marketplace/certified-operators-24qsq" Feb 18 09:58:56 crc kubenswrapper[4556]: I0218 09:58:56.520360 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-shswh\" (UniqueName: \"kubernetes.io/projected/8eebcd1f-e95a-46d3-a01d-e599eafc5c57-kube-api-access-shswh\") pod \"certified-operators-24qsq\" (UID: \"8eebcd1f-e95a-46d3-a01d-e599eafc5c57\") " pod="openshift-marketplace/certified-operators-24qsq" Feb 18 09:58:56 crc kubenswrapper[4556]: I0218 09:58:56.520750 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8eebcd1f-e95a-46d3-a01d-e599eafc5c57-catalog-content\") pod \"certified-operators-24qsq\" (UID: \"8eebcd1f-e95a-46d3-a01d-e599eafc5c57\") " pod="openshift-marketplace/certified-operators-24qsq" Feb 18 09:58:56 crc kubenswrapper[4556]: I0218 09:58:56.520895 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8eebcd1f-e95a-46d3-a01d-e599eafc5c57-utilities\") pod \"certified-operators-24qsq\" (UID: \"8eebcd1f-e95a-46d3-a01d-e599eafc5c57\") " pod="openshift-marketplace/certified-operators-24qsq" Feb 18 09:58:56 crc kubenswrapper[4556]: I0218 09:58:56.542335 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-shswh\" (UniqueName: \"kubernetes.io/projected/8eebcd1f-e95a-46d3-a01d-e599eafc5c57-kube-api-access-shswh\") pod \"certified-operators-24qsq\" (UID: \"8eebcd1f-e95a-46d3-a01d-e599eafc5c57\") " pod="openshift-marketplace/certified-operators-24qsq" Feb 18 09:58:56 crc kubenswrapper[4556]: I0218 09:58:56.682498 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-24qsq" Feb 18 09:58:57 crc kubenswrapper[4556]: I0218 09:58:57.179743 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-24qsq"] Feb 18 09:58:57 crc kubenswrapper[4556]: I0218 09:58:57.667593 4556 generic.go:334] "Generic (PLEG): container finished" podID="8eebcd1f-e95a-46d3-a01d-e599eafc5c57" containerID="970fe78ca41d10c5cae9b863cac830b6b08f4aae9dc5735925650500e7c68ee9" exitCode=0 Feb 18 09:58:57 crc kubenswrapper[4556]: I0218 09:58:57.667700 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-24qsq" event={"ID":"8eebcd1f-e95a-46d3-a01d-e599eafc5c57","Type":"ContainerDied","Data":"970fe78ca41d10c5cae9b863cac830b6b08f4aae9dc5735925650500e7c68ee9"} Feb 18 09:58:57 crc kubenswrapper[4556]: I0218 09:58:57.669680 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-24qsq" event={"ID":"8eebcd1f-e95a-46d3-a01d-e599eafc5c57","Type":"ContainerStarted","Data":"6ab6ce9f0975b2a98432610e3a8c83bd2075ee676fb75c5c3cc20860e7d245ca"} Feb 18 09:58:57 crc kubenswrapper[4556]: I0218 09:58:57.670842 4556 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 18 09:58:58 crc kubenswrapper[4556]: I0218 09:58:58.679168 4556 generic.go:334] "Generic (PLEG): container finished" podID="8eebcd1f-e95a-46d3-a01d-e599eafc5c57" containerID="b40fbfec35273bb0cb38f4261b0371efd6c51fb0ec74874128f1783c1baafcb8" exitCode=0 Feb 18 09:58:58 crc kubenswrapper[4556]: I0218 09:58:58.679222 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-24qsq" event={"ID":"8eebcd1f-e95a-46d3-a01d-e599eafc5c57","Type":"ContainerDied","Data":"b40fbfec35273bb0cb38f4261b0371efd6c51fb0ec74874128f1783c1baafcb8"} Feb 18 09:58:58 crc kubenswrapper[4556]: I0218 09:58:58.752879 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-lcq58"] Feb 18 09:58:58 crc kubenswrapper[4556]: I0218 09:58:58.756005 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lcq58" Feb 18 09:58:58 crc kubenswrapper[4556]: I0218 09:58:58.768193 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f03b1325-32ed-4912-95e9-2f9bfaeab947-catalog-content\") pod \"community-operators-lcq58\" (UID: \"f03b1325-32ed-4912-95e9-2f9bfaeab947\") " pod="openshift-marketplace/community-operators-lcq58" Feb 18 09:58:58 crc kubenswrapper[4556]: I0218 09:58:58.768327 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f03b1325-32ed-4912-95e9-2f9bfaeab947-utilities\") pod \"community-operators-lcq58\" (UID: \"f03b1325-32ed-4912-95e9-2f9bfaeab947\") " pod="openshift-marketplace/community-operators-lcq58" Feb 18 09:58:58 crc kubenswrapper[4556]: I0218 09:58:58.768464 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kzfz\" (UniqueName: \"kubernetes.io/projected/f03b1325-32ed-4912-95e9-2f9bfaeab947-kube-api-access-5kzfz\") pod \"community-operators-lcq58\" (UID: \"f03b1325-32ed-4912-95e9-2f9bfaeab947\") " pod="openshift-marketplace/community-operators-lcq58" Feb 18 09:58:58 crc kubenswrapper[4556]: I0218 09:58:58.775063 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lcq58"] Feb 18 09:58:58 crc kubenswrapper[4556]: I0218 09:58:58.870442 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f03b1325-32ed-4912-95e9-2f9bfaeab947-catalog-content\") pod \"community-operators-lcq58\" (UID: \"f03b1325-32ed-4912-95e9-2f9bfaeab947\") " pod="openshift-marketplace/community-operators-lcq58" Feb 18 09:58:58 crc kubenswrapper[4556]: I0218 09:58:58.870560 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f03b1325-32ed-4912-95e9-2f9bfaeab947-utilities\") pod \"community-operators-lcq58\" (UID: \"f03b1325-32ed-4912-95e9-2f9bfaeab947\") " pod="openshift-marketplace/community-operators-lcq58" Feb 18 09:58:58 crc kubenswrapper[4556]: I0218 09:58:58.870643 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kzfz\" (UniqueName: \"kubernetes.io/projected/f03b1325-32ed-4912-95e9-2f9bfaeab947-kube-api-access-5kzfz\") pod \"community-operators-lcq58\" (UID: \"f03b1325-32ed-4912-95e9-2f9bfaeab947\") " pod="openshift-marketplace/community-operators-lcq58" Feb 18 09:58:58 crc kubenswrapper[4556]: I0218 09:58:58.870992 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f03b1325-32ed-4912-95e9-2f9bfaeab947-catalog-content\") pod \"community-operators-lcq58\" (UID: \"f03b1325-32ed-4912-95e9-2f9bfaeab947\") " pod="openshift-marketplace/community-operators-lcq58" Feb 18 09:58:58 crc kubenswrapper[4556]: I0218 09:58:58.871137 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f03b1325-32ed-4912-95e9-2f9bfaeab947-utilities\") pod \"community-operators-lcq58\" (UID: \"f03b1325-32ed-4912-95e9-2f9bfaeab947\") " pod="openshift-marketplace/community-operators-lcq58" Feb 18 09:58:58 crc kubenswrapper[4556]: I0218 09:58:58.897942 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kzfz\" (UniqueName: \"kubernetes.io/projected/f03b1325-32ed-4912-95e9-2f9bfaeab947-kube-api-access-5kzfz\") pod \"community-operators-lcq58\" (UID: \"f03b1325-32ed-4912-95e9-2f9bfaeab947\") " pod="openshift-marketplace/community-operators-lcq58" Feb 18 09:58:59 crc kubenswrapper[4556]: I0218 09:58:59.078797 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lcq58" Feb 18 09:58:59 crc kubenswrapper[4556]: I0218 09:58:59.524129 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lcq58"] Feb 18 09:58:59 crc kubenswrapper[4556]: W0218 09:58:59.525257 4556 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf03b1325_32ed_4912_95e9_2f9bfaeab947.slice/crio-17570f80d6fdac5e62c57eac87e35af5988f30870fb0f202f00fe97b21354911 WatchSource:0}: Error finding container 17570f80d6fdac5e62c57eac87e35af5988f30870fb0f202f00fe97b21354911: Status 404 returned error can't find the container with id 17570f80d6fdac5e62c57eac87e35af5988f30870fb0f202f00fe97b21354911 Feb 18 09:58:59 crc kubenswrapper[4556]: I0218 09:58:59.689720 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lcq58" event={"ID":"f03b1325-32ed-4912-95e9-2f9bfaeab947","Type":"ContainerStarted","Data":"6bfcb362b7a203afee5b89b0d347d728614814f56d0cdfc21aec1b8df3e4c76a"} Feb 18 09:58:59 crc kubenswrapper[4556]: I0218 09:58:59.689770 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lcq58" event={"ID":"f03b1325-32ed-4912-95e9-2f9bfaeab947","Type":"ContainerStarted","Data":"17570f80d6fdac5e62c57eac87e35af5988f30870fb0f202f00fe97b21354911"} Feb 18 09:58:59 crc kubenswrapper[4556]: I0218 09:58:59.696750 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-24qsq" event={"ID":"8eebcd1f-e95a-46d3-a01d-e599eafc5c57","Type":"ContainerStarted","Data":"9c76c820acc0034a80c0612e0eb0306c672927d114328eb728787e0daa8a6fa7"} Feb 18 09:58:59 crc kubenswrapper[4556]: I0218 09:58:59.726537 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-24qsq" podStartSLOduration=2.219790195 podStartE2EDuration="3.72646841s" podCreationTimestamp="2026-02-18 09:58:56 +0000 UTC" firstStartedPulling="2026-02-18 09:58:57.670556165 +0000 UTC m=+3294.687517145" lastFinishedPulling="2026-02-18 09:58:59.177234379 +0000 UTC m=+3296.194195360" observedRunningTime="2026-02-18 09:58:59.719375641 +0000 UTC m=+3296.736336621" watchObservedRunningTime="2026-02-18 09:58:59.72646841 +0000 UTC m=+3296.743429389" Feb 18 09:59:00 crc kubenswrapper[4556]: I0218 09:59:00.712498 4556 generic.go:334] "Generic (PLEG): container finished" podID="f03b1325-32ed-4912-95e9-2f9bfaeab947" containerID="6bfcb362b7a203afee5b89b0d347d728614814f56d0cdfc21aec1b8df3e4c76a" exitCode=0 Feb 18 09:59:00 crc kubenswrapper[4556]: I0218 09:59:00.714611 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lcq58" event={"ID":"f03b1325-32ed-4912-95e9-2f9bfaeab947","Type":"ContainerDied","Data":"6bfcb362b7a203afee5b89b0d347d728614814f56d0cdfc21aec1b8df3e4c76a"} Feb 18 09:59:01 crc kubenswrapper[4556]: I0218 09:59:01.721436 4556 generic.go:334] "Generic (PLEG): container finished" podID="f03b1325-32ed-4912-95e9-2f9bfaeab947" containerID="3ef596b38d37eca4b37cf10af7b01cb0313ecc04d9ea0b9ac3b4da948d93670b" exitCode=0 Feb 18 09:59:01 crc kubenswrapper[4556]: I0218 09:59:01.721637 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lcq58" event={"ID":"f03b1325-32ed-4912-95e9-2f9bfaeab947","Type":"ContainerDied","Data":"3ef596b38d37eca4b37cf10af7b01cb0313ecc04d9ea0b9ac3b4da948d93670b"} Feb 18 09:59:02 crc kubenswrapper[4556]: I0218 09:59:02.732402 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lcq58" event={"ID":"f03b1325-32ed-4912-95e9-2f9bfaeab947","Type":"ContainerStarted","Data":"856a3bb195f7d25fa3582b3b3d4d539b6470ca7a9dbe049a7a4805a4176f1f9f"} Feb 18 09:59:02 crc kubenswrapper[4556]: I0218 09:59:02.755013 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-lcq58" podStartSLOduration=3.319721957 podStartE2EDuration="4.75499478s" podCreationTimestamp="2026-02-18 09:58:58 +0000 UTC" firstStartedPulling="2026-02-18 09:59:00.715739616 +0000 UTC m=+3297.732700597" lastFinishedPulling="2026-02-18 09:59:02.15101244 +0000 UTC m=+3299.167973420" observedRunningTime="2026-02-18 09:59:02.74885033 +0000 UTC m=+3299.765811310" watchObservedRunningTime="2026-02-18 09:59:02.75499478 +0000 UTC m=+3299.771955760" Feb 18 09:59:06 crc kubenswrapper[4556]: I0218 09:59:06.683201 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-24qsq" Feb 18 09:59:06 crc kubenswrapper[4556]: I0218 09:59:06.684202 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-24qsq" Feb 18 09:59:06 crc kubenswrapper[4556]: I0218 09:59:06.726217 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-24qsq" Feb 18 09:59:06 crc kubenswrapper[4556]: I0218 09:59:06.805435 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-24qsq" Feb 18 09:59:06 crc kubenswrapper[4556]: I0218 09:59:06.959958 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-24qsq"] Feb 18 09:59:08 crc kubenswrapper[4556]: I0218 09:59:08.785914 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-24qsq" podUID="8eebcd1f-e95a-46d3-a01d-e599eafc5c57" containerName="registry-server" containerID="cri-o://9c76c820acc0034a80c0612e0eb0306c672927d114328eb728787e0daa8a6fa7" gracePeriod=2 Feb 18 09:59:09 crc kubenswrapper[4556]: I0218 09:59:09.079609 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-lcq58" Feb 18 09:59:09 crc kubenswrapper[4556]: I0218 09:59:09.081113 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-lcq58" Feb 18 09:59:09 crc kubenswrapper[4556]: I0218 09:59:09.119997 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-lcq58" Feb 18 09:59:09 crc kubenswrapper[4556]: I0218 09:59:09.243019 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-24qsq" Feb 18 09:59:09 crc kubenswrapper[4556]: I0218 09:59:09.388182 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8eebcd1f-e95a-46d3-a01d-e599eafc5c57-utilities\") pod \"8eebcd1f-e95a-46d3-a01d-e599eafc5c57\" (UID: \"8eebcd1f-e95a-46d3-a01d-e599eafc5c57\") " Feb 18 09:59:09 crc kubenswrapper[4556]: I0218 09:59:09.388263 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8eebcd1f-e95a-46d3-a01d-e599eafc5c57-catalog-content\") pod \"8eebcd1f-e95a-46d3-a01d-e599eafc5c57\" (UID: \"8eebcd1f-e95a-46d3-a01d-e599eafc5c57\") " Feb 18 09:59:09 crc kubenswrapper[4556]: I0218 09:59:09.388302 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-shswh\" (UniqueName: \"kubernetes.io/projected/8eebcd1f-e95a-46d3-a01d-e599eafc5c57-kube-api-access-shswh\") pod \"8eebcd1f-e95a-46d3-a01d-e599eafc5c57\" (UID: \"8eebcd1f-e95a-46d3-a01d-e599eafc5c57\") " Feb 18 09:59:09 crc kubenswrapper[4556]: I0218 09:59:09.395072 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8eebcd1f-e95a-46d3-a01d-e599eafc5c57-kube-api-access-shswh" (OuterVolumeSpecName: "kube-api-access-shswh") pod "8eebcd1f-e95a-46d3-a01d-e599eafc5c57" (UID: "8eebcd1f-e95a-46d3-a01d-e599eafc5c57"). InnerVolumeSpecName "kube-api-access-shswh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:59:09 crc kubenswrapper[4556]: I0218 09:59:09.400295 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8eebcd1f-e95a-46d3-a01d-e599eafc5c57-utilities" (OuterVolumeSpecName: "utilities") pod "8eebcd1f-e95a-46d3-a01d-e599eafc5c57" (UID: "8eebcd1f-e95a-46d3-a01d-e599eafc5c57"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:59:09 crc kubenswrapper[4556]: I0218 09:59:09.429050 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8eebcd1f-e95a-46d3-a01d-e599eafc5c57-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8eebcd1f-e95a-46d3-a01d-e599eafc5c57" (UID: "8eebcd1f-e95a-46d3-a01d-e599eafc5c57"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:59:09 crc kubenswrapper[4556]: I0218 09:59:09.490373 4556 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8eebcd1f-e95a-46d3-a01d-e599eafc5c57-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 09:59:09 crc kubenswrapper[4556]: I0218 09:59:09.490411 4556 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8eebcd1f-e95a-46d3-a01d-e599eafc5c57-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 09:59:09 crc kubenswrapper[4556]: I0218 09:59:09.490426 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-shswh\" (UniqueName: \"kubernetes.io/projected/8eebcd1f-e95a-46d3-a01d-e599eafc5c57-kube-api-access-shswh\") on node \"crc\" DevicePath \"\"" Feb 18 09:59:09 crc kubenswrapper[4556]: I0218 09:59:09.801926 4556 generic.go:334] "Generic (PLEG): container finished" podID="8eebcd1f-e95a-46d3-a01d-e599eafc5c57" containerID="9c76c820acc0034a80c0612e0eb0306c672927d114328eb728787e0daa8a6fa7" exitCode=0 Feb 18 09:59:09 crc kubenswrapper[4556]: I0218 09:59:09.802780 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-24qsq" Feb 18 09:59:09 crc kubenswrapper[4556]: I0218 09:59:09.807351 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-24qsq" event={"ID":"8eebcd1f-e95a-46d3-a01d-e599eafc5c57","Type":"ContainerDied","Data":"9c76c820acc0034a80c0612e0eb0306c672927d114328eb728787e0daa8a6fa7"} Feb 18 09:59:09 crc kubenswrapper[4556]: I0218 09:59:09.807389 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-24qsq" event={"ID":"8eebcd1f-e95a-46d3-a01d-e599eafc5c57","Type":"ContainerDied","Data":"6ab6ce9f0975b2a98432610e3a8c83bd2075ee676fb75c5c3cc20860e7d245ca"} Feb 18 09:59:09 crc kubenswrapper[4556]: I0218 09:59:09.807410 4556 scope.go:117] "RemoveContainer" containerID="9c76c820acc0034a80c0612e0eb0306c672927d114328eb728787e0daa8a6fa7" Feb 18 09:59:09 crc kubenswrapper[4556]: I0218 09:59:09.834719 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-24qsq"] Feb 18 09:59:09 crc kubenswrapper[4556]: I0218 09:59:09.837534 4556 scope.go:117] "RemoveContainer" containerID="b40fbfec35273bb0cb38f4261b0371efd6c51fb0ec74874128f1783c1baafcb8" Feb 18 09:59:09 crc kubenswrapper[4556]: I0218 09:59:09.839938 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-24qsq"] Feb 18 09:59:09 crc kubenswrapper[4556]: I0218 09:59:09.857094 4556 scope.go:117] "RemoveContainer" containerID="970fe78ca41d10c5cae9b863cac830b6b08f4aae9dc5735925650500e7c68ee9" Feb 18 09:59:09 crc kubenswrapper[4556]: I0218 09:59:09.861753 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-lcq58" Feb 18 09:59:09 crc kubenswrapper[4556]: I0218 09:59:09.890951 4556 scope.go:117] "RemoveContainer" containerID="9c76c820acc0034a80c0612e0eb0306c672927d114328eb728787e0daa8a6fa7" Feb 18 09:59:09 crc kubenswrapper[4556]: E0218 09:59:09.891436 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c76c820acc0034a80c0612e0eb0306c672927d114328eb728787e0daa8a6fa7\": container with ID starting with 9c76c820acc0034a80c0612e0eb0306c672927d114328eb728787e0daa8a6fa7 not found: ID does not exist" containerID="9c76c820acc0034a80c0612e0eb0306c672927d114328eb728787e0daa8a6fa7" Feb 18 09:59:09 crc kubenswrapper[4556]: I0218 09:59:09.891468 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c76c820acc0034a80c0612e0eb0306c672927d114328eb728787e0daa8a6fa7"} err="failed to get container status \"9c76c820acc0034a80c0612e0eb0306c672927d114328eb728787e0daa8a6fa7\": rpc error: code = NotFound desc = could not find container \"9c76c820acc0034a80c0612e0eb0306c672927d114328eb728787e0daa8a6fa7\": container with ID starting with 9c76c820acc0034a80c0612e0eb0306c672927d114328eb728787e0daa8a6fa7 not found: ID does not exist" Feb 18 09:59:09 crc kubenswrapper[4556]: I0218 09:59:09.891489 4556 scope.go:117] "RemoveContainer" containerID="b40fbfec35273bb0cb38f4261b0371efd6c51fb0ec74874128f1783c1baafcb8" Feb 18 09:59:09 crc kubenswrapper[4556]: E0218 09:59:09.891798 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b40fbfec35273bb0cb38f4261b0371efd6c51fb0ec74874128f1783c1baafcb8\": container with ID starting with b40fbfec35273bb0cb38f4261b0371efd6c51fb0ec74874128f1783c1baafcb8 not found: ID does not exist" containerID="b40fbfec35273bb0cb38f4261b0371efd6c51fb0ec74874128f1783c1baafcb8" Feb 18 09:59:09 crc kubenswrapper[4556]: I0218 09:59:09.891856 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b40fbfec35273bb0cb38f4261b0371efd6c51fb0ec74874128f1783c1baafcb8"} err="failed to get container status \"b40fbfec35273bb0cb38f4261b0371efd6c51fb0ec74874128f1783c1baafcb8\": rpc error: code = NotFound desc = could not find container \"b40fbfec35273bb0cb38f4261b0371efd6c51fb0ec74874128f1783c1baafcb8\": container with ID starting with b40fbfec35273bb0cb38f4261b0371efd6c51fb0ec74874128f1783c1baafcb8 not found: ID does not exist" Feb 18 09:59:09 crc kubenswrapper[4556]: I0218 09:59:09.891884 4556 scope.go:117] "RemoveContainer" containerID="970fe78ca41d10c5cae9b863cac830b6b08f4aae9dc5735925650500e7c68ee9" Feb 18 09:59:09 crc kubenswrapper[4556]: E0218 09:59:09.892169 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"970fe78ca41d10c5cae9b863cac830b6b08f4aae9dc5735925650500e7c68ee9\": container with ID starting with 970fe78ca41d10c5cae9b863cac830b6b08f4aae9dc5735925650500e7c68ee9 not found: ID does not exist" containerID="970fe78ca41d10c5cae9b863cac830b6b08f4aae9dc5735925650500e7c68ee9" Feb 18 09:59:09 crc kubenswrapper[4556]: I0218 09:59:09.892201 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"970fe78ca41d10c5cae9b863cac830b6b08f4aae9dc5735925650500e7c68ee9"} err="failed to get container status \"970fe78ca41d10c5cae9b863cac830b6b08f4aae9dc5735925650500e7c68ee9\": rpc error: code = NotFound desc = could not find container \"970fe78ca41d10c5cae9b863cac830b6b08f4aae9dc5735925650500e7c68ee9\": container with ID starting with 970fe78ca41d10c5cae9b863cac830b6b08f4aae9dc5735925650500e7c68ee9 not found: ID does not exist" Feb 18 09:59:11 crc kubenswrapper[4556]: I0218 09:59:11.293304 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8eebcd1f-e95a-46d3-a01d-e599eafc5c57" path="/var/lib/kubelet/pods/8eebcd1f-e95a-46d3-a01d-e599eafc5c57/volumes" Feb 18 09:59:12 crc kubenswrapper[4556]: I0218 09:59:12.159242 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lcq58"] Feb 18 09:59:12 crc kubenswrapper[4556]: I0218 09:59:12.833036 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-lcq58" podUID="f03b1325-32ed-4912-95e9-2f9bfaeab947" containerName="registry-server" containerID="cri-o://856a3bb195f7d25fa3582b3b3d4d539b6470ca7a9dbe049a7a4805a4176f1f9f" gracePeriod=2 Feb 18 09:59:12 crc kubenswrapper[4556]: I0218 09:59:12.848043 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-69bbfbf88f-l7dg7_cb943421-d850-4010-817d-15920051ce04/kube-rbac-proxy/0.log" Feb 18 09:59:13 crc kubenswrapper[4556]: I0218 09:59:12.999850 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-69bbfbf88f-l7dg7_cb943421-d850-4010-817d-15920051ce04/controller/0.log" Feb 18 09:59:13 crc kubenswrapper[4556]: I0218 09:59:13.066059 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qltwk_7cb5a650-ed14-42c7-8655-c5dfcaeb9d40/cp-frr-files/0.log" Feb 18 09:59:13 crc kubenswrapper[4556]: I0218 09:59:13.208711 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qltwk_7cb5a650-ed14-42c7-8655-c5dfcaeb9d40/cp-reloader/0.log" Feb 18 09:59:13 crc kubenswrapper[4556]: I0218 09:59:13.265315 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qltwk_7cb5a650-ed14-42c7-8655-c5dfcaeb9d40/cp-metrics/0.log" Feb 18 09:59:13 crc kubenswrapper[4556]: I0218 09:59:13.278949 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qltwk_7cb5a650-ed14-42c7-8655-c5dfcaeb9d40/cp-frr-files/0.log" Feb 18 09:59:13 crc kubenswrapper[4556]: I0218 09:59:13.279074 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qltwk_7cb5a650-ed14-42c7-8655-c5dfcaeb9d40/cp-reloader/0.log" Feb 18 09:59:13 crc kubenswrapper[4556]: I0218 09:59:13.279085 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lcq58" Feb 18 09:59:13 crc kubenswrapper[4556]: I0218 09:59:13.377788 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f03b1325-32ed-4912-95e9-2f9bfaeab947-utilities\") pod \"f03b1325-32ed-4912-95e9-2f9bfaeab947\" (UID: \"f03b1325-32ed-4912-95e9-2f9bfaeab947\") " Feb 18 09:59:13 crc kubenswrapper[4556]: I0218 09:59:13.377832 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5kzfz\" (UniqueName: \"kubernetes.io/projected/f03b1325-32ed-4912-95e9-2f9bfaeab947-kube-api-access-5kzfz\") pod \"f03b1325-32ed-4912-95e9-2f9bfaeab947\" (UID: \"f03b1325-32ed-4912-95e9-2f9bfaeab947\") " Feb 18 09:59:13 crc kubenswrapper[4556]: I0218 09:59:13.377972 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f03b1325-32ed-4912-95e9-2f9bfaeab947-catalog-content\") pod \"f03b1325-32ed-4912-95e9-2f9bfaeab947\" (UID: \"f03b1325-32ed-4912-95e9-2f9bfaeab947\") " Feb 18 09:59:13 crc kubenswrapper[4556]: I0218 09:59:13.379556 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f03b1325-32ed-4912-95e9-2f9bfaeab947-utilities" (OuterVolumeSpecName: "utilities") pod "f03b1325-32ed-4912-95e9-2f9bfaeab947" (UID: "f03b1325-32ed-4912-95e9-2f9bfaeab947"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:59:13 crc kubenswrapper[4556]: I0218 09:59:13.387379 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f03b1325-32ed-4912-95e9-2f9bfaeab947-kube-api-access-5kzfz" (OuterVolumeSpecName: "kube-api-access-5kzfz") pod "f03b1325-32ed-4912-95e9-2f9bfaeab947" (UID: "f03b1325-32ed-4912-95e9-2f9bfaeab947"). InnerVolumeSpecName "kube-api-access-5kzfz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 09:59:13 crc kubenswrapper[4556]: I0218 09:59:13.436987 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f03b1325-32ed-4912-95e9-2f9bfaeab947-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f03b1325-32ed-4912-95e9-2f9bfaeab947" (UID: "f03b1325-32ed-4912-95e9-2f9bfaeab947"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 09:59:13 crc kubenswrapper[4556]: I0218 09:59:13.453856 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qltwk_7cb5a650-ed14-42c7-8655-c5dfcaeb9d40/cp-metrics/0.log" Feb 18 09:59:13 crc kubenswrapper[4556]: I0218 09:59:13.467273 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qltwk_7cb5a650-ed14-42c7-8655-c5dfcaeb9d40/cp-reloader/0.log" Feb 18 09:59:13 crc kubenswrapper[4556]: I0218 09:59:13.480878 4556 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f03b1325-32ed-4912-95e9-2f9bfaeab947-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 09:59:13 crc kubenswrapper[4556]: I0218 09:59:13.480908 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5kzfz\" (UniqueName: \"kubernetes.io/projected/f03b1325-32ed-4912-95e9-2f9bfaeab947-kube-api-access-5kzfz\") on node \"crc\" DevicePath \"\"" Feb 18 09:59:13 crc kubenswrapper[4556]: I0218 09:59:13.480922 4556 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f03b1325-32ed-4912-95e9-2f9bfaeab947-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 09:59:13 crc kubenswrapper[4556]: I0218 09:59:13.489108 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qltwk_7cb5a650-ed14-42c7-8655-c5dfcaeb9d40/cp-metrics/0.log" Feb 18 09:59:13 crc kubenswrapper[4556]: I0218 09:59:13.497042 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qltwk_7cb5a650-ed14-42c7-8655-c5dfcaeb9d40/cp-frr-files/0.log" Feb 18 09:59:13 crc kubenswrapper[4556]: I0218 09:59:13.611834 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qltwk_7cb5a650-ed14-42c7-8655-c5dfcaeb9d40/cp-reloader/0.log" Feb 18 09:59:13 crc kubenswrapper[4556]: I0218 09:59:13.615572 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qltwk_7cb5a650-ed14-42c7-8655-c5dfcaeb9d40/cp-frr-files/0.log" Feb 18 09:59:13 crc kubenswrapper[4556]: I0218 09:59:13.656940 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qltwk_7cb5a650-ed14-42c7-8655-c5dfcaeb9d40/cp-metrics/0.log" Feb 18 09:59:13 crc kubenswrapper[4556]: I0218 09:59:13.663494 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qltwk_7cb5a650-ed14-42c7-8655-c5dfcaeb9d40/controller/0.log" Feb 18 09:59:13 crc kubenswrapper[4556]: I0218 09:59:13.775905 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qltwk_7cb5a650-ed14-42c7-8655-c5dfcaeb9d40/frr-metrics/0.log" Feb 18 09:59:13 crc kubenswrapper[4556]: I0218 09:59:13.812065 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qltwk_7cb5a650-ed14-42c7-8655-c5dfcaeb9d40/kube-rbac-proxy-frr/0.log" Feb 18 09:59:13 crc kubenswrapper[4556]: I0218 09:59:13.839317 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qltwk_7cb5a650-ed14-42c7-8655-c5dfcaeb9d40/kube-rbac-proxy/0.log" Feb 18 09:59:13 crc kubenswrapper[4556]: I0218 09:59:13.845997 4556 generic.go:334] "Generic (PLEG): container finished" podID="f03b1325-32ed-4912-95e9-2f9bfaeab947" containerID="856a3bb195f7d25fa3582b3b3d4d539b6470ca7a9dbe049a7a4805a4176f1f9f" exitCode=0 Feb 18 09:59:13 crc kubenswrapper[4556]: I0218 09:59:13.846048 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lcq58" event={"ID":"f03b1325-32ed-4912-95e9-2f9bfaeab947","Type":"ContainerDied","Data":"856a3bb195f7d25fa3582b3b3d4d539b6470ca7a9dbe049a7a4805a4176f1f9f"} Feb 18 09:59:13 crc kubenswrapper[4556]: I0218 09:59:13.846077 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lcq58" event={"ID":"f03b1325-32ed-4912-95e9-2f9bfaeab947","Type":"ContainerDied","Data":"17570f80d6fdac5e62c57eac87e35af5988f30870fb0f202f00fe97b21354911"} Feb 18 09:59:13 crc kubenswrapper[4556]: I0218 09:59:13.846103 4556 scope.go:117] "RemoveContainer" containerID="856a3bb195f7d25fa3582b3b3d4d539b6470ca7a9dbe049a7a4805a4176f1f9f" Feb 18 09:59:13 crc kubenswrapper[4556]: I0218 09:59:13.846344 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lcq58" Feb 18 09:59:13 crc kubenswrapper[4556]: I0218 09:59:13.877458 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lcq58"] Feb 18 09:59:13 crc kubenswrapper[4556]: I0218 09:59:13.877497 4556 scope.go:117] "RemoveContainer" containerID="3ef596b38d37eca4b37cf10af7b01cb0313ecc04d9ea0b9ac3b4da948d93670b" Feb 18 09:59:13 crc kubenswrapper[4556]: I0218 09:59:13.884015 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-lcq58"] Feb 18 09:59:13 crc kubenswrapper[4556]: I0218 09:59:13.906261 4556 scope.go:117] "RemoveContainer" containerID="6bfcb362b7a203afee5b89b0d347d728614814f56d0cdfc21aec1b8df3e4c76a" Feb 18 09:59:13 crc kubenswrapper[4556]: I0218 09:59:13.931972 4556 scope.go:117] "RemoveContainer" containerID="856a3bb195f7d25fa3582b3b3d4d539b6470ca7a9dbe049a7a4805a4176f1f9f" Feb 18 09:59:13 crc kubenswrapper[4556]: E0218 09:59:13.933760 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"856a3bb195f7d25fa3582b3b3d4d539b6470ca7a9dbe049a7a4805a4176f1f9f\": container with ID starting with 856a3bb195f7d25fa3582b3b3d4d539b6470ca7a9dbe049a7a4805a4176f1f9f not found: ID does not exist" containerID="856a3bb195f7d25fa3582b3b3d4d539b6470ca7a9dbe049a7a4805a4176f1f9f" Feb 18 09:59:13 crc kubenswrapper[4556]: I0218 09:59:13.933793 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"856a3bb195f7d25fa3582b3b3d4d539b6470ca7a9dbe049a7a4805a4176f1f9f"} err="failed to get container status \"856a3bb195f7d25fa3582b3b3d4d539b6470ca7a9dbe049a7a4805a4176f1f9f\": rpc error: code = NotFound desc = could not find container \"856a3bb195f7d25fa3582b3b3d4d539b6470ca7a9dbe049a7a4805a4176f1f9f\": container with ID starting with 856a3bb195f7d25fa3582b3b3d4d539b6470ca7a9dbe049a7a4805a4176f1f9f not found: ID does not exist" Feb 18 09:59:13 crc kubenswrapper[4556]: I0218 09:59:13.933842 4556 scope.go:117] "RemoveContainer" containerID="3ef596b38d37eca4b37cf10af7b01cb0313ecc04d9ea0b9ac3b4da948d93670b" Feb 18 09:59:13 crc kubenswrapper[4556]: E0218 09:59:13.934170 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ef596b38d37eca4b37cf10af7b01cb0313ecc04d9ea0b9ac3b4da948d93670b\": container with ID starting with 3ef596b38d37eca4b37cf10af7b01cb0313ecc04d9ea0b9ac3b4da948d93670b not found: ID does not exist" containerID="3ef596b38d37eca4b37cf10af7b01cb0313ecc04d9ea0b9ac3b4da948d93670b" Feb 18 09:59:13 crc kubenswrapper[4556]: I0218 09:59:13.934221 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ef596b38d37eca4b37cf10af7b01cb0313ecc04d9ea0b9ac3b4da948d93670b"} err="failed to get container status \"3ef596b38d37eca4b37cf10af7b01cb0313ecc04d9ea0b9ac3b4da948d93670b\": rpc error: code = NotFound desc = could not find container \"3ef596b38d37eca4b37cf10af7b01cb0313ecc04d9ea0b9ac3b4da948d93670b\": container with ID starting with 3ef596b38d37eca4b37cf10af7b01cb0313ecc04d9ea0b9ac3b4da948d93670b not found: ID does not exist" Feb 18 09:59:13 crc kubenswrapper[4556]: I0218 09:59:13.934246 4556 scope.go:117] "RemoveContainer" containerID="6bfcb362b7a203afee5b89b0d347d728614814f56d0cdfc21aec1b8df3e4c76a" Feb 18 09:59:13 crc kubenswrapper[4556]: E0218 09:59:13.934554 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6bfcb362b7a203afee5b89b0d347d728614814f56d0cdfc21aec1b8df3e4c76a\": container with ID starting with 6bfcb362b7a203afee5b89b0d347d728614814f56d0cdfc21aec1b8df3e4c76a not found: ID does not exist" containerID="6bfcb362b7a203afee5b89b0d347d728614814f56d0cdfc21aec1b8df3e4c76a" Feb 18 09:59:13 crc kubenswrapper[4556]: I0218 09:59:13.934632 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6bfcb362b7a203afee5b89b0d347d728614814f56d0cdfc21aec1b8df3e4c76a"} err="failed to get container status \"6bfcb362b7a203afee5b89b0d347d728614814f56d0cdfc21aec1b8df3e4c76a\": rpc error: code = NotFound desc = could not find container \"6bfcb362b7a203afee5b89b0d347d728614814f56d0cdfc21aec1b8df3e4c76a\": container with ID starting with 6bfcb362b7a203afee5b89b0d347d728614814f56d0cdfc21aec1b8df3e4c76a not found: ID does not exist" Feb 18 09:59:13 crc kubenswrapper[4556]: I0218 09:59:13.986466 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qltwk_7cb5a650-ed14-42c7-8655-c5dfcaeb9d40/reloader/0.log" Feb 18 09:59:14 crc kubenswrapper[4556]: I0218 09:59:14.023815 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-78b44bf5bb-spppg_04e46782-3a29-4a3d-8b82-ff9f8d34090c/frr-k8s-webhook-server/0.log" Feb 18 09:59:14 crc kubenswrapper[4556]: I0218 09:59:14.207465 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-7cccb48f48-gbb9s_6750b0ea-8f96-404d-bd38-e04e8be95127/manager/0.log" Feb 18 09:59:14 crc kubenswrapper[4556]: I0218 09:59:14.439743 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-749d56bf4-c8bg9_414515f8-ef03-41cd-a1c8-7811efcca959/webhook-server/0.log" Feb 18 09:59:14 crc kubenswrapper[4556]: I0218 09:59:14.546931 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-cjpmm_90e14ccf-fe48-4f1a-85af-5e7aa2cc1874/kube-rbac-proxy/0.log" Feb 18 09:59:15 crc kubenswrapper[4556]: I0218 09:59:15.025450 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-cjpmm_90e14ccf-fe48-4f1a-85af-5e7aa2cc1874/speaker/0.log" Feb 18 09:59:15 crc kubenswrapper[4556]: I0218 09:59:15.074252 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qltwk_7cb5a650-ed14-42c7-8655-c5dfcaeb9d40/frr/0.log" Feb 18 09:59:15 crc kubenswrapper[4556]: I0218 09:59:15.292376 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f03b1325-32ed-4912-95e9-2f9bfaeab947" path="/var/lib/kubelet/pods/f03b1325-32ed-4912-95e9-2f9bfaeab947/volumes" Feb 18 09:59:25 crc kubenswrapper[4556]: I0218 09:59:25.653281 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gp95f_3adcbc7e-17be-4546-bd1b-38b6ddee82b6/util/0.log" Feb 18 09:59:25 crc kubenswrapper[4556]: I0218 09:59:25.871209 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gp95f_3adcbc7e-17be-4546-bd1b-38b6ddee82b6/pull/0.log" Feb 18 09:59:25 crc kubenswrapper[4556]: I0218 09:59:25.879546 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gp95f_3adcbc7e-17be-4546-bd1b-38b6ddee82b6/pull/0.log" Feb 18 09:59:25 crc kubenswrapper[4556]: I0218 09:59:25.886590 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gp95f_3adcbc7e-17be-4546-bd1b-38b6ddee82b6/util/0.log" Feb 18 09:59:26 crc kubenswrapper[4556]: I0218 09:59:26.020648 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gp95f_3adcbc7e-17be-4546-bd1b-38b6ddee82b6/pull/0.log" Feb 18 09:59:26 crc kubenswrapper[4556]: I0218 09:59:26.026745 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gp95f_3adcbc7e-17be-4546-bd1b-38b6ddee82b6/extract/0.log" Feb 18 09:59:26 crc kubenswrapper[4556]: I0218 09:59:26.033519 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gp95f_3adcbc7e-17be-4546-bd1b-38b6ddee82b6/util/0.log" Feb 18 09:59:26 crc kubenswrapper[4556]: I0218 09:59:26.172889 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bfrcb_6ba6b57c-a8a5-4692-955d-e851c1014fa4/extract-utilities/0.log" Feb 18 09:59:26 crc kubenswrapper[4556]: I0218 09:59:26.332011 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bfrcb_6ba6b57c-a8a5-4692-955d-e851c1014fa4/extract-content/0.log" Feb 18 09:59:26 crc kubenswrapper[4556]: I0218 09:59:26.345481 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bfrcb_6ba6b57c-a8a5-4692-955d-e851c1014fa4/extract-utilities/0.log" Feb 18 09:59:26 crc kubenswrapper[4556]: I0218 09:59:26.359516 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bfrcb_6ba6b57c-a8a5-4692-955d-e851c1014fa4/extract-content/0.log" Feb 18 09:59:26 crc kubenswrapper[4556]: I0218 09:59:26.479514 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bfrcb_6ba6b57c-a8a5-4692-955d-e851c1014fa4/extract-utilities/0.log" Feb 18 09:59:26 crc kubenswrapper[4556]: I0218 09:59:26.524573 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bfrcb_6ba6b57c-a8a5-4692-955d-e851c1014fa4/extract-content/0.log" Feb 18 09:59:26 crc kubenswrapper[4556]: I0218 09:59:26.713772 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ldrzh_745b3a46-b0b9-443b-95d2-6bfd635d1243/extract-utilities/0.log" Feb 18 09:59:26 crc kubenswrapper[4556]: I0218 09:59:26.852381 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ldrzh_745b3a46-b0b9-443b-95d2-6bfd635d1243/extract-utilities/0.log" Feb 18 09:59:26 crc kubenswrapper[4556]: I0218 09:59:26.886350 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ldrzh_745b3a46-b0b9-443b-95d2-6bfd635d1243/extract-content/0.log" Feb 18 09:59:26 crc kubenswrapper[4556]: I0218 09:59:26.950530 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bfrcb_6ba6b57c-a8a5-4692-955d-e851c1014fa4/registry-server/0.log" Feb 18 09:59:26 crc kubenswrapper[4556]: I0218 09:59:26.953186 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ldrzh_745b3a46-b0b9-443b-95d2-6bfd635d1243/extract-content/0.log" Feb 18 09:59:27 crc kubenswrapper[4556]: I0218 09:59:27.269961 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ldrzh_745b3a46-b0b9-443b-95d2-6bfd635d1243/extract-utilities/0.log" Feb 18 09:59:27 crc kubenswrapper[4556]: I0218 09:59:27.310563 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ldrzh_745b3a46-b0b9-443b-95d2-6bfd635d1243/extract-content/0.log" Feb 18 09:59:27 crc kubenswrapper[4556]: I0218 09:59:27.493806 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecal8bm8_6b5ce21b-7d77-4d8c-b2d6-272ab408d929/util/0.log" Feb 18 09:59:27 crc kubenswrapper[4556]: I0218 09:59:27.668570 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecal8bm8_6b5ce21b-7d77-4d8c-b2d6-272ab408d929/pull/0.log" Feb 18 09:59:27 crc kubenswrapper[4556]: I0218 09:59:27.693662 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecal8bm8_6b5ce21b-7d77-4d8c-b2d6-272ab408d929/pull/0.log" Feb 18 09:59:27 crc kubenswrapper[4556]: I0218 09:59:27.701823 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ldrzh_745b3a46-b0b9-443b-95d2-6bfd635d1243/registry-server/0.log" Feb 18 09:59:27 crc kubenswrapper[4556]: I0218 09:59:27.721961 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecal8bm8_6b5ce21b-7d77-4d8c-b2d6-272ab408d929/util/0.log" Feb 18 09:59:27 crc kubenswrapper[4556]: I0218 09:59:27.819421 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecal8bm8_6b5ce21b-7d77-4d8c-b2d6-272ab408d929/pull/0.log" Feb 18 09:59:27 crc kubenswrapper[4556]: I0218 09:59:27.842837 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecal8bm8_6b5ce21b-7d77-4d8c-b2d6-272ab408d929/util/0.log" Feb 18 09:59:27 crc kubenswrapper[4556]: I0218 09:59:27.859270 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecal8bm8_6b5ce21b-7d77-4d8c-b2d6-272ab408d929/extract/0.log" Feb 18 09:59:27 crc kubenswrapper[4556]: I0218 09:59:27.973249 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-vh264_070f724e-944f-48d2-bb39-67b98f7667ce/marketplace-operator/0.log" Feb 18 09:59:28 crc kubenswrapper[4556]: I0218 09:59:28.021063 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m8g6v_7e5017d0-70be-4177-9c9c-827f97ed470b/extract-utilities/0.log" Feb 18 09:59:28 crc kubenswrapper[4556]: I0218 09:59:28.207960 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m8g6v_7e5017d0-70be-4177-9c9c-827f97ed470b/extract-content/0.log" Feb 18 09:59:28 crc kubenswrapper[4556]: I0218 09:59:28.214966 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m8g6v_7e5017d0-70be-4177-9c9c-827f97ed470b/extract-utilities/0.log" Feb 18 09:59:28 crc kubenswrapper[4556]: I0218 09:59:28.221516 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m8g6v_7e5017d0-70be-4177-9c9c-827f97ed470b/extract-content/0.log" Feb 18 09:59:28 crc kubenswrapper[4556]: I0218 09:59:28.396750 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m8g6v_7e5017d0-70be-4177-9c9c-827f97ed470b/extract-content/0.log" Feb 18 09:59:28 crc kubenswrapper[4556]: I0218 09:59:28.491593 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m8g6v_7e5017d0-70be-4177-9c9c-827f97ed470b/extract-utilities/0.log" Feb 18 09:59:28 crc kubenswrapper[4556]: I0218 09:59:28.546134 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m8g6v_7e5017d0-70be-4177-9c9c-827f97ed470b/registry-server/0.log" Feb 18 09:59:28 crc kubenswrapper[4556]: I0218 09:59:28.622029 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-w6td5_e82c70e2-ba17-4919-a836-a6e252af6380/extract-utilities/0.log" Feb 18 09:59:28 crc kubenswrapper[4556]: I0218 09:59:28.812726 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-w6td5_e82c70e2-ba17-4919-a836-a6e252af6380/extract-utilities/0.log" Feb 18 09:59:28 crc kubenswrapper[4556]: I0218 09:59:28.838704 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-w6td5_e82c70e2-ba17-4919-a836-a6e252af6380/extract-content/0.log" Feb 18 09:59:28 crc kubenswrapper[4556]: I0218 09:59:28.851051 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-w6td5_e82c70e2-ba17-4919-a836-a6e252af6380/extract-content/0.log" Feb 18 09:59:28 crc kubenswrapper[4556]: I0218 09:59:28.997513 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-w6td5_e82c70e2-ba17-4919-a836-a6e252af6380/extract-content/0.log" Feb 18 09:59:29 crc kubenswrapper[4556]: I0218 09:59:29.014468 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-w6td5_e82c70e2-ba17-4919-a836-a6e252af6380/extract-utilities/0.log" Feb 18 09:59:29 crc kubenswrapper[4556]: I0218 09:59:29.382611 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-w6td5_e82c70e2-ba17-4919-a836-a6e252af6380/registry-server/0.log" Feb 18 10:00:00 crc kubenswrapper[4556]: I0218 10:00:00.145680 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29523480-qhqsp"] Feb 18 10:00:00 crc kubenswrapper[4556]: E0218 10:00:00.146477 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8eebcd1f-e95a-46d3-a01d-e599eafc5c57" containerName="extract-utilities" Feb 18 10:00:00 crc kubenswrapper[4556]: I0218 10:00:00.146493 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="8eebcd1f-e95a-46d3-a01d-e599eafc5c57" containerName="extract-utilities" Feb 18 10:00:00 crc kubenswrapper[4556]: E0218 10:00:00.146511 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f03b1325-32ed-4912-95e9-2f9bfaeab947" containerName="extract-content" Feb 18 10:00:00 crc kubenswrapper[4556]: I0218 10:00:00.146516 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="f03b1325-32ed-4912-95e9-2f9bfaeab947" containerName="extract-content" Feb 18 10:00:00 crc kubenswrapper[4556]: E0218 10:00:00.146527 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f03b1325-32ed-4912-95e9-2f9bfaeab947" containerName="registry-server" Feb 18 10:00:00 crc kubenswrapper[4556]: I0218 10:00:00.146532 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="f03b1325-32ed-4912-95e9-2f9bfaeab947" containerName="registry-server" Feb 18 10:00:00 crc kubenswrapper[4556]: E0218 10:00:00.146556 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f03b1325-32ed-4912-95e9-2f9bfaeab947" containerName="extract-utilities" Feb 18 10:00:00 crc kubenswrapper[4556]: I0218 10:00:00.146561 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="f03b1325-32ed-4912-95e9-2f9bfaeab947" containerName="extract-utilities" Feb 18 10:00:00 crc kubenswrapper[4556]: E0218 10:00:00.146572 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8eebcd1f-e95a-46d3-a01d-e599eafc5c57" containerName="extract-content" Feb 18 10:00:00 crc kubenswrapper[4556]: I0218 10:00:00.146577 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="8eebcd1f-e95a-46d3-a01d-e599eafc5c57" containerName="extract-content" Feb 18 10:00:00 crc kubenswrapper[4556]: E0218 10:00:00.146591 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8eebcd1f-e95a-46d3-a01d-e599eafc5c57" containerName="registry-server" Feb 18 10:00:00 crc kubenswrapper[4556]: I0218 10:00:00.146803 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="8eebcd1f-e95a-46d3-a01d-e599eafc5c57" containerName="registry-server" Feb 18 10:00:00 crc kubenswrapper[4556]: I0218 10:00:00.148049 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="8eebcd1f-e95a-46d3-a01d-e599eafc5c57" containerName="registry-server" Feb 18 10:00:00 crc kubenswrapper[4556]: I0218 10:00:00.148083 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="f03b1325-32ed-4912-95e9-2f9bfaeab947" containerName="registry-server" Feb 18 10:00:00 crc kubenswrapper[4556]: I0218 10:00:00.148691 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29523480-qhqsp" Feb 18 10:00:00 crc kubenswrapper[4556]: I0218 10:00:00.150713 4556 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 18 10:00:00 crc kubenswrapper[4556]: I0218 10:00:00.154901 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29523480-qhqsp"] Feb 18 10:00:00 crc kubenswrapper[4556]: I0218 10:00:00.159432 4556 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 18 10:00:00 crc kubenswrapper[4556]: I0218 10:00:00.197426 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czpwj\" (UniqueName: \"kubernetes.io/projected/7c77aaed-e8e3-4b33-8744-0f96ce8e9c53-kube-api-access-czpwj\") pod \"collect-profiles-29523480-qhqsp\" (UID: \"7c77aaed-e8e3-4b33-8744-0f96ce8e9c53\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523480-qhqsp" Feb 18 10:00:00 crc kubenswrapper[4556]: I0218 10:00:00.197585 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7c77aaed-e8e3-4b33-8744-0f96ce8e9c53-secret-volume\") pod \"collect-profiles-29523480-qhqsp\" (UID: \"7c77aaed-e8e3-4b33-8744-0f96ce8e9c53\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523480-qhqsp" Feb 18 10:00:00 crc kubenswrapper[4556]: I0218 10:00:00.197670 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7c77aaed-e8e3-4b33-8744-0f96ce8e9c53-config-volume\") pod \"collect-profiles-29523480-qhqsp\" (UID: \"7c77aaed-e8e3-4b33-8744-0f96ce8e9c53\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523480-qhqsp" Feb 18 10:00:00 crc kubenswrapper[4556]: I0218 10:00:00.298199 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czpwj\" (UniqueName: \"kubernetes.io/projected/7c77aaed-e8e3-4b33-8744-0f96ce8e9c53-kube-api-access-czpwj\") pod \"collect-profiles-29523480-qhqsp\" (UID: \"7c77aaed-e8e3-4b33-8744-0f96ce8e9c53\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523480-qhqsp" Feb 18 10:00:00 crc kubenswrapper[4556]: I0218 10:00:00.298405 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7c77aaed-e8e3-4b33-8744-0f96ce8e9c53-secret-volume\") pod \"collect-profiles-29523480-qhqsp\" (UID: \"7c77aaed-e8e3-4b33-8744-0f96ce8e9c53\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523480-qhqsp" Feb 18 10:00:00 crc kubenswrapper[4556]: I0218 10:00:00.298506 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7c77aaed-e8e3-4b33-8744-0f96ce8e9c53-config-volume\") pod \"collect-profiles-29523480-qhqsp\" (UID: \"7c77aaed-e8e3-4b33-8744-0f96ce8e9c53\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523480-qhqsp" Feb 18 10:00:00 crc kubenswrapper[4556]: I0218 10:00:00.299465 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7c77aaed-e8e3-4b33-8744-0f96ce8e9c53-config-volume\") pod \"collect-profiles-29523480-qhqsp\" (UID: \"7c77aaed-e8e3-4b33-8744-0f96ce8e9c53\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523480-qhqsp" Feb 18 10:00:00 crc kubenswrapper[4556]: I0218 10:00:00.311525 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7c77aaed-e8e3-4b33-8744-0f96ce8e9c53-secret-volume\") pod \"collect-profiles-29523480-qhqsp\" (UID: \"7c77aaed-e8e3-4b33-8744-0f96ce8e9c53\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523480-qhqsp" Feb 18 10:00:00 crc kubenswrapper[4556]: I0218 10:00:00.312040 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czpwj\" (UniqueName: \"kubernetes.io/projected/7c77aaed-e8e3-4b33-8744-0f96ce8e9c53-kube-api-access-czpwj\") pod \"collect-profiles-29523480-qhqsp\" (UID: \"7c77aaed-e8e3-4b33-8744-0f96ce8e9c53\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523480-qhqsp" Feb 18 10:00:00 crc kubenswrapper[4556]: I0218 10:00:00.465713 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29523480-qhqsp" Feb 18 10:00:00 crc kubenswrapper[4556]: I0218 10:00:00.870099 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29523480-qhqsp"] Feb 18 10:00:01 crc kubenswrapper[4556]: I0218 10:00:01.266520 4556 generic.go:334] "Generic (PLEG): container finished" podID="7c77aaed-e8e3-4b33-8744-0f96ce8e9c53" containerID="1d0e6d679a9525c72bd8322d47600da1952a2eec17b3d8b5bca6dee6f4b999b6" exitCode=0 Feb 18 10:00:01 crc kubenswrapper[4556]: I0218 10:00:01.266592 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29523480-qhqsp" event={"ID":"7c77aaed-e8e3-4b33-8744-0f96ce8e9c53","Type":"ContainerDied","Data":"1d0e6d679a9525c72bd8322d47600da1952a2eec17b3d8b5bca6dee6f4b999b6"} Feb 18 10:00:01 crc kubenswrapper[4556]: I0218 10:00:01.266630 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29523480-qhqsp" event={"ID":"7c77aaed-e8e3-4b33-8744-0f96ce8e9c53","Type":"ContainerStarted","Data":"04f125f9d92e3dccfcec87390d59dc8704caf2b296f41ddf4a1172c357489a3c"} Feb 18 10:00:01 crc kubenswrapper[4556]: I0218 10:00:01.727900 4556 patch_prober.go:28] interesting pod/machine-config-daemon-f76hs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 10:00:01 crc kubenswrapper[4556]: I0218 10:00:01.727974 4556 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 10:00:02 crc kubenswrapper[4556]: I0218 10:00:02.613693 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29523480-qhqsp" Feb 18 10:00:02 crc kubenswrapper[4556]: I0218 10:00:02.650512 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-czpwj\" (UniqueName: \"kubernetes.io/projected/7c77aaed-e8e3-4b33-8744-0f96ce8e9c53-kube-api-access-czpwj\") pod \"7c77aaed-e8e3-4b33-8744-0f96ce8e9c53\" (UID: \"7c77aaed-e8e3-4b33-8744-0f96ce8e9c53\") " Feb 18 10:00:02 crc kubenswrapper[4556]: I0218 10:00:02.650731 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7c77aaed-e8e3-4b33-8744-0f96ce8e9c53-config-volume\") pod \"7c77aaed-e8e3-4b33-8744-0f96ce8e9c53\" (UID: \"7c77aaed-e8e3-4b33-8744-0f96ce8e9c53\") " Feb 18 10:00:02 crc kubenswrapper[4556]: I0218 10:00:02.650851 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7c77aaed-e8e3-4b33-8744-0f96ce8e9c53-secret-volume\") pod \"7c77aaed-e8e3-4b33-8744-0f96ce8e9c53\" (UID: \"7c77aaed-e8e3-4b33-8744-0f96ce8e9c53\") " Feb 18 10:00:02 crc kubenswrapper[4556]: I0218 10:00:02.653063 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c77aaed-e8e3-4b33-8744-0f96ce8e9c53-config-volume" (OuterVolumeSpecName: "config-volume") pod "7c77aaed-e8e3-4b33-8744-0f96ce8e9c53" (UID: "7c77aaed-e8e3-4b33-8744-0f96ce8e9c53"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 10:00:02 crc kubenswrapper[4556]: I0218 10:00:02.656935 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c77aaed-e8e3-4b33-8744-0f96ce8e9c53-kube-api-access-czpwj" (OuterVolumeSpecName: "kube-api-access-czpwj") pod "7c77aaed-e8e3-4b33-8744-0f96ce8e9c53" (UID: "7c77aaed-e8e3-4b33-8744-0f96ce8e9c53"). InnerVolumeSpecName "kube-api-access-czpwj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 10:00:02 crc kubenswrapper[4556]: I0218 10:00:02.656979 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c77aaed-e8e3-4b33-8744-0f96ce8e9c53-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "7c77aaed-e8e3-4b33-8744-0f96ce8e9c53" (UID: "7c77aaed-e8e3-4b33-8744-0f96ce8e9c53"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 10:00:02 crc kubenswrapper[4556]: I0218 10:00:02.752732 4556 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7c77aaed-e8e3-4b33-8744-0f96ce8e9c53-config-volume\") on node \"crc\" DevicePath \"\"" Feb 18 10:00:02 crc kubenswrapper[4556]: I0218 10:00:02.753123 4556 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7c77aaed-e8e3-4b33-8744-0f96ce8e9c53-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 18 10:00:02 crc kubenswrapper[4556]: I0218 10:00:02.753140 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-czpwj\" (UniqueName: \"kubernetes.io/projected/7c77aaed-e8e3-4b33-8744-0f96ce8e9c53-kube-api-access-czpwj\") on node \"crc\" DevicePath \"\"" Feb 18 10:00:03 crc kubenswrapper[4556]: I0218 10:00:03.293139 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29523480-qhqsp" Feb 18 10:00:03 crc kubenswrapper[4556]: I0218 10:00:03.296225 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29523480-qhqsp" event={"ID":"7c77aaed-e8e3-4b33-8744-0f96ce8e9c53","Type":"ContainerDied","Data":"04f125f9d92e3dccfcec87390d59dc8704caf2b296f41ddf4a1172c357489a3c"} Feb 18 10:00:03 crc kubenswrapper[4556]: I0218 10:00:03.296306 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="04f125f9d92e3dccfcec87390d59dc8704caf2b296f41ddf4a1172c357489a3c" Feb 18 10:00:03 crc kubenswrapper[4556]: I0218 10:00:03.683464 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29523435-9ztbm"] Feb 18 10:00:03 crc kubenswrapper[4556]: I0218 10:00:03.690716 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29523435-9ztbm"] Feb 18 10:00:05 crc kubenswrapper[4556]: I0218 10:00:05.292591 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62835cbc-94a5-4b92-9b0c-eb44d39a5af5" path="/var/lib/kubelet/pods/62835cbc-94a5-4b92-9b0c-eb44d39a5af5/volumes" Feb 18 10:00:11 crc kubenswrapper[4556]: I0218 10:00:11.612477 4556 scope.go:117] "RemoveContainer" containerID="7e284035a6bf3b5a640f304fc8621b89e3a15039dd83d408395f86d674659470" Feb 18 10:00:12 crc kubenswrapper[4556]: E0218 10:00:12.347460 4556 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7c77aaed_e8e3_4b33_8744_0f96ce8e9c53.slice\": RecentStats: unable to find data in memory cache]" Feb 18 10:00:22 crc kubenswrapper[4556]: E0218 10:00:22.529738 4556 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7c77aaed_e8e3_4b33_8744_0f96ce8e9c53.slice\": RecentStats: unable to find data in memory cache]" Feb 18 10:00:31 crc kubenswrapper[4556]: I0218 10:00:31.727250 4556 patch_prober.go:28] interesting pod/machine-config-daemon-f76hs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 10:00:31 crc kubenswrapper[4556]: I0218 10:00:31.727636 4556 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 10:00:32 crc kubenswrapper[4556]: E0218 10:00:32.739001 4556 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7c77aaed_e8e3_4b33_8744_0f96ce8e9c53.slice\": RecentStats: unable to find data in memory cache]" Feb 18 10:00:42 crc kubenswrapper[4556]: E0218 10:00:42.939266 4556 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7c77aaed_e8e3_4b33_8744_0f96ce8e9c53.slice\": RecentStats: unable to find data in memory cache]" Feb 18 10:00:53 crc kubenswrapper[4556]: E0218 10:00:53.133561 4556 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7c77aaed_e8e3_4b33_8744_0f96ce8e9c53.slice\": RecentStats: unable to find data in memory cache]" Feb 18 10:00:59 crc kubenswrapper[4556]: I0218 10:00:59.765219 4556 generic.go:334] "Generic (PLEG): container finished" podID="106c448a-8930-4be7-aa0e-970ab3d8b33e" containerID="ab98b80451544ff53c43502227fd54e98ead544320b3850fccc3876aa72bd333" exitCode=0 Feb 18 10:00:59 crc kubenswrapper[4556]: I0218 10:00:59.765301 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rnjw6/must-gather-vhc8n" event={"ID":"106c448a-8930-4be7-aa0e-970ab3d8b33e","Type":"ContainerDied","Data":"ab98b80451544ff53c43502227fd54e98ead544320b3850fccc3876aa72bd333"} Feb 18 10:00:59 crc kubenswrapper[4556]: I0218 10:00:59.766271 4556 scope.go:117] "RemoveContainer" containerID="ab98b80451544ff53c43502227fd54e98ead544320b3850fccc3876aa72bd333" Feb 18 10:01:00 crc kubenswrapper[4556]: I0218 10:01:00.152657 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29523481-bmw8g"] Feb 18 10:01:00 crc kubenswrapper[4556]: E0218 10:01:00.153191 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c77aaed-e8e3-4b33-8744-0f96ce8e9c53" containerName="collect-profiles" Feb 18 10:01:00 crc kubenswrapper[4556]: I0218 10:01:00.153209 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c77aaed-e8e3-4b33-8744-0f96ce8e9c53" containerName="collect-profiles" Feb 18 10:01:00 crc kubenswrapper[4556]: I0218 10:01:00.153384 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c77aaed-e8e3-4b33-8744-0f96ce8e9c53" containerName="collect-profiles" Feb 18 10:01:00 crc kubenswrapper[4556]: I0218 10:01:00.154032 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29523481-bmw8g" Feb 18 10:01:00 crc kubenswrapper[4556]: I0218 10:01:00.161230 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29523481-bmw8g"] Feb 18 10:01:00 crc kubenswrapper[4556]: I0218 10:01:00.314585 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/db14278b-852a-47e6-aded-fadfbee5fd52-fernet-keys\") pod \"keystone-cron-29523481-bmw8g\" (UID: \"db14278b-852a-47e6-aded-fadfbee5fd52\") " pod="openstack/keystone-cron-29523481-bmw8g" Feb 18 10:01:00 crc kubenswrapper[4556]: I0218 10:01:00.314806 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db14278b-852a-47e6-aded-fadfbee5fd52-combined-ca-bundle\") pod \"keystone-cron-29523481-bmw8g\" (UID: \"db14278b-852a-47e6-aded-fadfbee5fd52\") " pod="openstack/keystone-cron-29523481-bmw8g" Feb 18 10:01:00 crc kubenswrapper[4556]: I0218 10:01:00.314964 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xh9sk\" (UniqueName: \"kubernetes.io/projected/db14278b-852a-47e6-aded-fadfbee5fd52-kube-api-access-xh9sk\") pod \"keystone-cron-29523481-bmw8g\" (UID: \"db14278b-852a-47e6-aded-fadfbee5fd52\") " pod="openstack/keystone-cron-29523481-bmw8g" Feb 18 10:01:00 crc kubenswrapper[4556]: I0218 10:01:00.315042 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db14278b-852a-47e6-aded-fadfbee5fd52-config-data\") pod \"keystone-cron-29523481-bmw8g\" (UID: \"db14278b-852a-47e6-aded-fadfbee5fd52\") " pod="openstack/keystone-cron-29523481-bmw8g" Feb 18 10:01:00 crc kubenswrapper[4556]: I0218 10:01:00.416448 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db14278b-852a-47e6-aded-fadfbee5fd52-config-data\") pod \"keystone-cron-29523481-bmw8g\" (UID: \"db14278b-852a-47e6-aded-fadfbee5fd52\") " pod="openstack/keystone-cron-29523481-bmw8g" Feb 18 10:01:00 crc kubenswrapper[4556]: I0218 10:01:00.416589 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/db14278b-852a-47e6-aded-fadfbee5fd52-fernet-keys\") pod \"keystone-cron-29523481-bmw8g\" (UID: \"db14278b-852a-47e6-aded-fadfbee5fd52\") " pod="openstack/keystone-cron-29523481-bmw8g" Feb 18 10:01:00 crc kubenswrapper[4556]: I0218 10:01:00.416611 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db14278b-852a-47e6-aded-fadfbee5fd52-combined-ca-bundle\") pod \"keystone-cron-29523481-bmw8g\" (UID: \"db14278b-852a-47e6-aded-fadfbee5fd52\") " pod="openstack/keystone-cron-29523481-bmw8g" Feb 18 10:01:00 crc kubenswrapper[4556]: I0218 10:01:00.416750 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xh9sk\" (UniqueName: \"kubernetes.io/projected/db14278b-852a-47e6-aded-fadfbee5fd52-kube-api-access-xh9sk\") pod \"keystone-cron-29523481-bmw8g\" (UID: \"db14278b-852a-47e6-aded-fadfbee5fd52\") " pod="openstack/keystone-cron-29523481-bmw8g" Feb 18 10:01:00 crc kubenswrapper[4556]: I0218 10:01:00.423388 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db14278b-852a-47e6-aded-fadfbee5fd52-combined-ca-bundle\") pod \"keystone-cron-29523481-bmw8g\" (UID: \"db14278b-852a-47e6-aded-fadfbee5fd52\") " pod="openstack/keystone-cron-29523481-bmw8g" Feb 18 10:01:00 crc kubenswrapper[4556]: I0218 10:01:00.423975 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db14278b-852a-47e6-aded-fadfbee5fd52-config-data\") pod \"keystone-cron-29523481-bmw8g\" (UID: \"db14278b-852a-47e6-aded-fadfbee5fd52\") " pod="openstack/keystone-cron-29523481-bmw8g" Feb 18 10:01:00 crc kubenswrapper[4556]: I0218 10:01:00.424039 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/db14278b-852a-47e6-aded-fadfbee5fd52-fernet-keys\") pod \"keystone-cron-29523481-bmw8g\" (UID: \"db14278b-852a-47e6-aded-fadfbee5fd52\") " pod="openstack/keystone-cron-29523481-bmw8g" Feb 18 10:01:00 crc kubenswrapper[4556]: I0218 10:01:00.431418 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xh9sk\" (UniqueName: \"kubernetes.io/projected/db14278b-852a-47e6-aded-fadfbee5fd52-kube-api-access-xh9sk\") pod \"keystone-cron-29523481-bmw8g\" (UID: \"db14278b-852a-47e6-aded-fadfbee5fd52\") " pod="openstack/keystone-cron-29523481-bmw8g" Feb 18 10:01:00 crc kubenswrapper[4556]: I0218 10:01:00.443384 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-rnjw6_must-gather-vhc8n_106c448a-8930-4be7-aa0e-970ab3d8b33e/gather/0.log" Feb 18 10:01:00 crc kubenswrapper[4556]: I0218 10:01:00.471177 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29523481-bmw8g" Feb 18 10:01:00 crc kubenswrapper[4556]: I0218 10:01:00.869703 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29523481-bmw8g"] Feb 18 10:01:01 crc kubenswrapper[4556]: I0218 10:01:01.727355 4556 patch_prober.go:28] interesting pod/machine-config-daemon-f76hs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 10:01:01 crc kubenswrapper[4556]: I0218 10:01:01.727682 4556 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 10:01:01 crc kubenswrapper[4556]: I0218 10:01:01.727723 4556 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" Feb 18 10:01:01 crc kubenswrapper[4556]: I0218 10:01:01.728230 4556 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"34f3a3491bd005389624fe9685306e5b96d58407e240c9903fadf6abf52b3afd"} pod="openshift-machine-config-operator/machine-config-daemon-f76hs" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 18 10:01:01 crc kubenswrapper[4556]: I0218 10:01:01.728323 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerName="machine-config-daemon" containerID="cri-o://34f3a3491bd005389624fe9685306e5b96d58407e240c9903fadf6abf52b3afd" gracePeriod=600 Feb 18 10:01:01 crc kubenswrapper[4556]: I0218 10:01:01.782393 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29523481-bmw8g" event={"ID":"db14278b-852a-47e6-aded-fadfbee5fd52","Type":"ContainerStarted","Data":"53a127fee440895e9d4bfd6cc3a5571f98f327f1d0ae614b6305f26be28ed77f"} Feb 18 10:01:01 crc kubenswrapper[4556]: I0218 10:01:01.782650 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29523481-bmw8g" event={"ID":"db14278b-852a-47e6-aded-fadfbee5fd52","Type":"ContainerStarted","Data":"026de6ca18429a313588a851350b15b820997fb74614a66eec1dcf7a963c6bc8"} Feb 18 10:01:01 crc kubenswrapper[4556]: I0218 10:01:01.799240 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29523481-bmw8g" podStartSLOduration=1.799226335 podStartE2EDuration="1.799226335s" podCreationTimestamp="2026-02-18 10:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 10:01:01.793791172 +0000 UTC m=+3418.810752143" watchObservedRunningTime="2026-02-18 10:01:01.799226335 +0000 UTC m=+3418.816187315" Feb 18 10:01:02 crc kubenswrapper[4556]: I0218 10:01:02.794657 4556 generic.go:334] "Generic (PLEG): container finished" podID="db14278b-852a-47e6-aded-fadfbee5fd52" containerID="53a127fee440895e9d4bfd6cc3a5571f98f327f1d0ae614b6305f26be28ed77f" exitCode=0 Feb 18 10:01:02 crc kubenswrapper[4556]: I0218 10:01:02.794758 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29523481-bmw8g" event={"ID":"db14278b-852a-47e6-aded-fadfbee5fd52","Type":"ContainerDied","Data":"53a127fee440895e9d4bfd6cc3a5571f98f327f1d0ae614b6305f26be28ed77f"} Feb 18 10:01:02 crc kubenswrapper[4556]: I0218 10:01:02.805828 4556 generic.go:334] "Generic (PLEG): container finished" podID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerID="34f3a3491bd005389624fe9685306e5b96d58407e240c9903fadf6abf52b3afd" exitCode=0 Feb 18 10:01:02 crc kubenswrapper[4556]: I0218 10:01:02.805931 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" event={"ID":"8dac7f27-d3d1-4778-9e54-f273035a1d37","Type":"ContainerDied","Data":"34f3a3491bd005389624fe9685306e5b96d58407e240c9903fadf6abf52b3afd"} Feb 18 10:01:02 crc kubenswrapper[4556]: I0218 10:01:02.806032 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" event={"ID":"8dac7f27-d3d1-4778-9e54-f273035a1d37","Type":"ContainerStarted","Data":"a714b4253c0b6300d44d42956a1cc0e98a0a818e7e867014b610f7e385fb3ddb"} Feb 18 10:01:02 crc kubenswrapper[4556]: I0218 10:01:02.806103 4556 scope.go:117] "RemoveContainer" containerID="9627d5bcc0cabda20a5fc92b3cc4b3dfae93225012caf89894f557937f69e6f7" Feb 18 10:01:03 crc kubenswrapper[4556]: E0218 10:01:03.370676 4556 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7c77aaed_e8e3_4b33_8744_0f96ce8e9c53.slice\": RecentStats: unable to find data in memory cache]" Feb 18 10:01:04 crc kubenswrapper[4556]: I0218 10:01:04.108587 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29523481-bmw8g" Feb 18 10:01:04 crc kubenswrapper[4556]: I0218 10:01:04.202942 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db14278b-852a-47e6-aded-fadfbee5fd52-combined-ca-bundle\") pod \"db14278b-852a-47e6-aded-fadfbee5fd52\" (UID: \"db14278b-852a-47e6-aded-fadfbee5fd52\") " Feb 18 10:01:04 crc kubenswrapper[4556]: I0218 10:01:04.203353 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xh9sk\" (UniqueName: \"kubernetes.io/projected/db14278b-852a-47e6-aded-fadfbee5fd52-kube-api-access-xh9sk\") pod \"db14278b-852a-47e6-aded-fadfbee5fd52\" (UID: \"db14278b-852a-47e6-aded-fadfbee5fd52\") " Feb 18 10:01:04 crc kubenswrapper[4556]: I0218 10:01:04.203408 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/db14278b-852a-47e6-aded-fadfbee5fd52-fernet-keys\") pod \"db14278b-852a-47e6-aded-fadfbee5fd52\" (UID: \"db14278b-852a-47e6-aded-fadfbee5fd52\") " Feb 18 10:01:04 crc kubenswrapper[4556]: I0218 10:01:04.203494 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db14278b-852a-47e6-aded-fadfbee5fd52-config-data\") pod \"db14278b-852a-47e6-aded-fadfbee5fd52\" (UID: \"db14278b-852a-47e6-aded-fadfbee5fd52\") " Feb 18 10:01:04 crc kubenswrapper[4556]: I0218 10:01:04.209989 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db14278b-852a-47e6-aded-fadfbee5fd52-kube-api-access-xh9sk" (OuterVolumeSpecName: "kube-api-access-xh9sk") pod "db14278b-852a-47e6-aded-fadfbee5fd52" (UID: "db14278b-852a-47e6-aded-fadfbee5fd52"). InnerVolumeSpecName "kube-api-access-xh9sk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 10:01:04 crc kubenswrapper[4556]: I0218 10:01:04.211367 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db14278b-852a-47e6-aded-fadfbee5fd52-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "db14278b-852a-47e6-aded-fadfbee5fd52" (UID: "db14278b-852a-47e6-aded-fadfbee5fd52"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 10:01:04 crc kubenswrapper[4556]: I0218 10:01:04.232411 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db14278b-852a-47e6-aded-fadfbee5fd52-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "db14278b-852a-47e6-aded-fadfbee5fd52" (UID: "db14278b-852a-47e6-aded-fadfbee5fd52"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 10:01:04 crc kubenswrapper[4556]: I0218 10:01:04.252646 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db14278b-852a-47e6-aded-fadfbee5fd52-config-data" (OuterVolumeSpecName: "config-data") pod "db14278b-852a-47e6-aded-fadfbee5fd52" (UID: "db14278b-852a-47e6-aded-fadfbee5fd52"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 10:01:04 crc kubenswrapper[4556]: I0218 10:01:04.308641 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xh9sk\" (UniqueName: \"kubernetes.io/projected/db14278b-852a-47e6-aded-fadfbee5fd52-kube-api-access-xh9sk\") on node \"crc\" DevicePath \"\"" Feb 18 10:01:04 crc kubenswrapper[4556]: I0218 10:01:04.308679 4556 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/db14278b-852a-47e6-aded-fadfbee5fd52-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 18 10:01:04 crc kubenswrapper[4556]: I0218 10:01:04.308694 4556 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db14278b-852a-47e6-aded-fadfbee5fd52-config-data\") on node \"crc\" DevicePath \"\"" Feb 18 10:01:04 crc kubenswrapper[4556]: I0218 10:01:04.308707 4556 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db14278b-852a-47e6-aded-fadfbee5fd52-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 10:01:04 crc kubenswrapper[4556]: I0218 10:01:04.834984 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29523481-bmw8g" event={"ID":"db14278b-852a-47e6-aded-fadfbee5fd52","Type":"ContainerDied","Data":"026de6ca18429a313588a851350b15b820997fb74614a66eec1dcf7a963c6bc8"} Feb 18 10:01:04 crc kubenswrapper[4556]: I0218 10:01:04.835498 4556 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="026de6ca18429a313588a851350b15b820997fb74614a66eec1dcf7a963c6bc8" Feb 18 10:01:04 crc kubenswrapper[4556]: I0218 10:01:04.835049 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29523481-bmw8g" Feb 18 10:01:10 crc kubenswrapper[4556]: I0218 10:01:10.091752 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-rnjw6/must-gather-vhc8n"] Feb 18 10:01:10 crc kubenswrapper[4556]: I0218 10:01:10.092423 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-rnjw6/must-gather-vhc8n" podUID="106c448a-8930-4be7-aa0e-970ab3d8b33e" containerName="copy" containerID="cri-o://8ffa55cc5a0ecee01dde67ff58189f72bcb9b29e97e75c66ab7db9e99bff8698" gracePeriod=2 Feb 18 10:01:10 crc kubenswrapper[4556]: I0218 10:01:10.102548 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-rnjw6/must-gather-vhc8n"] Feb 18 10:01:10 crc kubenswrapper[4556]: I0218 10:01:10.544598 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-rnjw6_must-gather-vhc8n_106c448a-8930-4be7-aa0e-970ab3d8b33e/copy/0.log" Feb 18 10:01:10 crc kubenswrapper[4556]: I0218 10:01:10.545319 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rnjw6/must-gather-vhc8n" Feb 18 10:01:10 crc kubenswrapper[4556]: I0218 10:01:10.743713 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d2pjv\" (UniqueName: \"kubernetes.io/projected/106c448a-8930-4be7-aa0e-970ab3d8b33e-kube-api-access-d2pjv\") pod \"106c448a-8930-4be7-aa0e-970ab3d8b33e\" (UID: \"106c448a-8930-4be7-aa0e-970ab3d8b33e\") " Feb 18 10:01:10 crc kubenswrapper[4556]: I0218 10:01:10.744104 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/106c448a-8930-4be7-aa0e-970ab3d8b33e-must-gather-output\") pod \"106c448a-8930-4be7-aa0e-970ab3d8b33e\" (UID: \"106c448a-8930-4be7-aa0e-970ab3d8b33e\") " Feb 18 10:01:10 crc kubenswrapper[4556]: I0218 10:01:10.750428 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/106c448a-8930-4be7-aa0e-970ab3d8b33e-kube-api-access-d2pjv" (OuterVolumeSpecName: "kube-api-access-d2pjv") pod "106c448a-8930-4be7-aa0e-970ab3d8b33e" (UID: "106c448a-8930-4be7-aa0e-970ab3d8b33e"). InnerVolumeSpecName "kube-api-access-d2pjv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 10:01:10 crc kubenswrapper[4556]: I0218 10:01:10.847237 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d2pjv\" (UniqueName: \"kubernetes.io/projected/106c448a-8930-4be7-aa0e-970ab3d8b33e-kube-api-access-d2pjv\") on node \"crc\" DevicePath \"\"" Feb 18 10:01:10 crc kubenswrapper[4556]: I0218 10:01:10.868769 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/106c448a-8930-4be7-aa0e-970ab3d8b33e-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "106c448a-8930-4be7-aa0e-970ab3d8b33e" (UID: "106c448a-8930-4be7-aa0e-970ab3d8b33e"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 10:01:10 crc kubenswrapper[4556]: I0218 10:01:10.881909 4556 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-rnjw6_must-gather-vhc8n_106c448a-8930-4be7-aa0e-970ab3d8b33e/copy/0.log" Feb 18 10:01:10 crc kubenswrapper[4556]: I0218 10:01:10.882285 4556 generic.go:334] "Generic (PLEG): container finished" podID="106c448a-8930-4be7-aa0e-970ab3d8b33e" containerID="8ffa55cc5a0ecee01dde67ff58189f72bcb9b29e97e75c66ab7db9e99bff8698" exitCode=143 Feb 18 10:01:10 crc kubenswrapper[4556]: I0218 10:01:10.882339 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rnjw6/must-gather-vhc8n" Feb 18 10:01:10 crc kubenswrapper[4556]: I0218 10:01:10.882354 4556 scope.go:117] "RemoveContainer" containerID="8ffa55cc5a0ecee01dde67ff58189f72bcb9b29e97e75c66ab7db9e99bff8698" Feb 18 10:01:10 crc kubenswrapper[4556]: I0218 10:01:10.897383 4556 scope.go:117] "RemoveContainer" containerID="ab98b80451544ff53c43502227fd54e98ead544320b3850fccc3876aa72bd333" Feb 18 10:01:10 crc kubenswrapper[4556]: I0218 10:01:10.949525 4556 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/106c448a-8930-4be7-aa0e-970ab3d8b33e-must-gather-output\") on node \"crc\" DevicePath \"\"" Feb 18 10:01:10 crc kubenswrapper[4556]: I0218 10:01:10.958320 4556 scope.go:117] "RemoveContainer" containerID="8ffa55cc5a0ecee01dde67ff58189f72bcb9b29e97e75c66ab7db9e99bff8698" Feb 18 10:01:10 crc kubenswrapper[4556]: E0218 10:01:10.959108 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ffa55cc5a0ecee01dde67ff58189f72bcb9b29e97e75c66ab7db9e99bff8698\": container with ID starting with 8ffa55cc5a0ecee01dde67ff58189f72bcb9b29e97e75c66ab7db9e99bff8698 not found: ID does not exist" containerID="8ffa55cc5a0ecee01dde67ff58189f72bcb9b29e97e75c66ab7db9e99bff8698" Feb 18 10:01:10 crc kubenswrapper[4556]: I0218 10:01:10.959174 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ffa55cc5a0ecee01dde67ff58189f72bcb9b29e97e75c66ab7db9e99bff8698"} err="failed to get container status \"8ffa55cc5a0ecee01dde67ff58189f72bcb9b29e97e75c66ab7db9e99bff8698\": rpc error: code = NotFound desc = could not find container \"8ffa55cc5a0ecee01dde67ff58189f72bcb9b29e97e75c66ab7db9e99bff8698\": container with ID starting with 8ffa55cc5a0ecee01dde67ff58189f72bcb9b29e97e75c66ab7db9e99bff8698 not found: ID does not exist" Feb 18 10:01:10 crc kubenswrapper[4556]: I0218 10:01:10.959211 4556 scope.go:117] "RemoveContainer" containerID="ab98b80451544ff53c43502227fd54e98ead544320b3850fccc3876aa72bd333" Feb 18 10:01:10 crc kubenswrapper[4556]: E0218 10:01:10.959784 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab98b80451544ff53c43502227fd54e98ead544320b3850fccc3876aa72bd333\": container with ID starting with ab98b80451544ff53c43502227fd54e98ead544320b3850fccc3876aa72bd333 not found: ID does not exist" containerID="ab98b80451544ff53c43502227fd54e98ead544320b3850fccc3876aa72bd333" Feb 18 10:01:10 crc kubenswrapper[4556]: I0218 10:01:10.959834 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab98b80451544ff53c43502227fd54e98ead544320b3850fccc3876aa72bd333"} err="failed to get container status \"ab98b80451544ff53c43502227fd54e98ead544320b3850fccc3876aa72bd333\": rpc error: code = NotFound desc = could not find container \"ab98b80451544ff53c43502227fd54e98ead544320b3850fccc3876aa72bd333\": container with ID starting with ab98b80451544ff53c43502227fd54e98ead544320b3850fccc3876aa72bd333 not found: ID does not exist" Feb 18 10:01:11 crc kubenswrapper[4556]: I0218 10:01:11.292569 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="106c448a-8930-4be7-aa0e-970ab3d8b33e" path="/var/lib/kubelet/pods/106c448a-8930-4be7-aa0e-970ab3d8b33e/volumes" Feb 18 10:02:13 crc kubenswrapper[4556]: I0218 10:02:13.329527 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rqbq6"] Feb 18 10:02:13 crc kubenswrapper[4556]: E0218 10:02:13.330421 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="106c448a-8930-4be7-aa0e-970ab3d8b33e" containerName="copy" Feb 18 10:02:13 crc kubenswrapper[4556]: I0218 10:02:13.330436 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="106c448a-8930-4be7-aa0e-970ab3d8b33e" containerName="copy" Feb 18 10:02:13 crc kubenswrapper[4556]: E0218 10:02:13.330460 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="106c448a-8930-4be7-aa0e-970ab3d8b33e" containerName="gather" Feb 18 10:02:13 crc kubenswrapper[4556]: I0218 10:02:13.330466 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="106c448a-8930-4be7-aa0e-970ab3d8b33e" containerName="gather" Feb 18 10:02:13 crc kubenswrapper[4556]: E0218 10:02:13.330486 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db14278b-852a-47e6-aded-fadfbee5fd52" containerName="keystone-cron" Feb 18 10:02:13 crc kubenswrapper[4556]: I0218 10:02:13.330493 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="db14278b-852a-47e6-aded-fadfbee5fd52" containerName="keystone-cron" Feb 18 10:02:13 crc kubenswrapper[4556]: I0218 10:02:13.330690 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="106c448a-8930-4be7-aa0e-970ab3d8b33e" containerName="copy" Feb 18 10:02:13 crc kubenswrapper[4556]: I0218 10:02:13.330703 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="db14278b-852a-47e6-aded-fadfbee5fd52" containerName="keystone-cron" Feb 18 10:02:13 crc kubenswrapper[4556]: I0218 10:02:13.330710 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="106c448a-8930-4be7-aa0e-970ab3d8b33e" containerName="gather" Feb 18 10:02:13 crc kubenswrapper[4556]: I0218 10:02:13.332029 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rqbq6" Feb 18 10:02:13 crc kubenswrapper[4556]: I0218 10:02:13.335335 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q578t\" (UniqueName: \"kubernetes.io/projected/9ac2a19b-6445-4c8b-adbf-a0d81076ba21-kube-api-access-q578t\") pod \"redhat-marketplace-rqbq6\" (UID: \"9ac2a19b-6445-4c8b-adbf-a0d81076ba21\") " pod="openshift-marketplace/redhat-marketplace-rqbq6" Feb 18 10:02:13 crc kubenswrapper[4556]: I0218 10:02:13.335422 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ac2a19b-6445-4c8b-adbf-a0d81076ba21-utilities\") pod \"redhat-marketplace-rqbq6\" (UID: \"9ac2a19b-6445-4c8b-adbf-a0d81076ba21\") " pod="openshift-marketplace/redhat-marketplace-rqbq6" Feb 18 10:02:13 crc kubenswrapper[4556]: I0218 10:02:13.335978 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ac2a19b-6445-4c8b-adbf-a0d81076ba21-catalog-content\") pod \"redhat-marketplace-rqbq6\" (UID: \"9ac2a19b-6445-4c8b-adbf-a0d81076ba21\") " pod="openshift-marketplace/redhat-marketplace-rqbq6" Feb 18 10:02:13 crc kubenswrapper[4556]: I0218 10:02:13.338105 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rqbq6"] Feb 18 10:02:13 crc kubenswrapper[4556]: I0218 10:02:13.437878 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q578t\" (UniqueName: \"kubernetes.io/projected/9ac2a19b-6445-4c8b-adbf-a0d81076ba21-kube-api-access-q578t\") pod \"redhat-marketplace-rqbq6\" (UID: \"9ac2a19b-6445-4c8b-adbf-a0d81076ba21\") " pod="openshift-marketplace/redhat-marketplace-rqbq6" Feb 18 10:02:13 crc kubenswrapper[4556]: I0218 10:02:13.437942 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ac2a19b-6445-4c8b-adbf-a0d81076ba21-utilities\") pod \"redhat-marketplace-rqbq6\" (UID: \"9ac2a19b-6445-4c8b-adbf-a0d81076ba21\") " pod="openshift-marketplace/redhat-marketplace-rqbq6" Feb 18 10:02:13 crc kubenswrapper[4556]: I0218 10:02:13.438128 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ac2a19b-6445-4c8b-adbf-a0d81076ba21-catalog-content\") pod \"redhat-marketplace-rqbq6\" (UID: \"9ac2a19b-6445-4c8b-adbf-a0d81076ba21\") " pod="openshift-marketplace/redhat-marketplace-rqbq6" Feb 18 10:02:13 crc kubenswrapper[4556]: I0218 10:02:13.438519 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ac2a19b-6445-4c8b-adbf-a0d81076ba21-utilities\") pod \"redhat-marketplace-rqbq6\" (UID: \"9ac2a19b-6445-4c8b-adbf-a0d81076ba21\") " pod="openshift-marketplace/redhat-marketplace-rqbq6" Feb 18 10:02:13 crc kubenswrapper[4556]: I0218 10:02:13.438579 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ac2a19b-6445-4c8b-adbf-a0d81076ba21-catalog-content\") pod \"redhat-marketplace-rqbq6\" (UID: \"9ac2a19b-6445-4c8b-adbf-a0d81076ba21\") " pod="openshift-marketplace/redhat-marketplace-rqbq6" Feb 18 10:02:13 crc kubenswrapper[4556]: I0218 10:02:13.454567 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q578t\" (UniqueName: \"kubernetes.io/projected/9ac2a19b-6445-4c8b-adbf-a0d81076ba21-kube-api-access-q578t\") pod \"redhat-marketplace-rqbq6\" (UID: \"9ac2a19b-6445-4c8b-adbf-a0d81076ba21\") " pod="openshift-marketplace/redhat-marketplace-rqbq6" Feb 18 10:02:13 crc kubenswrapper[4556]: I0218 10:02:13.650387 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rqbq6" Feb 18 10:02:14 crc kubenswrapper[4556]: I0218 10:02:14.066555 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rqbq6"] Feb 18 10:02:14 crc kubenswrapper[4556]: I0218 10:02:14.435084 4556 generic.go:334] "Generic (PLEG): container finished" podID="9ac2a19b-6445-4c8b-adbf-a0d81076ba21" containerID="fd04420b78923c4e3b1e22a3bf4c7c03bdcfccc5545175e3eb80ea66321d242d" exitCode=0 Feb 18 10:02:14 crc kubenswrapper[4556]: I0218 10:02:14.435183 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rqbq6" event={"ID":"9ac2a19b-6445-4c8b-adbf-a0d81076ba21","Type":"ContainerDied","Data":"fd04420b78923c4e3b1e22a3bf4c7c03bdcfccc5545175e3eb80ea66321d242d"} Feb 18 10:02:14 crc kubenswrapper[4556]: I0218 10:02:14.435453 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rqbq6" event={"ID":"9ac2a19b-6445-4c8b-adbf-a0d81076ba21","Type":"ContainerStarted","Data":"c6e8a4c3337049f9052defc61374282272361c61eb5684881ab422025d99b7a5"} Feb 18 10:02:15 crc kubenswrapper[4556]: I0218 10:02:15.451205 4556 generic.go:334] "Generic (PLEG): container finished" podID="9ac2a19b-6445-4c8b-adbf-a0d81076ba21" containerID="868e86db500ae0c1a951a4e5485fb2819c2e710eb17bed1cdffe5819c9f69dbe" exitCode=0 Feb 18 10:02:15 crc kubenswrapper[4556]: I0218 10:02:15.451505 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rqbq6" event={"ID":"9ac2a19b-6445-4c8b-adbf-a0d81076ba21","Type":"ContainerDied","Data":"868e86db500ae0c1a951a4e5485fb2819c2e710eb17bed1cdffe5819c9f69dbe"} Feb 18 10:02:16 crc kubenswrapper[4556]: I0218 10:02:16.461811 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rqbq6" event={"ID":"9ac2a19b-6445-4c8b-adbf-a0d81076ba21","Type":"ContainerStarted","Data":"7baab1d4453495cd6e36ab1cbee23aec524a2e1316777a234fc1d8bf12ac5495"} Feb 18 10:02:16 crc kubenswrapper[4556]: I0218 10:02:16.477143 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rqbq6" podStartSLOduration=1.859384676 podStartE2EDuration="3.477130913s" podCreationTimestamp="2026-02-18 10:02:13 +0000 UTC" firstStartedPulling="2026-02-18 10:02:14.436268687 +0000 UTC m=+3491.453229667" lastFinishedPulling="2026-02-18 10:02:16.054014924 +0000 UTC m=+3493.070975904" observedRunningTime="2026-02-18 10:02:16.47603202 +0000 UTC m=+3493.492993000" watchObservedRunningTime="2026-02-18 10:02:16.477130913 +0000 UTC m=+3493.494091892" Feb 18 10:02:23 crc kubenswrapper[4556]: I0218 10:02:23.650601 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rqbq6" Feb 18 10:02:23 crc kubenswrapper[4556]: I0218 10:02:23.651168 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rqbq6" Feb 18 10:02:23 crc kubenswrapper[4556]: I0218 10:02:23.684918 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rqbq6" Feb 18 10:02:24 crc kubenswrapper[4556]: I0218 10:02:24.555168 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rqbq6" Feb 18 10:02:24 crc kubenswrapper[4556]: I0218 10:02:24.600505 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rqbq6"] Feb 18 10:02:26 crc kubenswrapper[4556]: I0218 10:02:26.531201 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-rqbq6" podUID="9ac2a19b-6445-4c8b-adbf-a0d81076ba21" containerName="registry-server" containerID="cri-o://7baab1d4453495cd6e36ab1cbee23aec524a2e1316777a234fc1d8bf12ac5495" gracePeriod=2 Feb 18 10:02:26 crc kubenswrapper[4556]: I0218 10:02:26.887930 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rqbq6" Feb 18 10:02:27 crc kubenswrapper[4556]: I0218 10:02:27.005064 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q578t\" (UniqueName: \"kubernetes.io/projected/9ac2a19b-6445-4c8b-adbf-a0d81076ba21-kube-api-access-q578t\") pod \"9ac2a19b-6445-4c8b-adbf-a0d81076ba21\" (UID: \"9ac2a19b-6445-4c8b-adbf-a0d81076ba21\") " Feb 18 10:02:27 crc kubenswrapper[4556]: I0218 10:02:27.005200 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ac2a19b-6445-4c8b-adbf-a0d81076ba21-utilities\") pod \"9ac2a19b-6445-4c8b-adbf-a0d81076ba21\" (UID: \"9ac2a19b-6445-4c8b-adbf-a0d81076ba21\") " Feb 18 10:02:27 crc kubenswrapper[4556]: I0218 10:02:27.005277 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ac2a19b-6445-4c8b-adbf-a0d81076ba21-catalog-content\") pod \"9ac2a19b-6445-4c8b-adbf-a0d81076ba21\" (UID: \"9ac2a19b-6445-4c8b-adbf-a0d81076ba21\") " Feb 18 10:02:27 crc kubenswrapper[4556]: I0218 10:02:27.006259 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ac2a19b-6445-4c8b-adbf-a0d81076ba21-utilities" (OuterVolumeSpecName: "utilities") pod "9ac2a19b-6445-4c8b-adbf-a0d81076ba21" (UID: "9ac2a19b-6445-4c8b-adbf-a0d81076ba21"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 10:02:27 crc kubenswrapper[4556]: I0218 10:02:27.011040 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ac2a19b-6445-4c8b-adbf-a0d81076ba21-kube-api-access-q578t" (OuterVolumeSpecName: "kube-api-access-q578t") pod "9ac2a19b-6445-4c8b-adbf-a0d81076ba21" (UID: "9ac2a19b-6445-4c8b-adbf-a0d81076ba21"). InnerVolumeSpecName "kube-api-access-q578t". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 10:02:27 crc kubenswrapper[4556]: I0218 10:02:27.024928 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ac2a19b-6445-4c8b-adbf-a0d81076ba21-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9ac2a19b-6445-4c8b-adbf-a0d81076ba21" (UID: "9ac2a19b-6445-4c8b-adbf-a0d81076ba21"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 10:02:27 crc kubenswrapper[4556]: I0218 10:02:27.107345 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q578t\" (UniqueName: \"kubernetes.io/projected/9ac2a19b-6445-4c8b-adbf-a0d81076ba21-kube-api-access-q578t\") on node \"crc\" DevicePath \"\"" Feb 18 10:02:27 crc kubenswrapper[4556]: I0218 10:02:27.107379 4556 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ac2a19b-6445-4c8b-adbf-a0d81076ba21-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 10:02:27 crc kubenswrapper[4556]: I0218 10:02:27.107390 4556 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ac2a19b-6445-4c8b-adbf-a0d81076ba21-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 10:02:27 crc kubenswrapper[4556]: I0218 10:02:27.539981 4556 generic.go:334] "Generic (PLEG): container finished" podID="9ac2a19b-6445-4c8b-adbf-a0d81076ba21" containerID="7baab1d4453495cd6e36ab1cbee23aec524a2e1316777a234fc1d8bf12ac5495" exitCode=0 Feb 18 10:02:27 crc kubenswrapper[4556]: I0218 10:02:27.540030 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rqbq6" Feb 18 10:02:27 crc kubenswrapper[4556]: I0218 10:02:27.540047 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rqbq6" event={"ID":"9ac2a19b-6445-4c8b-adbf-a0d81076ba21","Type":"ContainerDied","Data":"7baab1d4453495cd6e36ab1cbee23aec524a2e1316777a234fc1d8bf12ac5495"} Feb 18 10:02:27 crc kubenswrapper[4556]: I0218 10:02:27.540443 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rqbq6" event={"ID":"9ac2a19b-6445-4c8b-adbf-a0d81076ba21","Type":"ContainerDied","Data":"c6e8a4c3337049f9052defc61374282272361c61eb5684881ab422025d99b7a5"} Feb 18 10:02:27 crc kubenswrapper[4556]: I0218 10:02:27.540480 4556 scope.go:117] "RemoveContainer" containerID="7baab1d4453495cd6e36ab1cbee23aec524a2e1316777a234fc1d8bf12ac5495" Feb 18 10:02:27 crc kubenswrapper[4556]: I0218 10:02:27.556087 4556 scope.go:117] "RemoveContainer" containerID="868e86db500ae0c1a951a4e5485fb2819c2e710eb17bed1cdffe5819c9f69dbe" Feb 18 10:02:27 crc kubenswrapper[4556]: I0218 10:02:27.556650 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rqbq6"] Feb 18 10:02:27 crc kubenswrapper[4556]: I0218 10:02:27.562982 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-rqbq6"] Feb 18 10:02:27 crc kubenswrapper[4556]: I0218 10:02:27.573857 4556 scope.go:117] "RemoveContainer" containerID="fd04420b78923c4e3b1e22a3bf4c7c03bdcfccc5545175e3eb80ea66321d242d" Feb 18 10:02:27 crc kubenswrapper[4556]: I0218 10:02:27.606903 4556 scope.go:117] "RemoveContainer" containerID="7baab1d4453495cd6e36ab1cbee23aec524a2e1316777a234fc1d8bf12ac5495" Feb 18 10:02:27 crc kubenswrapper[4556]: E0218 10:02:27.607175 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7baab1d4453495cd6e36ab1cbee23aec524a2e1316777a234fc1d8bf12ac5495\": container with ID starting with 7baab1d4453495cd6e36ab1cbee23aec524a2e1316777a234fc1d8bf12ac5495 not found: ID does not exist" containerID="7baab1d4453495cd6e36ab1cbee23aec524a2e1316777a234fc1d8bf12ac5495" Feb 18 10:02:27 crc kubenswrapper[4556]: I0218 10:02:27.607207 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7baab1d4453495cd6e36ab1cbee23aec524a2e1316777a234fc1d8bf12ac5495"} err="failed to get container status \"7baab1d4453495cd6e36ab1cbee23aec524a2e1316777a234fc1d8bf12ac5495\": rpc error: code = NotFound desc = could not find container \"7baab1d4453495cd6e36ab1cbee23aec524a2e1316777a234fc1d8bf12ac5495\": container with ID starting with 7baab1d4453495cd6e36ab1cbee23aec524a2e1316777a234fc1d8bf12ac5495 not found: ID does not exist" Feb 18 10:02:27 crc kubenswrapper[4556]: I0218 10:02:27.607229 4556 scope.go:117] "RemoveContainer" containerID="868e86db500ae0c1a951a4e5485fb2819c2e710eb17bed1cdffe5819c9f69dbe" Feb 18 10:02:27 crc kubenswrapper[4556]: E0218 10:02:27.607626 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"868e86db500ae0c1a951a4e5485fb2819c2e710eb17bed1cdffe5819c9f69dbe\": container with ID starting with 868e86db500ae0c1a951a4e5485fb2819c2e710eb17bed1cdffe5819c9f69dbe not found: ID does not exist" containerID="868e86db500ae0c1a951a4e5485fb2819c2e710eb17bed1cdffe5819c9f69dbe" Feb 18 10:02:27 crc kubenswrapper[4556]: I0218 10:02:27.607648 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"868e86db500ae0c1a951a4e5485fb2819c2e710eb17bed1cdffe5819c9f69dbe"} err="failed to get container status \"868e86db500ae0c1a951a4e5485fb2819c2e710eb17bed1cdffe5819c9f69dbe\": rpc error: code = NotFound desc = could not find container \"868e86db500ae0c1a951a4e5485fb2819c2e710eb17bed1cdffe5819c9f69dbe\": container with ID starting with 868e86db500ae0c1a951a4e5485fb2819c2e710eb17bed1cdffe5819c9f69dbe not found: ID does not exist" Feb 18 10:02:27 crc kubenswrapper[4556]: I0218 10:02:27.607660 4556 scope.go:117] "RemoveContainer" containerID="fd04420b78923c4e3b1e22a3bf4c7c03bdcfccc5545175e3eb80ea66321d242d" Feb 18 10:02:27 crc kubenswrapper[4556]: E0218 10:02:27.607916 4556 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd04420b78923c4e3b1e22a3bf4c7c03bdcfccc5545175e3eb80ea66321d242d\": container with ID starting with fd04420b78923c4e3b1e22a3bf4c7c03bdcfccc5545175e3eb80ea66321d242d not found: ID does not exist" containerID="fd04420b78923c4e3b1e22a3bf4c7c03bdcfccc5545175e3eb80ea66321d242d" Feb 18 10:02:27 crc kubenswrapper[4556]: I0218 10:02:27.607937 4556 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd04420b78923c4e3b1e22a3bf4c7c03bdcfccc5545175e3eb80ea66321d242d"} err="failed to get container status \"fd04420b78923c4e3b1e22a3bf4c7c03bdcfccc5545175e3eb80ea66321d242d\": rpc error: code = NotFound desc = could not find container \"fd04420b78923c4e3b1e22a3bf4c7c03bdcfccc5545175e3eb80ea66321d242d\": container with ID starting with fd04420b78923c4e3b1e22a3bf4c7c03bdcfccc5545175e3eb80ea66321d242d not found: ID does not exist" Feb 18 10:02:29 crc kubenswrapper[4556]: I0218 10:02:29.292759 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ac2a19b-6445-4c8b-adbf-a0d81076ba21" path="/var/lib/kubelet/pods/9ac2a19b-6445-4c8b-adbf-a0d81076ba21/volumes" Feb 18 10:02:42 crc kubenswrapper[4556]: I0218 10:02:42.228681 4556 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-ghqhx"] Feb 18 10:02:42 crc kubenswrapper[4556]: E0218 10:02:42.229768 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ac2a19b-6445-4c8b-adbf-a0d81076ba21" containerName="registry-server" Feb 18 10:02:42 crc kubenswrapper[4556]: I0218 10:02:42.229789 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ac2a19b-6445-4c8b-adbf-a0d81076ba21" containerName="registry-server" Feb 18 10:02:42 crc kubenswrapper[4556]: E0218 10:02:42.229832 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ac2a19b-6445-4c8b-adbf-a0d81076ba21" containerName="extract-utilities" Feb 18 10:02:42 crc kubenswrapper[4556]: I0218 10:02:42.229838 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ac2a19b-6445-4c8b-adbf-a0d81076ba21" containerName="extract-utilities" Feb 18 10:02:42 crc kubenswrapper[4556]: E0218 10:02:42.229854 4556 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ac2a19b-6445-4c8b-adbf-a0d81076ba21" containerName="extract-content" Feb 18 10:02:42 crc kubenswrapper[4556]: I0218 10:02:42.229862 4556 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ac2a19b-6445-4c8b-adbf-a0d81076ba21" containerName="extract-content" Feb 18 10:02:42 crc kubenswrapper[4556]: I0218 10:02:42.230069 4556 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ac2a19b-6445-4c8b-adbf-a0d81076ba21" containerName="registry-server" Feb 18 10:02:42 crc kubenswrapper[4556]: I0218 10:02:42.231438 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ghqhx" Feb 18 10:02:42 crc kubenswrapper[4556]: I0218 10:02:42.247303 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ghqhx"] Feb 18 10:02:42 crc kubenswrapper[4556]: I0218 10:02:42.360900 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/803def4a-8ae5-4489-9a44-e31c040113b7-catalog-content\") pod \"redhat-operators-ghqhx\" (UID: \"803def4a-8ae5-4489-9a44-e31c040113b7\") " pod="openshift-marketplace/redhat-operators-ghqhx" Feb 18 10:02:42 crc kubenswrapper[4556]: I0218 10:02:42.361532 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/803def4a-8ae5-4489-9a44-e31c040113b7-utilities\") pod \"redhat-operators-ghqhx\" (UID: \"803def4a-8ae5-4489-9a44-e31c040113b7\") " pod="openshift-marketplace/redhat-operators-ghqhx" Feb 18 10:02:42 crc kubenswrapper[4556]: I0218 10:02:42.361610 4556 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nckn5\" (UniqueName: \"kubernetes.io/projected/803def4a-8ae5-4489-9a44-e31c040113b7-kube-api-access-nckn5\") pod \"redhat-operators-ghqhx\" (UID: \"803def4a-8ae5-4489-9a44-e31c040113b7\") " pod="openshift-marketplace/redhat-operators-ghqhx" Feb 18 10:02:42 crc kubenswrapper[4556]: I0218 10:02:42.464031 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/803def4a-8ae5-4489-9a44-e31c040113b7-catalog-content\") pod \"redhat-operators-ghqhx\" (UID: \"803def4a-8ae5-4489-9a44-e31c040113b7\") " pod="openshift-marketplace/redhat-operators-ghqhx" Feb 18 10:02:42 crc kubenswrapper[4556]: I0218 10:02:42.464453 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/803def4a-8ae5-4489-9a44-e31c040113b7-utilities\") pod \"redhat-operators-ghqhx\" (UID: \"803def4a-8ae5-4489-9a44-e31c040113b7\") " pod="openshift-marketplace/redhat-operators-ghqhx" Feb 18 10:02:42 crc kubenswrapper[4556]: I0218 10:02:42.464518 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/803def4a-8ae5-4489-9a44-e31c040113b7-catalog-content\") pod \"redhat-operators-ghqhx\" (UID: \"803def4a-8ae5-4489-9a44-e31c040113b7\") " pod="openshift-marketplace/redhat-operators-ghqhx" Feb 18 10:02:42 crc kubenswrapper[4556]: I0218 10:02:42.464628 4556 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nckn5\" (UniqueName: \"kubernetes.io/projected/803def4a-8ae5-4489-9a44-e31c040113b7-kube-api-access-nckn5\") pod \"redhat-operators-ghqhx\" (UID: \"803def4a-8ae5-4489-9a44-e31c040113b7\") " pod="openshift-marketplace/redhat-operators-ghqhx" Feb 18 10:02:42 crc kubenswrapper[4556]: I0218 10:02:42.464833 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/803def4a-8ae5-4489-9a44-e31c040113b7-utilities\") pod \"redhat-operators-ghqhx\" (UID: \"803def4a-8ae5-4489-9a44-e31c040113b7\") " pod="openshift-marketplace/redhat-operators-ghqhx" Feb 18 10:02:42 crc kubenswrapper[4556]: I0218 10:02:42.482419 4556 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nckn5\" (UniqueName: \"kubernetes.io/projected/803def4a-8ae5-4489-9a44-e31c040113b7-kube-api-access-nckn5\") pod \"redhat-operators-ghqhx\" (UID: \"803def4a-8ae5-4489-9a44-e31c040113b7\") " pod="openshift-marketplace/redhat-operators-ghqhx" Feb 18 10:02:42 crc kubenswrapper[4556]: I0218 10:02:42.550231 4556 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ghqhx" Feb 18 10:02:42 crc kubenswrapper[4556]: I0218 10:02:42.956142 4556 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ghqhx"] Feb 18 10:02:43 crc kubenswrapper[4556]: I0218 10:02:43.672906 4556 generic.go:334] "Generic (PLEG): container finished" podID="803def4a-8ae5-4489-9a44-e31c040113b7" containerID="b63b0a2512ed8488a25f1bd1792cf9ba73dd0c1e9c9207fcdda3ea899bf42a90" exitCode=0 Feb 18 10:02:43 crc kubenswrapper[4556]: I0218 10:02:43.673143 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ghqhx" event={"ID":"803def4a-8ae5-4489-9a44-e31c040113b7","Type":"ContainerDied","Data":"b63b0a2512ed8488a25f1bd1792cf9ba73dd0c1e9c9207fcdda3ea899bf42a90"} Feb 18 10:02:43 crc kubenswrapper[4556]: I0218 10:02:43.673187 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ghqhx" event={"ID":"803def4a-8ae5-4489-9a44-e31c040113b7","Type":"ContainerStarted","Data":"45d02f6a5df397b63a9f64b10e0d37cb8a8703aa453e3d9fec757466c3ab8780"} Feb 18 10:02:44 crc kubenswrapper[4556]: I0218 10:02:44.684311 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ghqhx" event={"ID":"803def4a-8ae5-4489-9a44-e31c040113b7","Type":"ContainerStarted","Data":"9a223c1dd430323b08570a1cbe47c2ddcebcfdd463e7cd9c59367d7076079ae6"} Feb 18 10:02:45 crc kubenswrapper[4556]: I0218 10:02:45.694506 4556 generic.go:334] "Generic (PLEG): container finished" podID="803def4a-8ae5-4489-9a44-e31c040113b7" containerID="9a223c1dd430323b08570a1cbe47c2ddcebcfdd463e7cd9c59367d7076079ae6" exitCode=0 Feb 18 10:02:45 crc kubenswrapper[4556]: I0218 10:02:45.694592 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ghqhx" event={"ID":"803def4a-8ae5-4489-9a44-e31c040113b7","Type":"ContainerDied","Data":"9a223c1dd430323b08570a1cbe47c2ddcebcfdd463e7cd9c59367d7076079ae6"} Feb 18 10:02:46 crc kubenswrapper[4556]: I0218 10:02:46.705827 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ghqhx" event={"ID":"803def4a-8ae5-4489-9a44-e31c040113b7","Type":"ContainerStarted","Data":"6de276ca1c8e3078c6326c2d167f9ab00efe143573f6a5a0e4b87c53fc7cd289"} Feb 18 10:02:46 crc kubenswrapper[4556]: I0218 10:02:46.728305 4556 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-ghqhx" podStartSLOduration=2.211286328 podStartE2EDuration="4.728281076s" podCreationTimestamp="2026-02-18 10:02:42 +0000 UTC" firstStartedPulling="2026-02-18 10:02:43.674827496 +0000 UTC m=+3520.691788476" lastFinishedPulling="2026-02-18 10:02:46.191822243 +0000 UTC m=+3523.208783224" observedRunningTime="2026-02-18 10:02:46.721288336 +0000 UTC m=+3523.738249317" watchObservedRunningTime="2026-02-18 10:02:46.728281076 +0000 UTC m=+3523.745242057" Feb 18 10:02:52 crc kubenswrapper[4556]: I0218 10:02:52.550939 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-ghqhx" Feb 18 10:02:52 crc kubenswrapper[4556]: I0218 10:02:52.551785 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-ghqhx" Feb 18 10:02:52 crc kubenswrapper[4556]: I0218 10:02:52.588463 4556 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-ghqhx" Feb 18 10:02:52 crc kubenswrapper[4556]: I0218 10:02:52.797674 4556 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-ghqhx" Feb 18 10:02:52 crc kubenswrapper[4556]: I0218 10:02:52.846071 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ghqhx"] Feb 18 10:02:54 crc kubenswrapper[4556]: I0218 10:02:54.776405 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-ghqhx" podUID="803def4a-8ae5-4489-9a44-e31c040113b7" containerName="registry-server" containerID="cri-o://6de276ca1c8e3078c6326c2d167f9ab00efe143573f6a5a0e4b87c53fc7cd289" gracePeriod=2 Feb 18 10:02:56 crc kubenswrapper[4556]: I0218 10:02:56.790848 4556 generic.go:334] "Generic (PLEG): container finished" podID="803def4a-8ae5-4489-9a44-e31c040113b7" containerID="6de276ca1c8e3078c6326c2d167f9ab00efe143573f6a5a0e4b87c53fc7cd289" exitCode=0 Feb 18 10:02:56 crc kubenswrapper[4556]: I0218 10:02:56.790929 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ghqhx" event={"ID":"803def4a-8ae5-4489-9a44-e31c040113b7","Type":"ContainerDied","Data":"6de276ca1c8e3078c6326c2d167f9ab00efe143573f6a5a0e4b87c53fc7cd289"} Feb 18 10:02:56 crc kubenswrapper[4556]: I0218 10:02:56.977520 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ghqhx" Feb 18 10:02:57 crc kubenswrapper[4556]: I0218 10:02:57.149778 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/803def4a-8ae5-4489-9a44-e31c040113b7-catalog-content\") pod \"803def4a-8ae5-4489-9a44-e31c040113b7\" (UID: \"803def4a-8ae5-4489-9a44-e31c040113b7\") " Feb 18 10:02:57 crc kubenswrapper[4556]: I0218 10:02:57.149830 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nckn5\" (UniqueName: \"kubernetes.io/projected/803def4a-8ae5-4489-9a44-e31c040113b7-kube-api-access-nckn5\") pod \"803def4a-8ae5-4489-9a44-e31c040113b7\" (UID: \"803def4a-8ae5-4489-9a44-e31c040113b7\") " Feb 18 10:02:57 crc kubenswrapper[4556]: I0218 10:02:57.150684 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/803def4a-8ae5-4489-9a44-e31c040113b7-utilities" (OuterVolumeSpecName: "utilities") pod "803def4a-8ae5-4489-9a44-e31c040113b7" (UID: "803def4a-8ae5-4489-9a44-e31c040113b7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 10:02:57 crc kubenswrapper[4556]: I0218 10:02:57.150729 4556 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/803def4a-8ae5-4489-9a44-e31c040113b7-utilities\") pod \"803def4a-8ae5-4489-9a44-e31c040113b7\" (UID: \"803def4a-8ae5-4489-9a44-e31c040113b7\") " Feb 18 10:02:57 crc kubenswrapper[4556]: I0218 10:02:57.151485 4556 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/803def4a-8ae5-4489-9a44-e31c040113b7-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 10:02:57 crc kubenswrapper[4556]: I0218 10:02:57.154518 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/803def4a-8ae5-4489-9a44-e31c040113b7-kube-api-access-nckn5" (OuterVolumeSpecName: "kube-api-access-nckn5") pod "803def4a-8ae5-4489-9a44-e31c040113b7" (UID: "803def4a-8ae5-4489-9a44-e31c040113b7"). InnerVolumeSpecName "kube-api-access-nckn5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 10:02:57 crc kubenswrapper[4556]: I0218 10:02:57.239533 4556 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/803def4a-8ae5-4489-9a44-e31c040113b7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "803def4a-8ae5-4489-9a44-e31c040113b7" (UID: "803def4a-8ae5-4489-9a44-e31c040113b7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 10:02:57 crc kubenswrapper[4556]: I0218 10:02:57.253114 4556 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/803def4a-8ae5-4489-9a44-e31c040113b7-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 10:02:57 crc kubenswrapper[4556]: I0218 10:02:57.253138 4556 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nckn5\" (UniqueName: \"kubernetes.io/projected/803def4a-8ae5-4489-9a44-e31c040113b7-kube-api-access-nckn5\") on node \"crc\" DevicePath \"\"" Feb 18 10:02:57 crc kubenswrapper[4556]: I0218 10:02:57.800793 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ghqhx" event={"ID":"803def4a-8ae5-4489-9a44-e31c040113b7","Type":"ContainerDied","Data":"45d02f6a5df397b63a9f64b10e0d37cb8a8703aa453e3d9fec757466c3ab8780"} Feb 18 10:02:57 crc kubenswrapper[4556]: I0218 10:02:57.800839 4556 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ghqhx" Feb 18 10:02:57 crc kubenswrapper[4556]: I0218 10:02:57.800860 4556 scope.go:117] "RemoveContainer" containerID="6de276ca1c8e3078c6326c2d167f9ab00efe143573f6a5a0e4b87c53fc7cd289" Feb 18 10:02:57 crc kubenswrapper[4556]: I0218 10:02:57.818263 4556 scope.go:117] "RemoveContainer" containerID="9a223c1dd430323b08570a1cbe47c2ddcebcfdd463e7cd9c59367d7076079ae6" Feb 18 10:02:57 crc kubenswrapper[4556]: I0218 10:02:57.819091 4556 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ghqhx"] Feb 18 10:02:57 crc kubenswrapper[4556]: I0218 10:02:57.825183 4556 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-ghqhx"] Feb 18 10:02:57 crc kubenswrapper[4556]: I0218 10:02:57.839682 4556 scope.go:117] "RemoveContainer" containerID="b63b0a2512ed8488a25f1bd1792cf9ba73dd0c1e9c9207fcdda3ea899bf42a90" Feb 18 10:02:59 crc kubenswrapper[4556]: I0218 10:02:59.291878 4556 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="803def4a-8ae5-4489-9a44-e31c040113b7" path="/var/lib/kubelet/pods/803def4a-8ae5-4489-9a44-e31c040113b7/volumes" Feb 18 10:03:01 crc kubenswrapper[4556]: I0218 10:03:01.727379 4556 patch_prober.go:28] interesting pod/machine-config-daemon-f76hs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 10:03:01 crc kubenswrapper[4556]: I0218 10:03:01.727724 4556 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 10:03:11 crc kubenswrapper[4556]: I0218 10:03:11.733495 4556 scope.go:117] "RemoveContainer" containerID="6b05317fa9359cf529529b893f3580663d1cba39dddbed9ef83b550afce96760" Feb 18 10:03:31 crc kubenswrapper[4556]: I0218 10:03:31.727275 4556 patch_prober.go:28] interesting pod/machine-config-daemon-f76hs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 10:03:31 crc kubenswrapper[4556]: I0218 10:03:31.727968 4556 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 10:04:01 crc kubenswrapper[4556]: I0218 10:04:01.727177 4556 patch_prober.go:28] interesting pod/machine-config-daemon-f76hs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 10:04:01 crc kubenswrapper[4556]: I0218 10:04:01.727938 4556 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 10:04:01 crc kubenswrapper[4556]: I0218 10:04:01.728002 4556 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" Feb 18 10:04:01 crc kubenswrapper[4556]: I0218 10:04:01.729143 4556 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a714b4253c0b6300d44d42956a1cc0e98a0a818e7e867014b610f7e385fb3ddb"} pod="openshift-machine-config-operator/machine-config-daemon-f76hs" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 18 10:04:01 crc kubenswrapper[4556]: I0218 10:04:01.729245 4556 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerName="machine-config-daemon" containerID="cri-o://a714b4253c0b6300d44d42956a1cc0e98a0a818e7e867014b610f7e385fb3ddb" gracePeriod=600 Feb 18 10:04:01 crc kubenswrapper[4556]: E0218 10:04:01.850177 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 10:04:02 crc kubenswrapper[4556]: I0218 10:04:02.340300 4556 generic.go:334] "Generic (PLEG): container finished" podID="8dac7f27-d3d1-4778-9e54-f273035a1d37" containerID="a714b4253c0b6300d44d42956a1cc0e98a0a818e7e867014b610f7e385fb3ddb" exitCode=0 Feb 18 10:04:02 crc kubenswrapper[4556]: I0218 10:04:02.340355 4556 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" event={"ID":"8dac7f27-d3d1-4778-9e54-f273035a1d37","Type":"ContainerDied","Data":"a714b4253c0b6300d44d42956a1cc0e98a0a818e7e867014b610f7e385fb3ddb"} Feb 18 10:04:02 crc kubenswrapper[4556]: I0218 10:04:02.340396 4556 scope.go:117] "RemoveContainer" containerID="34f3a3491bd005389624fe9685306e5b96d58407e240c9903fadf6abf52b3afd" Feb 18 10:04:02 crc kubenswrapper[4556]: I0218 10:04:02.340846 4556 scope.go:117] "RemoveContainer" containerID="a714b4253c0b6300d44d42956a1cc0e98a0a818e7e867014b610f7e385fb3ddb" Feb 18 10:04:02 crc kubenswrapper[4556]: E0218 10:04:02.341070 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 10:04:15 crc kubenswrapper[4556]: I0218 10:04:15.287006 4556 scope.go:117] "RemoveContainer" containerID="a714b4253c0b6300d44d42956a1cc0e98a0a818e7e867014b610f7e385fb3ddb" Feb 18 10:04:15 crc kubenswrapper[4556]: E0218 10:04:15.287790 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 10:04:30 crc kubenswrapper[4556]: I0218 10:04:30.283193 4556 scope.go:117] "RemoveContainer" containerID="a714b4253c0b6300d44d42956a1cc0e98a0a818e7e867014b610f7e385fb3ddb" Feb 18 10:04:30 crc kubenswrapper[4556]: E0218 10:04:30.284010 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 10:04:44 crc kubenswrapper[4556]: I0218 10:04:44.283375 4556 scope.go:117] "RemoveContainer" containerID="a714b4253c0b6300d44d42956a1cc0e98a0a818e7e867014b610f7e385fb3ddb" Feb 18 10:04:44 crc kubenswrapper[4556]: E0218 10:04:44.284407 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 10:04:58 crc kubenswrapper[4556]: I0218 10:04:58.283127 4556 scope.go:117] "RemoveContainer" containerID="a714b4253c0b6300d44d42956a1cc0e98a0a818e7e867014b610f7e385fb3ddb" Feb 18 10:04:58 crc kubenswrapper[4556]: E0218 10:04:58.284057 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 10:05:13 crc kubenswrapper[4556]: I0218 10:05:13.287828 4556 scope.go:117] "RemoveContainer" containerID="a714b4253c0b6300d44d42956a1cc0e98a0a818e7e867014b610f7e385fb3ddb" Feb 18 10:05:13 crc kubenswrapper[4556]: E0218 10:05:13.288885 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" Feb 18 10:05:26 crc kubenswrapper[4556]: I0218 10:05:26.282263 4556 scope.go:117] "RemoveContainer" containerID="a714b4253c0b6300d44d42956a1cc0e98a0a818e7e867014b610f7e385fb3ddb" Feb 18 10:05:26 crc kubenswrapper[4556]: E0218 10:05:26.283255 4556 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f76hs_openshift-machine-config-operator(8dac7f27-d3d1-4778-9e54-f273035a1d37)\"" pod="openshift-machine-config-operator/machine-config-daemon-f76hs" podUID="8dac7f27-d3d1-4778-9e54-f273035a1d37" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515145307356024456 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015145307357017374 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015145277765016530 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015145277765015500 5ustar corecore